Raising yield potential in wheat: increasing photosynthesis capacity and efficiency
USDA-ARS?s Scientific Manuscript database
Increasing wheat yields to help to ensure food security is a major challenge. Meeting this challenge requires a quantum improvement in the yield potential of wheat. Past increases in yield potential have largely resulted from improvements in harvest index not through increased biomass. Further large...
Querying Large Biological Network Datasets
ERIC Educational Resources Information Center
Gulsoy, Gunhan
2013-01-01
New experimental methods has resulted in increasing amount of genetic interaction data to be generated every day. Biological networks are used to store genetic interaction data gathered. Increasing amount of data available requires fast large scale analysis methods. Therefore, we address the problem of querying large biological network datasets.…
Big questions, big science: meeting the challenges of global ecology.
Schimel, David; Keller, Michael
2015-04-01
Ecologists are increasingly tackling questions that require significant infrastucture, large experiments, networks of observations, and complex data and computation. Key hypotheses in ecology increasingly require more investment, and larger data sets to be tested than can be collected by a single investigator's or s group of investigator's labs, sustained for longer than a typical grant. Large-scale projects are expensive, so their scientific return on the investment has to justify the opportunity cost-the science foregone because resources were expended on a large project rather than supporting a number of individual projects. In addition, their management must be accountable and efficient in the use of significant resources, requiring the use of formal systems engineering and project management to mitigate risk of failure. Mapping the scientific method into formal project management requires both scientists able to work in the context, and a project implementation team sensitive to the unique requirements of ecology. Sponsoring agencies, under pressure from external and internal forces, experience many pressures that push them towards counterproductive project management but a scientific community aware and experienced in large project science can mitigate these tendencies. For big ecology to result in great science, ecologists must become informed, aware and engaged in the advocacy and governance of large ecological projects.
ERIC Educational Resources Information Center
Atkinson, Maxine P.; Czaja, Ronald F.; Brewster, Zachary B.
2006-01-01
Sociologists can make meaningful contributions to quantitative literacy by teaching sociological research skills in sociology classes, including introductory courses. We report on the effectiveness of requiring a research module in a large introductory class. The module is designed to teach both basic research skills and to increase awareness of…
Correlation between Academic and Skills-Based Tests in Computer Networks
ERIC Educational Resources Information Center
Buchanan, William
2006-01-01
Computing-related programmes and modules have many problems, especially related to large class sizes, large-scale plagiarism, module franchising, and an increased requirement from students for increased amounts of hands-on, practical work. This paper presents a practical computer networks module which uses a mixture of online examinations and a…
Detwiler, R.L.; Mehl, S.; Rajaram, H.; Cheung, W.W.
2002-01-01
Numerical solution of large-scale ground water flow and transport problems is often constrained by the convergence behavior of the iterative solvers used to solve the resulting systems of equations. We demonstrate the ability of an algebraic multigrid algorithm (AMG) to efficiently solve the large, sparse systems of equations that result from computational models of ground water flow and transport in large and complex domains. Unlike geometric multigrid methods, this algorithm is applicable to problems in complex flow geometries, such as those encountered in pore-scale modeling of two-phase flow and transport. We integrated AMG into MODFLOW 2000 to compare two- and three-dimensional flow simulations using AMG to simulations using PCG2, a preconditioned conjugate gradient solver that uses the modified incomplete Cholesky preconditioner and is included with MODFLOW 2000. CPU times required for convergence with AMG were up to 140 times faster than those for PCG2. The cost of this increased speed was up to a nine-fold increase in required random access memory (RAM) for the three-dimensional problems and up to a four-fold increase in required RAM for the two-dimensional problems. We also compared two-dimensional numerical simulations of steady-state transport using AMG and the generalized minimum residual method with an incomplete LU-decomposition preconditioner. For these transport simulations, AMG yielded increased speeds of up to 17 times with only a 20% increase in required RAM. The ability of AMG to solve flow and transport problems in large, complex flow systems and its ready availability make it an ideal solver for use in both field-scale and pore-scale modeling.
DEM Based Modeling: Grid or TIN? The Answer Depends
NASA Astrophysics Data System (ADS)
Ogden, F. L.; Moreno, H. A.
2015-12-01
The availability of petascale supercomputing power has enabled process-based hydrological simulations on large watersheds and two-way coupling with mesoscale atmospheric models. Of course with increasing watershed scale come corresponding increases in watershed complexity, including wide ranging water management infrastructure and objectives, and ever increasing demands for forcing data. Simulations of large watersheds using grid-based models apply a fixed resolution over the entire watershed. In large watersheds, this means an enormous number of grids, or coarsening of the grid resolution to reduce memory requirements. One alternative to grid-based methods is the triangular irregular network (TIN) approach. TINs provide the flexibility of variable resolution, which allows optimization of computational resources by providing high resolution where necessary and low resolution elsewhere. TINs also increase required effort in model setup, parameter estimation, and coupling with forcing data which are often gridded. This presentation discusses the costs and benefits of the use of TINs compared to grid-based methods, in the context of large watershed simulations within the traditional gridded WRF-HYDRO framework and the new TIN-based ADHydro high performance computing watershed simulator.
To increase controllability of a large flexible antenna by modal optimization
NASA Astrophysics Data System (ADS)
Wang, Feng; Wang, Pengpeng; Jiang, Wenjian
2017-12-01
Large deployable antennas are widely used in aerospace engineering to meet the envelop limit of rocket fairing. The high flexibility and low damping of antenna has proposed critical requirement not only for stability control of the antenna itself, but also for attitude control of the satellite. This paper aims to increase controllability of a large flexible antenna by modal optimization. Firstly, Sensitivity analysis of antenna modal frequencies to stiffness of support structure and stiffness of scanning mechanism are conducted respectively. Secondly, Modal simulation results of antenna frequencies are given, influences of scanning angles on moment of inertia and modal frequencies are evaluated, and modal test is carried out to validate the simulation results. All the simulation and test results show that, after modal optimization the modal characteristic of the large deployable antenna meets the controllability requirement well.
NASA Technical Reports Server (NTRS)
Griffin, Roy N., Jr.; Holzhauser, Curt A.; Weiberg, James A.
1958-01-01
An investigation was made to determine the lifting effectiveness and flow requirements of blowing over the trailing-edge flaps and ailerons on a large-scale model of a twin-engine, propeller-driven airplane having a high-aspect-ratio, thick, straight wing. With sufficient blowing jet momentum to prevent flow separation on the flap, the lift increment increased for flap deflections up to 80 deg (the maximum tested). This lift increment also increased with increasing propeller thrust coefficient. The blowing jet momentum coefficient required for attached flow on the flaps was not significantly affected by thrust coefficient, angle of attack, or blowing nozzle height.
Efficient Bayesian mixed model analysis increases association power in large cohorts
Loh, Po-Ru; Tucker, George; Bulik-Sullivan, Brendan K; Vilhjálmsson, Bjarni J; Finucane, Hilary K; Salem, Rany M; Chasman, Daniel I; Ridker, Paul M; Neale, Benjamin M; Berger, Bonnie; Patterson, Nick; Price, Alkes L
2014-01-01
Linear mixed models are a powerful statistical tool for identifying genetic associations and avoiding confounding. However, existing methods are computationally intractable in large cohorts, and may not optimize power. All existing methods require time cost O(MN2) (where N = #samples and M = #SNPs) and implicitly assume an infinitesimal genetic architecture in which effect sizes are normally distributed, which can limit power. Here, we present a far more efficient mixed model association method, BOLT-LMM, which requires only a small number of O(MN)-time iterations and increases power by modeling more realistic, non-infinitesimal genetic architectures via a Bayesian mixture prior on marker effect sizes. We applied BOLT-LMM to nine quantitative traits in 23,294 samples from the Women’s Genome Health Study (WGHS) and observed significant increases in power, consistent with simulations. Theory and simulations show that the boost in power increases with cohort size, making BOLT-LMM appealing for GWAS in large cohorts. PMID:25642633
NASA Technical Reports Server (NTRS)
Hollenbach, D. (Editor)
1983-01-01
The scientific rationale for the large deployable reflector (LDR) and the overall technological requirements are discussed. The main scientific objectives include studies of the origins of planets, stars and galaxies, and of the ultimate fate of the universe. The envisioned studies require a telescope with a diameter of at least 20 m, diffraction-limited to wavelengths as short as 30-50 micron. In addition, light-bucket operation with 1 arcsec spatial resolution in the 2-4 microns wavelength region would be useful in studies of high-redshifted galaxies. Such a telescope would provide a large increase in spectroscopic sensitivity and spatial resolving power compared with existing or planned infrared telescopes.
A Novel Concept for a Deformable Membrane Mirror for Correction of Large Amplitude Aberrations
NASA Technical Reports Server (NTRS)
Moore, Jim; Patrick, Brian
2006-01-01
Very large, light weight mirrors are being developed for applications in space. Due to launch mass and volume restrictions these mirrors will need to be much more flexible than traditional optics. The use of primary mirrors with these characteristics will lead to requirements for adaptive optics capable of correcting wave front errors with large amplitude relatively low spatial frequency aberrations. The use of low modulus membrane mirrors actuated with electrostatic attraction forces is a potential solution for this application. Several different electrostatic membrane mirrors are now available commercially. However, as the dynamic range requirement of the adaptive mirror is increased the separation distance between the membrane and the electrodes must increase to accommodate the required face sheet deformations. The actuation force applied to the mirror decreases inversely proportional to the square of the separation distance; thus for large dynamic ranges the voltage requirement can rapidly increase into the high voltage regime. Experimentation with mirrors operating in the KV range has shown that at the higher voltages a serious problem with electrostatic field cross coupling between actuators can occur. Voltage changes on individual actuators affect the voltage of other actuators making the system very difficult to control. A novel solution has been proposed that combines high voltage electrodes with mechanical actuation to overcome this problem. In this design an array of electrodes are mounted to a backing structure via light weight large dynamic range flextensional actuators. With this design the control input becomes the separation distance between the electrode and the mirror. The voltage on each of the actuators is set to a uniform relatively high voltage, thus the problem of cross talk between actuators is avoided and the favorable distributed load characteristic of electrostatic actuation is retained. Initial testing and modeling of this concept demonstrates that this is an attractive concept for increasing the dynamic range capability of electrostatic deformable mirrors.
NASA Technical Reports Server (NTRS)
1976-01-01
The six themes identified by the Workshop have many common navigation guidance and control needs. All the earth orbit themes have a strong requirement for attitude, figure and stabilization control of large space structures, a requirement not currently being supported. All but the space transportation theme have need for precision pointing of spacecraft and instruments. In addition all the themes have requirements for increasing autonomous operations for such activities as spacecraft and experiment operations, onboard mission modification, rendezvous and docking, spacecraft assembly and maintenance, navigation and guidance, and self-checkout, test and repair. Major new efforts are required to conceptualize new approaches to large space antennas and arrays that are lightweight, readily deployable, and capable of precise attitude and figure control. Conventional approaches offer little hope of meeting these requirements. Functions that can benefit from increasing automation or autonomous operations are listed.
Big questions, big science: meeting the challenges of global ecology
David Schimel; Michael Keller
2015-01-01
Ecologists are increasingly tackling questions that require significant infrastucture, large experiments, networks of observations, and complex data and computation. Key hypotheses in ecology increasingly require more investment, and larger data sets to be tested than can be collected by a single investigatorâs or s group of investigatorâs labs, sustained for longer...
A Segmented Ion-Propulsion Engine
NASA Technical Reports Server (NTRS)
Brophy, John R.
1992-01-01
New design approach for high-power (100-kW class or greater) ion engines conceptually divides single engine into combination of smaller discharge chambers integrated to operate as single large engine. Analogous to multicylinder automobile engine, benefits include reduction in required accelerator system span-to-gap ratio for large-area engines, reduction in required hollow-cathode emission current, mitigation of plasma-uniformity problem, increased tolerance to accelerator system faults, and reduction in vacuum-system pumping speed.
Post-Attack Economic Stabilization Issues for Federal, State, and Local Governments
1985-02-01
workers being transfered from large urban areas to production facilities in areas of lower risk . In another case, rent control staff should be quickly...food supermarkets , which do not universally accept bank cards. 3 0 A requirement will still exist for a large number of credit cards. While there is some...separate system is required for rationing. For example, the increasingly popular automatic teller machine ( ATM ) debit card routinely accesses both a
A large high vacuum, high pumping speed space simulation chamber for electric propulsion
NASA Technical Reports Server (NTRS)
Grisnik, Stanley P.; Parkes, James E.
1994-01-01
Testing high power electric propulsion devices poses unique requirements on space simulation facilities. Very high pumping speeds are required to maintain high vacuum levels while handling large volumes of exhaust products. These pumping speeds are significantly higher than those available in most existing vacuum facilities. There is also a requirement for relatively large vacuum chamber dimensions to minimize facility wall/thruster plume interactions and to accommodate far field plume diagnostic measurements. A 4.57 m (15 ft) diameter by 19.2 m (63 ft) long vacuum chamber at NASA Lewis Research Center is described. The chamber utilizes oil diffusion pumps in combination with cryopanels to achieve high vacuum pumping speeds at high vacuum levels. The facility is computer controlled for all phases of operation from start-up, through testing, to shutdown. The computer control system increases the utilization of the facility and reduces the manpower requirements needed for facility operations.
ERIC Educational Resources Information Center
Wolf, Erika J.; Harrington, Kelly M.; Clark, Shaunna L.; Miller, Mark W.
2013-01-01
Determining sample size requirements for structural equation modeling (SEM) is a challenge often faced by investigators, peer reviewers, and grant writers. Recent years have seen a large increase in SEMs in the behavioral science literature, but consideration of sample size requirements for applied SEMs often relies on outdated rules-of-thumb.…
Innovative Double Bypass Engine for Increased Performance
NASA Astrophysics Data System (ADS)
Manoharan, Sanjivan
Engines continue to grow in size to meet the current thrust requirements of the civil aerospace industry. Large engines pose significant transportation problems and require them to be split in order to be shipped. Thus, large amounts of time have been spent in researching methods to increase thrust capabilities while maintaining a reasonable engine size. Unfortunately, much of this research has been focused on increasing the performance and efficiencies of individual components while limited research has been done on innovative engine configurations. This thesis focuses on an innovative engine configuration, the High Double Bypass Engine, aimed at increasing fuel efficiency and thrust while maintaining a competitive fan diameter and engine length. The 1-D analysis was done in Excel and then compared to the results from Numerical Propulsion Simulation System (NPSS) software and were found to be within 4% error. Flow performance characteristics were also determined and validated against their criteria.
Breeding and Genetics Symposium: really big data: processing and analysis of very large data sets.
Cole, J B; Newman, S; Foertter, F; Aguilar, I; Coffey, M
2012-03-01
Modern animal breeding data sets are large and getting larger, due in part to recent availability of high-density SNP arrays and cheap sequencing technology. High-performance computing methods for efficient data warehousing and analysis are under development. Financial and security considerations are important when using shared clusters. Sound software engineering practices are needed, and it is better to use existing solutions when possible. Storage requirements for genotypes are modest, although full-sequence data will require greater storage capacity. Storage requirements for intermediate and results files for genetic evaluations are much greater, particularly when multiple runs must be stored for research and validation studies. The greatest gains in accuracy from genomic selection have been realized for traits of low heritability, and there is increasing interest in new health and management traits. The collection of sufficient phenotypes to produce accurate evaluations may take many years, and high-reliability proofs for older bulls are needed to estimate marker effects. Data mining algorithms applied to large data sets may help identify unexpected relationships in the data, and improved visualization tools will provide insights. Genomic selection using large data requires a lot of computing power, particularly when large fractions of the population are genotyped. Theoretical improvements have made possible the inversion of large numerator relationship matrices, permitted the solving of large systems of equations, and produced fast algorithms for variance component estimation. Recent work shows that single-step approaches combining BLUP with a genomic relationship (G) matrix have similar computational requirements to traditional BLUP, and the limiting factor is the construction and inversion of G for many genotypes. A naïve algorithm for creating G for 14,000 individuals required almost 24 h to run, but custom libraries and parallel computing reduced that to 15 m. Large data sets also create challenges for the delivery of genetic evaluations that must be overcome in a way that does not disrupt the transition from conventional to genomic evaluations. Processing time is important, especially as real-time systems for on-farm decisions are developed. The ultimate value of these systems is to decrease time-to-results in research, increase accuracy in genomic evaluations, and accelerate rates of genetic improvement.
Really big data: Processing and analysis of large datasets
USDA-ARS?s Scientific Manuscript database
Modern animal breeding datasets are large and getting larger, due in part to the recent availability of DNA data for many animals. Computational methods for efficiently storing and analyzing those data are under development. The amount of storage space required for such datasets is increasing rapidl...
Improved microseismic event locations through large-N arrays and wave-equation imaging and inversion
NASA Astrophysics Data System (ADS)
Witten, B.; Shragge, J. C.
2016-12-01
The recent increased focus on small-scale seismicity, Mw < 4 has come about primarily for two reasons. First, there is an increase in induced seismicity related to injection operations primarily for wastewater disposal and hydraulic fracturing for oil and gas recovery and for geothermal energy production. While the seismicity associated with injection is sometimes felt, it is more often weak. Some weak events are detected on current sparse arrays; however, accurate location of the events often requires a larger number of (multi-component) sensors. This leads to the second reason for an increased focus on small magnitude seismicity: a greater number of seismometers are being deployed in large N-arrays. The greater number of sensors decreases the detection threshold and therefore significantly increases the number of weak events found. Overall, these two factors bring new challenges and opportunities. Many standard seismological location and inversion techniques are geared toward large, easily identifiable events recorded on a sparse number of stations. However, with large-N arrays we can detect small events by utilizing multi-trace processing techniques, and increased processing power equips us with tools that employ more complete physics for simultaneously locating events and inverting for P- and S-wave velocity structure. We present a method that uses large-N arrays and wave-equation-based imaging and inversion to jointly locate earthquakes and estimate the elastic velocities of the earth. The technique requires no picking and is thus suitable for weak events. We validate the methodology through synthetic and field data examples.
DOT National Transportation Integrated Search
2016-12-01
Geospatial data have been playing an increasingly important role in disaster response and recovery. For large-scale natural disasters such as Hurricanes which often have the capacity to topple a large region within a span of a few days, disaster prep...
Impact of rainfall on the moisture content of large woody fuels
Helen H. Mohr; Thomas A. Waldrop
2013-01-01
This unreplicated case study evaluates the impact of rainfall on large woody fuels over time. We know that one rainfall event may decrease the Keetch-Byram Drought Index, but this study shows no real increase in fuel moisture in 1,000- hour fuels after just one rainfall. Several rain events over time are required for the moisture content of large woody fuels to...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lerer, L.B.; Scudder, T.
1999-03-01
Large dams have been criticized because of their negative environmental and social impacts. Public health interest largely has focused on vector-borne diseases, such as schistosomiasis, associated with reservoirs and irrigation projects. Large dams also influence health through changes in water and food security, increases in communicable diseases, and the social disruption caused by construction and involuntary resettlement. Communities living in close proximity to large dams often do not benefit from water transfer and electricity generation revenues. A comprehensive health component is required in environmental and social impact assessments for large dam projects.
A comparison of quality and utilization problems in large and small group practices.
Gleason, S C; Richards, M J; Quinnell, J E
1995-12-01
Physicians practicing in large, multispecialty medical groups share an organizational culture that differs from that of physicians in small or independent practices. Since 1980, there has been a sharp increase in the size of multispecialty group practice organizations, in part because of increased efficiencies of large group practices. The greater number of physicians and support personnel in a large group practice also requires a relatively more sophisticated management structure. The efficiencies, conveniences, and management structure of a large group practice provide an optimal environment to practice medicine. However, a search of the literature found no data linking a large group practice environment to practice outcomes. The purpose of the study reported in this article was to determine if physicians in large practices have fewer quality and utilization problems than physicians in small or independent practices.
NASA Technical Reports Server (NTRS)
Farner, Bruce
2013-01-01
A moveable valve for controlling flow of a pressurized working fluid was designed. This valve consists of a hollow, moveable floating piston pressed against a stationary solid seat, and can use the working fluid to seal the valve. This open/closed, novel valve is able to use metal-to-metal seats, without requiring seat sliding action; therefore there are no associated damaging effects. During use, existing standard high-pressure ball valve seats tend to become damaged during rotation of the ball. Additionally, forces acting on the ball and stem create large amounts of friction. The combination of these effects can lead to system failure. In an attempt to reduce damaging effects and seat failures, soft seats in the ball valve have been eliminated; however, the sliding action of the ball across the highly loaded seat still tends to scratch the seat, causing failure. Also, in order to operate, ball valves require the use of large actuators. Positioning the metal-to-metal seats requires more loading, which tends to increase the size of the required actuator, and can also lead to other failures in other areas such as the stem and bearing mechanisms, thus increasing cost and maintenance. This novel non-sliding seat surface valve allows metal-to-metal seats without the damaging effects that can lead to failure, and enables large seating forces without damaging the valve. Additionally, this valve design, even when used with large, high-pressure applications, does not require large conventional valve actuators and the valve stem itself is eliminated. Actuation is achieved with the use of a small, simple solenoid valve. This design also eliminates the need for many seals used with existing ball valve and globe valve designs, which commonly cause failure, too. This, coupled with the elimination of the valve stem and conventional valve actuator, improves valve reliability and seat life. Other mechanical liftoff seats have been designed; however, they have only resulted in increased cost, and incurred other reliability issues. With this novel design, the seat is lifted by simply removing the working fluid pressure that presses it against the seat and no external force is required. By eliminating variables associated with existing ball and globe configurations that can have damaging effects upon a valve, this novel design reduces downtime in rocket engine test schedules and maintenance costs.
Federal Register 2010, 2011, 2012, 2013, 2014
2011-09-15
... DEPARTMENT OF HEALTH AND HUMAN SERVICES Announcement of Requirements and Registration for... Coordinator for Health Information Technology, HHS. ACTION: Notice. SUMMARY: Medical devices will play an increasingly large role in the monitoring and collection of patient data with the spread of electronic health...
NASA Astrophysics Data System (ADS)
Fader, Marianela; Shi, Sinan; von Bloh, Werner; Bondeau, Alberte; Cramer, Wolfgang
2017-04-01
Irrigation in the Mediterranean is of vital importance for food security, employment and economic development. We will present a recently published study1 that estimates the current level of water demand for Mediterranean agriculture and simulates the potential impacts of climate change, population growth and transitions to water-saving irrigation and conveyance technologies. The results indicate that, at present, Mediterranean region could save 35% of water by implementing more efficient irrigation and conveyance systems, with large differences in the saving potentials across countries. Under climate change, more efficient irrigation is of vital importance for counteracting increases in irrigation water requirements. The Mediterranean area as a whole might face an increase in gross irrigation requirements between 4% and 18% from climate change alone by the end of the century if irrigation systems and conveyance are not improved. Population growth increases these numbers to 22% and 74%, respectively, affecting mainly the Southern and Eastern Mediterranean. However, improved irrigation technologies and conveyance systems have large water saving potentials, especially in the Eastern Mediterranean. Both the Eastern and the Southern Mediterranean would need around 35% more water than today if they could afford some degree of modernization of irrigation and conveyance systems and benefit from the CO2-fertilization effect. However, in some scenarios water scarcity may constrain the supply of the irrigation water needed in future in Algeria, Libya, Israel, Jordan, Lebanon, Syria, Serbia, Morocco, Tunisia and Spain. In this study, vegetation growth, phenology, agricultural production and irrigation water requirements and withdrawal were simulated with the process-based ecohydrological and agro-ecosystem model LPJmL ("Lund-Potsdam-Jena managed Land") after a large development2 that comprised the improved representation of Mediterranean crops.
Supporting large scale applications on networks of workstations
NASA Technical Reports Server (NTRS)
Cooper, Robert; Birman, Kenneth P.
1989-01-01
Distributed applications on networks of workstations are an increasingly common way to satisfy computing needs. However, existing mechanisms for distributed programming exhibit poor performance and reliability as application size increases. Extension of the ISIS distributed programming system to support large scale distributed applications by providing hierarchical process groups is discussed. Incorporation of hierarchy in the program structure and exploitation of this to limit the communication and storage required in any one component of the distributed system is examined.
NASA Technical Reports Server (NTRS)
1975-01-01
Energy consumption in the United States has risen in response to both increasing population and to increasing levels of affluence. Depletion of domestic energy reserves requires consumption modulation, production of fossil fuels, more efficient conversion techniques, and large scale transitions to non-fossile fuel energy sources. Widening disparity between the wealthy and poor nations of the world contributes to trends that increase the likelihood of group action by the lesser developed countries to achieve political and economic goals. The formation of anticartel cartels is envisioned.
Noise-Reduction Benefits Analyzed for Over-the-Wing-Mounted Advanced Turbofan Engines
NASA Technical Reports Server (NTRS)
Berton, Jeffrey J.
2000-01-01
As we look to the future, increasingly stringent civilian aviation noise regulations will require the design and manufacture of extremely quiet commercial aircraft. Also, the large fan diameters of modern engines with increasingly higher bypass ratios pose significant packaging and aircraft installation challenges. One design approach that addresses both of these challenges is to mount the engines above the wing. In addition to allowing the performance trend towards large diameters and high bypass ratio cycles to continue, this approach allows the wing to shield much of the engine noise from people on the ground. The Propulsion Systems Analysis Office at the NASA Glenn Research Center at Lewis Field conducted independent analytical research to estimate the noise reduction potential of mounting advanced turbofan engines above the wing. Certification noise predictions were made for a notional long-haul commercial quadjet transport. A large quad was chosen because, even under current regulations, such aircraft sometimes experience difficulty in complying with certification noise requirements with a substantial margin. Also, because of its long wing chords, a large airplane would receive the greatest advantage of any noise-shielding benefit.
Paterson, R Russell M; Lima, Nelson
2018-01-01
Palm oil is used in various valued commodities and is a large global industry worth over US$ 50 billion annually. Oil palms (OP) are grown commercially in Indonesia and Malaysia and other countries within Latin America and Africa. The large-scale land-use change has high ecological, economic, and social impacts. Tropical countries in particular are affected negatively by climate change (CC) which also has a detrimental impact on OP agronomy, whereas the cultivation of OP increases CC. Amelioration of both is required. The reduced ability to grow OP will reduce CC, which may allow more cultivation tending to increase CC, in a decreasing cycle. OP could be increasingly grown in more suitable regions occurring under CC. Enhancing the soil fauna may compensate for the effect of CC on OP agriculture to some extent. The effect of OP cultivation on CC may be reduced by employing reduced emissions from deforestation and forest degradation plans, for example, by avoiding illegal fire land clearing. Other ameliorating methods are reported herein. More research is required involving good management practices that can offset the increases in CC by OP plantations. Overall, OP-growing countries should support the Paris convention on reducing CC as the most feasible scheme for reducing CC.
NASA Astrophysics Data System (ADS)
Li, Yan; Ai, Qinghui; Mai, Kangsen; Xu, Wei; Cheng, Zhenyan; He, Zhigang
2010-12-01
Dietary leucine requirement for juvenile large yellow croaker, Pseudosciaena crocea Richardson 1846 (initial body weight 6.0 g ± 0.1 g) was determined using dose-response method. Six isonitogenous (crude protein 43%) and isoenergetic (19 kJ g-1) practical diets containing six levels of leucine (Diets 1-6) ranging from 1.23% to 4.80% (dry matter) were made at about 0.7% increment of leucine. Equal amino acid nitrogen was maintained by replacing leucine with glutamic acid. Triplicate groups of 60 individuals were fed to apparent satiation by hand twice daily (05:00 and 17:30). The water temperature was 26-32°C, salinity 26-30 and dissolved oxygen approximately 7 mg L-1 during the experimental period. Final weight (FW) of large yellow croaker initially increased with increasing level of dietary leucine but then decreased at further higher level of leucine. The highest FW was obtained in fish fed diet with 3.30% Leucine (Diet 4). FW of fish fed the diet with 4.80% Leucine (Diet 6) was significantly lower than those fed Diet 4. However, no significant differences were observed between the other dietary treatments. Feed efficiency (FE) and whole body composition were independent of dietary leucine contents ( P > 0.05). The results indicated that leucine was essential for growth of juvenile large yellow croaker. On the basis of FW, the optimum dietary leucine requirement for juvenile large yellow croaker was estimated to be 2.92% of dry matter (6.79% of dietary protein).
2014-01-01
reaching a peak of $212M (constant 2010 dollars) over the baseline language program in 2010. Needless to say, neither this increase nor the increase...committed on a large scale, as has occurred with Afghanistan and Iraq. This figure assumes peaks and valleys in requirements as wars begin and wind...Support • SRM • Depot Maintenance • Reset • Service-Wide Transportation • Central Supply Activities • Logistic Support Activities • Ammunition
Agronomic and environmental consequences of using liquid mineral concentrates on arable farms.
Schils, René L M; Postma, Romke; van Rotterdam, Debby; Zwart, Kor B
2015-12-01
In regions with intensive livestock systems, the processing of manure into liquid mineral concentrates is seen as an option to increase the nutrient use efficiency of manures. The agricultural sector anticipates that these products may in future be regarded as regular mineral fertilisers. We assessed the agronomic suitability and impact on greenhouse gas (GHG) and ammonia emissions of using liquid mineral concentrates on arable farms. The phosphate requirements on arable farms were largely met by raw pig slurry, given its large regional availability. After the initial nutrient input by means of pig slurry, the nitrogen/phosphate ratio of the remaining nutrient crop requirements determined the additional amount of liquid mineral concentrates that can be used. For sandy soils, liquid mineral concentrates could supply 50% of the nitrogen requirement, whereas for clay soils the concentrates did not meet the required nitrogen/phosphate ratio. The total GHG emissions per kg of plant available nitrogen ranged from -65 to 33 kg CO2 -equivalents. It increased in the order digestates < mineral fertiliser < raw slurries. Liquid mineral concentrates had limited added value for arable farms. For an increased suitability it is necessary that liquid mineral concentrates do not contain phosphate and that the nitrogen availability is increased. In the manure-processing chain, anaerobic digestion had a dominant and beneficial effect on GHG emissions. © 2015 Society of Chemical Industry.
Large-scale seismic signal analysis with Hadoop
Addair, T. G.; Dodge, D. A.; Walter, W. R.; ...
2014-02-11
In seismology, waveform cross correlation has been used for years to produce high-precision hypocenter locations and for sensitive detectors. Because correlated seismograms generally are found only at small hypocenter separation distances, correlation detectors have historically been reserved for spotlight purposes. However, many regions have been found to produce large numbers of correlated seismograms, and there is growing interest in building next-generation pipelines that employ correlation as a core part of their operation. In an effort to better understand the distribution and behavior of correlated seismic events, we have cross correlated a global dataset consisting of over 300 million seismograms. Thismore » was done using a conventional distributed cluster, and required 42 days. In anticipation of processing much larger datasets, we have re-architected the system to run as a series of MapReduce jobs on a Hadoop cluster. In doing so we achieved a factor of 19 performance increase on a test dataset. We found that fundamental algorithmic transformations were required to achieve the maximum performance increase. Whereas in the original IO-bound implementation, we went to great lengths to minimize IO, in the Hadoop implementation where IO is cheap, we were able to greatly increase the parallelism of our algorithms by performing a tiered series of very fine-grained (highly parallelizable) transformations on the data. Each of these MapReduce jobs required reading and writing large amounts of data.« less
Large-scale seismic signal analysis with Hadoop
DOE Office of Scientific and Technical Information (OSTI.GOV)
Addair, T. G.; Dodge, D. A.; Walter, W. R.
In seismology, waveform cross correlation has been used for years to produce high-precision hypocenter locations and for sensitive detectors. Because correlated seismograms generally are found only at small hypocenter separation distances, correlation detectors have historically been reserved for spotlight purposes. However, many regions have been found to produce large numbers of correlated seismograms, and there is growing interest in building next-generation pipelines that employ correlation as a core part of their operation. In an effort to better understand the distribution and behavior of correlated seismic events, we have cross correlated a global dataset consisting of over 300 million seismograms. Thismore » was done using a conventional distributed cluster, and required 42 days. In anticipation of processing much larger datasets, we have re-architected the system to run as a series of MapReduce jobs on a Hadoop cluster. In doing so we achieved a factor of 19 performance increase on a test dataset. We found that fundamental algorithmic transformations were required to achieve the maximum performance increase. Whereas in the original IO-bound implementation, we went to great lengths to minimize IO, in the Hadoop implementation where IO is cheap, we were able to greatly increase the parallelism of our algorithms by performing a tiered series of very fine-grained (highly parallelizable) transformations on the data. Each of these MapReduce jobs required reading and writing large amounts of data.« less
Large historical eruptions at subaerial mud volcanoes, Italy
NASA Astrophysics Data System (ADS)
Manga, M.; Bonini, M.
2012-11-01
Active mud volcanoes in the northern Apennines, Italy, currently have gentle eruptions. There are, however, historical accounts of violent eruptions and outbursts. Evidence for large past eruptions is also recorded by large decimeter rock clasts preserved in erupted mud. We measured the rheological properties of mud currently being erupted in order to evaluate the conditions needed to transport such large clasts to the surface. The mud is well-characterized by the Herschel-Bulkley model, with yield stresses between 4 and 8 Pa. Yield stresses of this magnitude can support the weight of particles with diameters up to several mm. At present, particles larger than this size are not being carried to the surface. The transport of larger clasts to the surface requires ascent speeds greater than their settling speed in the mud. We use a model for the settling of particles and rheological parameters from laboratory measurements to show that the eruption of large clasts requires ascent velocities > 1 m s-1, at least three orders of magnitude greater than during the present, comparatively quiescent, activity. After regional earthquakes on 20 May and 29 May 2012, discharge also increased at locations where the stress changes produced by the earthquakes would have unclamped feeder dikes below the mud volcanoes. The magnitude of increased discharge, however, is less than that inferred from the large clasts. Both historical accounts and erupted deposits are consistent in recording episodic large eruptions.
Clark, Andrew G; Naufer, M Nabuan; Westerlund, Fredrik; Lincoln, Per; Rouzina, Ioulia; Paramanathan, Thayaparan; Williams, Mark C
2018-02-06
Molecules that bind DNA via threading intercalation show high binding affinity as well as slow dissociation kinetics, properties ideal for the development of anticancer drugs. To this end, it is critical to identify the specific molecular characteristics of threading intercalators that result in optimal DNA interactions. Using single-molecule techniques, we quantify the binding of a small metal-organic ruthenium threading intercalator (Δ,Δ-B) and compare its binding characteristics to a similar molecule with significantly larger threading moieties (Δ,Δ-P). The binding affinities of the two molecules are the same, while comparison of the binding kinetics reveals significantly faster kinetics for Δ,Δ-B. However, the kinetics is still much slower than that observed for conventional intercalators. Comparison of the two threading intercalators shows that the binding affinity is modulated independently by the intercalating section and the binding kinetics is modulated by the threading moiety. In order to thread DNA, Δ,Δ-P requires a "lock mechanism", in which a large length increase of the DNA duplex is required for both association and dissociation. In contrast, measurements of the force-dependent binding kinetics show that Δ,Δ-B requires a large DNA length increase for association but no length increase for dissociation from DNA. This contrasts strongly with conventional intercalators, for which almost no DNA length change is required for association but a large DNA length change must occur for dissociation. This result illustrates the fundamentally different mechanism of threading intercalation compared with conventional intercalation and will pave the way for the rational design of therapeutic drugs based on DNA threading intercalation.
Measuring discharge with ADCPs: Inferences from synthetic velocity profiles
Rehmann, C.R.; Mueller, D.S.; Oberg, K.A.
2009-01-01
Synthetic velocity profiles are used to determine guidelines for sampling discharge with acoustic Doppler current profilers (ADCPs). The analysis allows the effects of instrument characteristics, sampling parameters, and properties of the flow to be studied systematically. For mid-section measurements, the averaging time required for a single profile measurement always exceeded the 40 s usually recommended for velocity measurements, and it increased with increasing sample interval and increasing time scale of the large eddies. Similarly, simulations of transect measurements show that discharge error decreases as the number of large eddies sampled increases. The simulations allow sampling criteria that account for the physics of the flow to be developed. ?? 2009 ASCE.
Potential benefits of a ceramic thermal barrier coating on large power generation gas turbine
NASA Technical Reports Server (NTRS)
Clark, J. S.; Nainiger, J. J.
1977-01-01
Thermal barrier coating design option offers benefit in terms of reduced electricity costs when used in utility gas turbines. Options considered include: increased firing temperature, increased component life, reduced cooling air requirements, and increased corrosion resistance (resulting in increased tolerance for dirty fuels). Performance and cost data were obtained. Simple, recuperated and combined cycle applications were considered, and distillate and residual fuels were assumed. The results indicate that thermal barrier coatings could produce large electricity cost savings if these coatings permit turbine operation with residual fuels at distillate-rated firing temperatures. The results also show that increased turbine inlet temperature can result in substantial savings in fuel and capital costs.
NASA Astrophysics Data System (ADS)
Devilliers, C.; Du Jeu, C.; Costes, V.; Suau, A.; Girault, N.; Cornillon, L.
2017-11-01
Space telescopes pupil diameter increases continuously to reach higher resolutions and associated optical scheme become more sensitive. As a consequence the size of these telescopes but also their stability requirements increase. Therefore, mass of space telescopes becomes a strong design driver to be still compatible with price competitive launcher capabilities. Moreover satellite agility requirements are more and more severe and instruments shall be compatible with quick evolution of thermal environment.
Perspective on the span-distributed-load concept for application to large cargo aircraft design
NASA Technical Reports Server (NTRS)
Whitehead, A. H., Jr.
1975-01-01
Results of a simplified analysis of the span-distributed-load concept (in which payload is placed within the wing structure) are presented. It is shown that a design based on these principles has a high potential for application to future large air cargo transport. Significant improvements are foreseen in increased payload fraction and productivity and in reduced fuel consumption and operating costs. A review of the efforts in the 1940's to develop all-wing aircraft shows the potential of transferring those early technological developments to current design of distributed-load aircraft. Current market analyses are projected to 1990 to show the future commercial demand for large capacity freighters. Several configuration designs which would serve different market requirements for these large freighters are discussed as are some of the pacing-technology requirements.
The potential benefits of photonics in the computing platform
NASA Astrophysics Data System (ADS)
Bautista, Jerry
2005-03-01
The increase in computational requirements for real-time image processing, complex computational fluid dynamics, very large scale data mining in the health industry/Internet, and predictive models for financial markets are driving computer architects to consider new paradigms that rely upon very high speed interconnects within and between computing elements. Further challenges result from reduced power requirements, reduced transmission latency, and greater interconnect density. Optical interconnects may solve many of these problems with the added benefit extended reach. In addition, photonic interconnects provide relative EMI immunity which is becoming an increasing issue with a greater dependence on wireless connectivity. However, to be truly functional, the optical interconnect mesh should be able to support arbitration, addressing, etc. completely in the optical domain with a BER that is more stringent than "traditional" communication requirements. Outlined are challenges in the advanced computing environment, some possible optical architectures and relevant platform technologies, as well roughly sizing these opportunities which are quite large relative to the more "traditional" optical markets.
How Big Is Big Enough? Sample Size Requirements for CAST Item Parameter Estimation
ERIC Educational Resources Information Center
Chuah, Siang Chee; Drasgow, Fritz; Luecht, Richard
2006-01-01
Adaptive tests offer the advantages of reduced test length and increased accuracy in ability estimation. However, adaptive tests require large pools of precalibrated items. This study looks at the development of an item pool for 1 type of adaptive administration: the computer-adaptive sequential test. An important issue is the sample size required…
Some thoughts concerning large load-carrying vehicles
NASA Technical Reports Server (NTRS)
Spearman, M. L.
1983-01-01
Some implications relative to combat operations and force sustainability into the twenty-first century are discussed. The basic conjecture is that, sometime in the future, secure overseas basing may be denied to the United States by the Soviet Union or by unfriendly, unstable governments. In that event, the support of future battle itself, may be conducted from the continental U.S. and would introduce requirements for large, long-range, efficient, and sometimes, fast air vehicles. Some unusual design concepts and the technology requirements for such vehicles are suggested. It is concluded that, while much of the required technology is already being pursued, further advanced should be expected and sought in improved aerodynamics, propulsion, structures, and avionics with a view toward increased efficiency, utility, and affordability.
Hydraulic hoisting and backfilling
NASA Astrophysics Data System (ADS)
Sauermann, H. B.
In a country such as South Africa, with its large deep level mining industry, improvements in mining and hoisting techniques could result in substantial savings. Hoisting techniques, for example, may be improved by the introduction of hydraulic hoisting. The following are some of the advantages of hydraulic hoisting as against conventional skip hoisting: (1) smaller shafts are required because the pipes to hoist the same quantity of ore hydraulically require less space in the shaft than does skip hoisting equipment; (2) the hoisting capacity of a mine can easily be increased without the necessity of sinking new shafts. Large savings in capital costs can thus be made; (3) fully automatic control is possible with hydraulic hoisting and therefore less manpower is required; and (4) health and safety conditions will be improved.
Thermal/structural design verification strategies for large space structures
NASA Technical Reports Server (NTRS)
Benton, David
1988-01-01
Requirements for space structures of increasing size, complexity, and precision have engendered a search for thermal design verification methods that do not impose unreasonable costs, that fit within the capabilities of existing facilities, and that still adequately reduce technical risk. This requires a combination of analytical and testing methods. This requires two approaches. The first is to limit thermal testing to sub-elements of the total system only in a compact configuration (i.e., not fully deployed). The second approach is to use a simplified environment to correlate analytical models with test results. These models can then be used to predict flight performance. In practice, a combination of these approaches is needed to verify the thermal/structural design of future very large space systems.
Optimizing the scale of markets for water quality trading
NASA Astrophysics Data System (ADS)
Doyle, Martin W.; Patterson, Lauren A.; Chen, Yanyou; Schnier, Kurt E.; Yates, Andrew J.
2014-09-01
Applying market approaches to environmental regulations requires establishing a spatial scale for trading. Spatially large markets usually increase opportunities for abatement cost savings but increase the potential for pollution damages (hot spots), vice versa for spatially small markets. We develop a coupled hydrologic-economic modeling approach for application to point source emissions trading by a large number of sources and apply this approach to the wastewater treatment plants (WWTPs) within the watershed of the second largest estuary in the U.S. We consider two different administrative structures that govern the trade of emission permits: one-for-one trading (the number of permits required for each unit of emission is the same for every WWTP) and trading ratios (the number of permits required for each unit of emissions varies across WWTP). Results show that water quality regulators should allow trading to occur at the river basin scale as an appropriate first-step policy, as is being done in a limited number of cases via compliance associations. Larger spatial scales may be needed under conditions of increased abatement costs. The optimal scale of the market is generally the same regardless of whether one-for-one trading or trading ratios are employed.
ERIC Educational Resources Information Center
Fiedler, Klaus; Kareev, Yaakov
2006-01-01
Adaptive decision making requires that contingencies between decision options and their relative assets be assessed accurately and quickly. The present research addresses the challenging notion that contingencies may be more visible from small than from large samples of observations. An algorithmic account for such a seemingly paradoxical effect…
Kumar, Rajendra; Meis, Jeanne M; Amini, Behrang; McEnery, Kevin W; Madewell, John E; Rhines, Laurence D; Benjamin, Robert S
2017-05-15
Case report and literature review. To describe treatment of a unique case of acute airway obstruction by a large C7 giant cell tumor (GCT) with preoperative denosumab followed by surgical resection, and review the literature on this rare entity. Standard treatment for GCTs includes surgical resection or curettage and packing. Large lesions in the spine may require preoperative therapy with denosumab, a human monoclonal antibody to RANKL, to facilitate surgery. It is highly unusual for GCT arising in cervical spine to present with acute asphyxia (requiring tracheostomy). We report a patient with large C7 GCT that caused tracheal compression with almost complete airway obstruction requiring emergency intubation. The tumor responded to subcutaneously administered denosumab with marked decrease in size and relief of symptoms. Increased tumor mineralization in response to therapy facilitated subsequent successful surgical tumor resection. The patient remains symptom-free 2 years after surgery without tumor recurrence. Denosumab can shrink the size of large GCTs, providing symptom relief before surgery and facilitate tumor resection. 5.
Aversive properties of negative incentive shifts in Fischer 344 and Lewis rats
Brewer, Adam; Johnson, Patrick; Stein, Jeff; Schlund, Michael; Williams, Dean C.
2018-01-01
Research on incentive contrast highlights that reward value is not absolute but rather is based upon comparisons we make to rewards we have received and expect to receive. Both human and nonhuman studies on incentive contrast show that shifting from a larger more-valued reward to a smaller less-valued reward is associated with long periods of nonresponding—a negative contrast effect. In this investigation, we used two different genetic rat strains, Fischer 344 and Lewis rats that putatively differ in their sensitivity to aversive stimulation, to assess the aversive properties of large-to-small reward shifts (negative incentive shifts). Additionally, we examined the extent to which increasing cost (fixed-ratio requirements) modulates negative contrast effects. In the presence of a cue that signaled the upcoming reward magnitude, lever pressing was reinforced with one of two different magnitudes of food (large or small). This design created two contrast shifts (small-to-large, large-to-small) and two shifts used as control conditions (small-to-small, large-to-large). Results showed a significant interaction between rat strain and cost requirements only during the negative incentive shift with the emotionally reactive Fischer 344 rats exhibiting significantly longer response latencies with increasing cost, highlighting greater negative contrast. These findings are more consistent with emotionality accounts of negative contrast and results of neurophysiological research that suggests shifting from a large to a small reward is aversive. Findings also highlight how subjective reward value and motivation is a product of gene-environment interactions. PMID:27864048
Large bipolarons and oxide superconductivity
NASA Astrophysics Data System (ADS)
Emin, David
2017-02-01
Large-bipolaron superconductivity is plausible with carrier densities well below those of conventional metals. Bipolarons form when carriers self-trap in pairs. Coherently moving large-bipolarons require extremely large ratios of static to optical dielectric-constants. The mutual Coulomb repulsion of a planar large-bipolaron's paired carriers drives it to a four-lobed shape. A phonon-mediated attraction among large-bipolarons propels their condensation into a liquid. This liquid's excitations move slowly with a huge effective mass. Excitations' concomitant weak scattering by phonons produces a moderate low-temperature dc resistivity that increases linearly with rising temperature. With falling temperature an energy gap opens between large-bipolarons' excitations and those of their self-trapped electronic carriers.
Implementing Parquet equations using HPX
NASA Astrophysics Data System (ADS)
Kellar, Samuel; Wagle, Bibek; Yang, Shuxiang; Tam, Ka-Ming; Kaiser, Hartmut; Moreno, Juana; Jarrell, Mark
A new C++ runtime system (HPX) enables simulations of complex systems to run more efficiently on parallel and heterogeneous systems. This increased efficiency allows for solutions to larger simulations of the parquet approximation for a system with impurities. The relevancy of the parquet equations depends upon the ability to solve systems which require long runs and large amounts of memory. These limitations, in addition to numerical complications arising from stability of the solutions, necessitate running on large distributed systems. As the computational resources trend towards the exascale and the limitations arising from computational resources vanish efficiency of large scale simulations becomes a focus. HPX facilitates efficient simulations through intelligent overlapping of computation and communication. Simulations such as the parquet equations which require the transfer of large amounts of data should benefit from HPX implementations. Supported by the the NSF EPSCoR Cooperative Agreement No. EPS-1003897 with additional support from the Louisiana Board of Regents.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fishkind, H.H.
1982-04-01
The feasibility of large-scale plantation establishment by various methods was examined, and the following conclusions were reached: seedling plantations are limited in potential yield due to genetic variation among the planting stock and often inadequate supplies of appropriate seed; vegetative propagation by rooted cuttings can provide good genetic uniformity of select hybrid planting stock; however, large-scale production requires establishment and maintenance of extensive cutting orchards. The collection of shoots and preparation of cuttings, although successfully implemented in the Congo and Brazil, would not be economically feasible in Florida for large-scale plantations; tissue culture propagation of select hybrid eucalypts offers themore » only opportunity to produce the very large number of trees required to establish the energy plantation. The cost of tissue culture propagation, although higher than seedling production, is more than off-set by the increased productivity of vegetative plantations established from select hybrid Eucalyptus.« less
Site Management and Productivity in Tropical Forest Plantations
A. Tiarks; E.K.S. Nambiar; C. Cossalter
1998-01-01
Tropical countries are expanding plantation forestry to develop sustainable woodproduction systems. Much of this is based on short rotations of exotic species. These systems require large capital investments, represent intensive land use and increase the demands on the soil. To develop options for maintaining or increasing productivity a partner-project was initiated...
Cardiotachometer displays heart rate on a beat-to-beat basis
NASA Technical Reports Server (NTRS)
Rasquin, J. R.; Smith, H. E.; Taylor, R. A.
1974-01-01
Electronics for this system may be chosen so that complete calculation and display may be accomplished in a few milliseconds, far less than even the fastest heartbeat interval. Accuracy may be increased, if desired, by using higher-frequency timing oscillator, although this will require large capacity registers at increased cost.
ERIC Educational Resources Information Center
Kharabe, Amol T.
2012-01-01
Over the last two decades, firms have operated in "increasingly" accelerated "high-velocity" dynamic markets, which require them to become "agile." During the same time frame, firms have increasingly deployed complex enterprise systems--large-scale packaged software "innovations" that integrate and automate…
Effects of Teacher Evaluation on Teacher Job Satisfaction in Ohio
ERIC Educational Resources Information Center
Downing, Pamela R.
2016-01-01
The purpose of this quantitative study was to explore whether or not increased accountability measures found in the Ohio Teacher Evaluation System (OTES) impacted teacher job satisfaction. Student growth measures required by the OTES increased teacher accountability. Today, teachers are largely evaluated based on the results of what they do in the…
ERIC Educational Resources Information Center
Young, Kathleen D.; Snelling, Anastasia; Maroto, Maya; Young, Katherine A.
2013-01-01
Purpose/Objectives: In 2010, a large urban school district implemented a district-wide school wellness policy that addressed childhood obesity by requiring schools to increase health and physical education contact hours for students and to improve the nutritional standards of school meals. Schools were required to serve a different fruit and…
Large Format Arrays for Far Infrared and Millimeter Astronomy
NASA Technical Reports Server (NTRS)
Moseley, Harvey
2004-01-01
Some of the most compelling questions in modem astronomy are best addressed with submillimeter and millimeter observations. The question of the role of inflation in the early evolution of the universe is best addressed with large sensitive arrays of millimeter polarimeters. The study of the first generations of galaxies requires sensitive submillimeter imaging, which can help us to understand the history of energy release and nucleosynthesis in the universe. Our ability to address these questions is dramatically increasing, driven by dramatic steps in the sensitivity and size of available detector arrays. While the MIPS instrument on the SIRTF mission will revolutionize far infrared astronomy with its 1024 element array of photoconductors, thermal detectors remain the dominant technology for submillimeter and millimeter imaging and polarimetry. The last decade has seen the deployment of increasingly large arrays of bolometers, ranging from the 48 element arrays deployed on the KAO in the late 198Os, to the SHARC and SCUBA arrays in the 1990s. The past years have seen the deployment of a new generation of larger detector arrays in SHARC II (384 channels) and Bolocam (144 channels). These detectors are in operation and are beginning to make significant impacts on the field. Arrays of sensitive submillimeter bolometers on the SPIRE instrument on Herschel will allow the first large areas surveys of the sky, providing important insight into the evolution of galaxies. The next generation of detectors, led by SCUBA II, will increase the focal scale of these instruments by an order of magnitude. Two major missions are being planned by NASA for which further development of long wavelength detectors is essential, The SAFlR mission, a 10-m class telescope with large arrays of background limited detectors, will extend our reach into the epoch of initial galaxy formation. A major goal of modem cosmology is to test the inflationary paradigm in the early evolution of the universe. To this end, a mission is planned to detect the imprint of inflation on the CMB by precision measurement of its polarization. This work requires very large arrays of sensitive detectors which can provide unprecedented control of a wide range of systematic errors, given the small amplitude of the signal of interest. We will describe the current state of large format detector arrays, the performance requirements set by the new missions, and the different approaches being developed in the community to meet these requirements. We are confident that within a decade, these developments will lead to dramatic advances in our understanding of the evolution of the universe.
Riley, Karin L.; Loehman, Rachel A.
2016-01-01
Climate changes are expected to increase fire frequency, fire season length, and cumulative area burned in the western United States. We focus on the potential impact of mid-21st-century climate changes on annual burn probability, fire season length, and large fire characteristics including number and size for a study area in the Northern Rocky Mountains. Although large fires are rare they account for most of the area burned in western North America, burn under extreme weather conditions, and exhibit behaviors that preclude methods of direct control. Allocation of resources, development of management plans, and assessment of fire effects on ecosystems all require an understanding of when and where fires are likely to burn, particularly under altered climate regimes that may increase large fire occurrence. We used the large fire simulation model FSim to model ignition, growth, and containment of wildfires under two climate scenarios: contemporary (based on instrumental weather) and mid-century (based on an ensemble average of global climate models driven by the A1B SRES emissions scenario). Modeled changes in fire patterns include increased annual burn probability, particularly in areas of the study region with relatively short contemporary fire return intervals; increased individual fire size and annual area burned; and fewer years without large fires. High fire danger days, represented by threshold values of Energy Release Component (ERC), are projected to increase in number, especially in spring and fall, lengthening the climatic fire season. For fire managers, ERC is an indicator of fire intensity potential and fire economics, with higher ERC thresholds often associated with larger, more expensive fires. Longer periods of elevated ERC may significantly increase the cost and complexity of fire management activities, requiring new strategies to maintain desired ecological conditions and limit fire risk. Increased fire activity (within the historical range of frequency and severity, and depending on the extent to which ecosystems are adapted) may maintain or restore ecosystem functionality; however, in areas that are highly departed from historical fire regimes or where there is disequilibrium between climate and vegetation, ecosystems may be rapidly and persistently altered by wildfires, especially those that burn under extreme conditions.
Harwood, Rowan H.; Sayer, Avan Aihie; Hirschfeld, Miriam
2004-01-01
OBJECTIVE: To estimate the number of people worldwide requiring daily assistance from another person in carrying out health, domestic or personal tasks. METHODS: Data from the Global Burden of Disease Study were used to calculate the prevalence of severe levels of disability, and consequently, to estimate dependency. Population projections were used to forecast changes over the next 50 years. FINDINGS: The greatest burden of dependency currently falls in sub-Saharan Africa, where the "dependency ratio" (ratio of dependent people to the population of working age) is about 10%, compared with 7-8% elsewhere. Large increases in prevalence are predicted in sub-Saharan Africa, the Middle East, Asia and Latin America of up to 5-fold or 6-fold in some cases. These increases will occur in the context of generally increasing populations, and dependency ratios will increase modestly to about 10%. The dependency ratio will increase more in China (14%) and India (12%) than in other areas with large prevalence increases. Established market economies, especially Europe and Japan, will experience modest increases in the prevalence of dependency (30%), and in the dependency ratio (up to 10%). Former Socialist economies of Europe will have static or declining numbers of dependent people, but will have large increases in the dependency ratio (up to 13%). CONCLUSION: Many countries will be greatly affected by the increasing number of dependent people and will need to identify the human and financial resources to support them. Much improved collection of data on disability and on the needs of caregivers is required. The prevention of disability and provision of support for caregivers needs greater priority. PMID:15259253
The costs and effectiveness of large Phase III pre-licensure vaccine clinical trials.
Black, Steven
2015-01-01
Prior to the 1980s, most vaccines were licensed based upon safety and effectiveness studies in several hundred individuals. Beginning with the evaluation of Haemophilus influenzae type b conjugate vaccines, much larger pre-licensure trials became common. The pre-licensure trial for Haemophilus influenzae oligosaccharide conjugate vaccine had more than 60,000 children and that of the seven-valent pneumococcal conjugate vaccine included almost 38,000 children. Although trial sizes for both of these studies were driven by the sample size required to demonstrate efficacy, the sample size requirements for safety evaluations of other vaccines have subsequently increased. With the demonstration of an increased risk of intussusception following the Rotashield brand rotavirus vaccine, this trend has continued. However, routinely requiring safety studies of 20,000-50,000 or more participants has two major downsides. First, the cost of performing large safety trials routinely prior to licensure of a vaccine is very large, with some estimates as high at US$200 million euros for one vaccine. This high financial cost engenders an opportunity cost whereby the number of vaccines that a company is willing or able to develop to meet public health needs becomes limited by this financial barrier. The second downside is that in the pre-licensure setting, such studies are very time consuming and delay the availability of a beneficial vaccine substantially. One might argue that in some situations, this financial commitment is warranted such as for evaluations of the risk of intussusception following newer rotavirus vaccines. However, it must be noted that while an increased risk of intussusception was not identified in large pre-licensure studies, in post marketing evaluations an increased risk of this outcome has been identified. Thus, even the extensive pre-licensure evaluations conducted did not identify an associated risk. The limitations of large Phase III trials have also been demonstrated in efficacy trials. Notably, pre-licensure trials of pneumococcal conjugate severely underestimated their true effect and cost-effectiveness. In fact, in discussions prior to vaccine introduction in the USA for PCV7, the vaccine was said to be not cost-effective and some counseled against its introduction. In reality, following introduction, PCV7 has been shown to be highly cost-effective. In the last decade, new methods have been identified using large linked databases such as the Vaccine Safety Datalink in the USA that allow identification of an increased risk of an event within a few months of vaccine introduction and that can screen for unanticipated very rare events as well. In addition, the availability of electronic medical records and hospital discharge data in many settings allows for accurate assessment of vaccine effectiveness. Given the high financial and opportunity cost of requiring large pre-licensure safety studies, consideration could be given to 'conditional licensure' of vaccines whose delivery system is well characterized in a setting where sophisticated pharmacovigilance systems exist on the condition that such licensure would incorporate a requirement for rapid cycle and other real-time evaluations of safety and effectiveness following introduction. This would actually allow for a more complete and timely evaluation of vaccines, lower the financial barrier to development of new vaccines and thus allow a broader portfolio of vaccines to be developed and successfully introduced.
NASA Astrophysics Data System (ADS)
Silber, Armin; Gonzalez, Christian; Pino, Francisco; Escarate, Patricio; Gairing, Stefan
2014-08-01
With expanding sizes and increasing complexity of large astronomical observatories on remote observing sites, the call for an efficient and recourses saving maintenance concept becomes louder. The increasing number of subsystems on telescopes and instruments forces large observatories, like in industries, to rethink conventional maintenance strategies for reaching this demanding goal. The implementation of full-, or semi-automatic processes for standard service activities can help to keep the number of operating staff on an efficient level and to reduce significantly the consumption of valuable consumables or equipment. In this contribution we will demonstrate on the example of the 80 Cryogenic subsystems of the ALMA Front End instrument, how an implemented automatic service process increases the availability of spare parts and Line Replaceable Units. Furthermore how valuable staff recourses can be freed from continuous repetitive maintenance activities, to allow focusing more on system diagnostic tasks, troubleshooting and the interchanging of line replaceable units. The required service activities are decoupled from the day-to-day work, eliminating dependencies on workload peaks or logistic constrains. The automatic refurbishing processes running in parallel to the operational tasks with constant quality and without compromising the performance of the serviced system components. Consequentially that results in an efficiency increase, less down time and keeps the observing schedule on track. Automatic service processes in combination with proactive maintenance concepts are providing the necessary flexibility for the complex operational work structures of large observatories. The gained planning flexibility is allowing an optimization of operational procedures and sequences by considering the required cost efficiency.
Avoiding and tolerating latency in large-scale next-generation shared-memory multiprocessors
NASA Technical Reports Server (NTRS)
Probst, David K.
1993-01-01
A scalable solution to the memory-latency problem is necessary to prevent the large latencies of synchronization and memory operations inherent in large-scale shared-memory multiprocessors from reducing high performance. We distinguish latency avoidance and latency tolerance. Latency is avoided when data is brought to nearby locales for future reference. Latency is tolerated when references are overlapped with other computation. Latency-avoiding locales include: processor registers, data caches used temporally, and nearby memory modules. Tolerating communication latency requires parallelism, allowing the overlap of communication and computation. Latency-tolerating techniques include: vector pipelining, data caches used spatially, prefetching in various forms, and multithreading in various forms. Relaxing the consistency model permits increased use of avoidance and tolerance techniques. Each model is a mapping from the program text to sets of partial orders on program operations; it is a convention about which temporal precedences among program operations are necessary. Information about temporal locality and parallelism constrains the use of avoidance and tolerance techniques. Suitable architectural primitives and compiler technology are required to exploit the increased freedom to reorder and overlap operations in relaxed models.
Shape accuracy requirements on starshades for large and small apertures
NASA Astrophysics Data System (ADS)
Shaklan, Stuart B.; Marchen, Luis; Cady, Eric
2017-09-01
Starshades have been designed to work with large and small telescopes alike. With smaller telescopes, the targets tend to be brighter and closer to the Solar System, and their putative planetary systems span angles that require starshades with radii of 10-30 m at distances of 10s of Mm. With larger apertures, the light-collecting power enables studies of more numerous, fainter systems, requiring larger, more distant starshades with radii >50 m at distances of 100s of Mm. Characterization using infrared wavelengths requires even larger starshades. A mitigating approach is to observe planets between the petals, where one can observe regions closer to the star but with reduced throughput and increased instrument scatter. We compare the starshade shape requirements, including petal shape, petal positioning, and other key terms, for the WFIRST 26m starshade and the HABEX 72 m starshade concepts, over a range of working angles and telescope sizes. We also compare starshades having rippled and smooth edges and show that their performance is nearly identical.
Computers and Technological Forecasting
ERIC Educational Resources Information Center
Martino, Joseph P.
1971-01-01
Forecasting is becoming increasingly automated, thanks in large measure to the computer. It is now possible for a forecaster to submit his data to a computation center and call for the appropriate program. (No knowledge of statistics is required.) (Author)
Results of the harmonics measurement program at the John F. Long photovoltaic house
NASA Astrophysics Data System (ADS)
Campen, G. L.
1982-03-01
Photovoltaic (PV) systems used in single-family dwellings require an inverter to act as an interface between the direct-current (dc) power output of the PV unit and the alternating-current (ac) power needed by house loads. A type of inverter known as line commutated injects harmonic currents on the ac side and requires large amounts of reactive power. Large numbers of such PV installations could lead to unacceptable levels of harmonic voltages on the utility system, and the need to increase the utility's deliver of reactive power could result in significant cost increases. The harmonics and power-factor effects are examined for a single PV installation using a line-commutated inverter. The magnitude and phase of various currents and voltages from the fundamental to the 13th harmonic were recorded both with and without the operation of the PV system.
NASA Astrophysics Data System (ADS)
Meng, Lixin; Meng, Lingchen; Zhang, Yiqun; Zhang, Lizhong; Liu, Ming; Li, Xiaoming
2018-01-01
In the satellite to earth laser communication link, large-aperture ground laser communication terminals usually are used in order to realize the requirement of high rate and long distance communication and restrain the power fluctuation by atmospheric scintillation. With the increasing of the laser communication terminal caliber, the primary mirror weight should also be increased, and selfweight, thermal deformation and environment will affect the surface accuracy of the primary mirror surface. A high precision vehicular laser communication telescope unit with an effective aperture of 600mm was considered in this paper. The primary mirror is positioned with center hole, which back is supported by 9 floats and the side is supported by a mercury band. The secondary mirror adopts a spherical adjusting mechanism. Through simulation analysis, the system wave difference is better than λ/20 when the primary mirror is in different dip angle, which meets the requirements of laser communication.
NASA Technical Reports Server (NTRS)
Smith, R. L.; Lyubomirsky, A. S.
1981-01-01
Two techniques were analyzed. The first is a representation using Chebyshev expansions in three-dimensional cells. The second technique employs a temporary file for storing the components of the nonspherical gravity force. Computer storage requirements and relative CPU time requirements are presented. The Chebyshev gravity representation can provide a significant reduction in CPU time in precision orbit calculations, but at the cost of a large amount of direct-access storage space, which is required for a global model.
Low-Field-Triggered Large Magnetostriction in Iron-Palladium Strain Glass Alloys.
Ren, Shuai; Xue, Dezhen; Ji, Yuanchao; Liu, Xiaolian; Yang, Sen; Ren, Xiaobing
2017-09-22
Development of miniaturized magnetostriction-associated devices requires low-field-triggered large magnetostriction. In this study, we acquired a large magnetostriction (800 ppm) triggered by a low saturation field (0.8 kOe) in iron-palladium (Fe-Pd) alloys. Magnetostriction enhancement jumping from 340 to 800 ppm was obtained with a slight increase in Pd concentration from 31.3 to 32.3 at. %. Further analysis showed that such a slight increase led to suppression of the long-range ordered martensitic phase and resulted in a frozen short-range ordered strain glass state. This strain glass state possessed a two-phase nanostructure with nanosized frozen strain domains embedded in the austenite matrix, which was responsible for the unique magnetostriction behavior. Our study provides a way to design novel magnetostrictive materials with low-field-triggered large magnetostriction.
Expanding Computer Science Education in Schools: Understanding Teacher Experiences and Challenges
ERIC Educational Resources Information Center
Yadav, Aman; Gretter, Sarah; Hambrusch, Susanne; Sands, Phil
2017-01-01
The increased push for teaching computer science (CS) in schools in the United States requires training a large number of new K-12 teachers. The current efforts to increase the number of CS teachers have predominantly focused on training teachers from other content areas. In order to support these beginning CS teachers, we need to better…
Kepper, Nick; Ettig, Ramona; Dickmann, Frank; Stehr, Rene; Grosveld, Frank G; Wedemann, Gero; Knoch, Tobias A
2010-01-01
Especially in the life-science and the health-care sectors the huge IT requirements are imminent due to the large and complex systems to be analysed and simulated. Grid infrastructures play here a rapidly increasing role for research, diagnostics, and treatment, since they provide the necessary large-scale resources efficiently. Whereas grids were first used for huge number crunching of trivially parallelizable problems, increasingly parallel high-performance computing is required. Here, we show for the prime example of molecular dynamic simulations how the presence of large grid clusters including very fast network interconnects within grid infrastructures allows now parallel high-performance grid computing efficiently and thus combines the benefits of dedicated super-computing centres and grid infrastructures. The demands for this service class are the highest since the user group has very heterogeneous requirements: i) two to many thousands of CPUs, ii) different memory architectures, iii) huge storage capabilities, and iv) fast communication via network interconnects, are all needed in different combinations and must be considered in a highly dedicated manner to reach highest performance efficiency. Beyond, advanced and dedicated i) interaction with users, ii) the management of jobs, iii) accounting, and iv) billing, not only combines classic with parallel high-performance grid usage, but more importantly is also able to increase the efficiency of IT resource providers. Consequently, the mere "yes-we-can" becomes a huge opportunity like e.g. the life-science and health-care sectors as well as grid infrastructures by reaching higher level of resource efficiency.
Common procedures and strategies for anaesthesia in interventional radiology.
Landrigan-Ossar, Mary
2015-08-01
This review describes the range of cases now available in the interventional radiology suite and summarizes suggestions for their anaesthetic and perioperative management. The type and complexity of interventional radiology cases being performed increases from year to year. Anaesthesiologists' presence in interventional radiology is increasing in turn, due to increasingly ill patients and intricate procedures requiring more than local anaesthesia for well tolerated completion. The literature available describing this is largely written by radiologists, with little attention paid to anaesthetic considerations. Cases in interventional radiology are complex in terms of the logistics of working in an unfamiliar area, frequency of patient comorbidity and unfamiliar procedures. Ensuring familiarity with the variety of interventional radiology procedures and their periprocedure requirements can increase anaesthesiologists' comfort in interventional radiology.
Microworld Simulations: A New Dimension in Training Army Logistics Management Skills
2004-01-01
Providing effective training to Army personnelis always challenging, but the Army facessome new challenges in training its logisticsstaff managers in...soldiers are stationed and where materiel and services are readily available. The design and management of the Army’s Combat Ser- vice Support (CSS) large...scale logistics systems are increasingly important. The skills that are required to manage these systems are difficult to train. Large deployments
Crop water productivity and irrigation management
USDA-ARS?s Scientific Manuscript database
Modern irrigation systems offer large increases in crop water productivity compared with rainfed or gravity irrigation, but require different management approaches to achieve this. Flood, sprinkler, low-energy precision application, LEPA, and subsurface drip irrigation methods vary widely in water a...
NASA Technical Reports Server (NTRS)
Griffin, Charles F.; James, Arthur M.
1985-01-01
The damage-tolerance characteristics of high strain-to-failure graphite fibers and toughened resins were evaluated. Test results show that conventional fuel tank sealing techniques are applicable to composite structures. Techniques were developed to prevent fuel leaks due to low-energy impact damage. For wing panels subjected to swept stroke lightning strikes, a surface protection of graphite/aluminum wire fabric and a fastener treatment proved effective in eliminating internal sparking and reducing structural damage. The technology features developed were incorporated and demonstrated in a test panel designed to meet the strength, stiffness, and damage tolerance requirements of a large commercial transport aircraft. The panel test results exceeded design requirements for all test conditions. Wing surfaces constructed with composites offer large weight savings if design allowable strains for compression can be increased from current levels.
Normal myoblast fusion requires myoferlin
Doherty, Katherine R.; Cave, Andrew; Davis, Dawn Belt; Delmonte, Anthony J.; Posey, Avery; Earley, Judy U.; Hadhazy, Michele; McNally, Elizabeth M.
2014-01-01
Summary Muscle growth occurs during embryonic development and continues in adult life as regeneration. During embryonic muscle growth and regeneration in mature muscle, singly nucleated myoblasts fuse to each other to form myotubes. In muscle growth, singly nucleated myoblasts can also fuse to existing large, syncytial myofibers as a mechanism of increasing muscle mass without increasing myofiber number. Myoblast fusion requires the alignment and fusion of two apposed lipid bilayers. The repair of muscle plasma membrane disruptions also relies on the fusion of two apposed lipid bilayers. The protein dysferlin, the product of the Limb Girdle Muscular Dystrophy type 2 locus, has been shown to be necessary for efficient, calcium-sensitive, membrane resealing. We now show that the related protein myoferlin is highly expressed in myoblasts undergoing fusion, and is expressed at the site of myoblasts fusing to myotubes. Like dysferlin, we found that myoferlin binds phospholipids in a calcium-sensitive manner that requires the first C2A domain. We generated mice with a null allele of myoferlin. Myoferlin null myoblasts undergo initial fusion events, but they form large myotubes less efficiently in vitro, consistent with a defect in a later stage of myogenesis. In vivo, myoferlin null mice have smaller muscles than controls do, and myoferlin null muscle lacks large diameter myofibers. Additionally, myoferlin null muscle does not regenerate as well as wild-type muscle does, and instead displays a dystrophic phenotype. These data support a role for myoferlin in the maturation of myotubes and the formation of large myotubes that arise from the fusion of myoblasts to multinucleate myotubes. PMID:16280346
Wood, Tim J; Moore, Craig S; Horsfield, Carl J; Saunderson, John R; Beavis, Andrew W
2015-01-01
The purpose of this study was to develop size-based radiotherapy kilovoltage cone beam CT (CBCT) protocols for the pelvis. Image noise was measured in an elliptical phantom of varying size for a range of exposure factors. Based on a previously defined "small pelvis" reference patient and CBCT protocol, appropriate exposure factors for small, medium, large and extra-large patients were derived which approximate the image noise behaviour observed on a Philips CT scanner (Philips Medical Systems, Best, Netherlands) with automatic exposure control (AEC). Selection criteria, based on maximum tube current-time product per rotation selected during the radiotherapy treatment planning scan, were derived based on an audit of patient size. It has been demonstrated that 110 kVp yields acceptable image noise for reduced patient dose in pelvic CBCT scans of small, medium and large patients, when compared with manufacturer's default settings (125 kVp). Conversely, extra-large patients require increased exposure factors to give acceptable images. 57% of patients in the local population now receive much lower radiation doses, whereas 13% require higher doses (but now yield acceptable images). The implementation of size-based exposure protocols has significantly reduced radiation dose to the majority of patients with no negative impact on image quality. Increased doses are required on the largest patients to give adequate image quality. The development of size-based CBCT protocols that use the planning CT scan (with AEC) to determine which protocol is appropriate ensures adequate image quality whilst minimizing patient radiation dose.
Tyrannosaurus en pointe: allometry minimized rotational inertia of large carnivorous dinosaurs.
Henderson, Donald M; Snively, Eric
2004-01-01
Theropod dinosaurs attained the largest body sizes among terrestrial predators, and were also unique in being exclusively bipedal. With only two limbs for propulsion and balance, theropods would have been greatly constrained in their locomotor performance at large body size. Using three-dimensional restorations of the axial bodies and limbs of 12 theropod dinosaurs, and determining their rotational inertias (RIs) about a vertical axis, we show that these animals expressed a pattern of phyletic size increase that minimized the increase in RI associated with increases in body size. By contrast, the RI of six quadrupedal, carnivorous archosaurs exhibited changes in body proportions that were closer to those predicted by isometry. Correlations of low RI with high agility in lizards suggest that large theropods, with low relative RI, could engage in activities requiring higher agility than would be possible with isometric scaling. PMID:15101419
A digital gigapixel large-format tile-scan camera.
Ben-Ezra, M
2011-01-01
Although the resolution of single-lens reflex (SLR) and medium-format digital cameras has increased in recent years, applications for cultural-heritage preservation and computational photography require even higher resolutions. Addressing this issue, a large-format cameras' large image planes can achieve very high resolution without compromising pixel size and thus can provide high-quality, high-resolution images.This digital large-format tile scan camera can acquire high-quality, high-resolution images of static scenes. It employs unique calibration techniques and a simple algorithm for focal-stack processing of very large images with significant magnification variations. The camera automatically collects overlapping focal stacks and processes them into a high-resolution, extended-depth-of-field image.
How much electrical energy storage do we need? A synthesis for the U.S., Europe, and Germany
Cebulla, Felix; Haas, Jannik; Eichman, Josh; ...
2018-02-03
Electrical energy storage (EES) is a promising flexibility source for prospective low-carbon energy systems. In the last couple of years, many studies for EES capacity planning have been produced. However, these resulted in a very broad range of power and energy capacity requirements for storage, making it difficult for policymakers to identify clear storage planning recommendations. Therefore, we studied 17 recent storage expansion studies pertinent to the U.S., Europe, and Germany. We then systemized the storage requirement per variable renewable energy (VRE) share and generation technology. Our synthesis reveals that with increasing VRE shares, the EES power capacity increases linearly;more » and the energy capacity, exponentially. Further, by analyzing the outliers, the EES energy requirements can be at least halved. It becomes clear that grids dominated by photovoltaic energy call for more EES, while large shares of wind rely more on transmission capacity. Taking into account the energy mix clarifies - to a large degree - the apparent conflict of the storage requirements between the existing studies. Finally, there might exist a negative bias towards storage because transmission costs are frequently optimistic (by neglecting execution delays and social opposition) and storage can cope with uncertainties, but these issues are rarely acknowledged in the planning process.« less
How much electrical energy storage do we need? A synthesis for the U.S., Europe, and Germany
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cebulla, Felix; Haas, Jannik; Eichman, Josh
Electrical energy storage (EES) is a promising flexibility source for prospective low-carbon energy systems. In the last couple of years, many studies for EES capacity planning have been produced. However, these resulted in a very broad range of power and energy capacity requirements for storage, making it difficult for policymakers to identify clear storage planning recommendations. Therefore, we studied 17 recent storage expansion studies pertinent to the U.S., Europe, and Germany. We then systemized the storage requirement per variable renewable energy (VRE) share and generation technology. Our synthesis reveals that with increasing VRE shares, the EES power capacity increases linearly;more » and the energy capacity, exponentially. Further, by analyzing the outliers, the EES energy requirements can be at least halved. It becomes clear that grids dominated by photovoltaic energy call for more EES, while large shares of wind rely more on transmission capacity. Taking into account the energy mix clarifies - to a large degree - the apparent conflict of the storage requirements between the existing studies. Finally, there might exist a negative bias towards storage because transmission costs are frequently optimistic (by neglecting execution delays and social opposition) and storage can cope with uncertainties, but these issues are rarely acknowledged in the planning process.« less
Cereal area and nitrogen use efficiency are drivers of future nitrogen fertilizer consumption.
Dobermann, Achim; Cassman, Kenneth G
2005-09-01
At a global scale, cereal yields and fertilizer N consumption have increased in a near-linear fashion during the past 40 years and are highly correlated with one another. However, large differences exist in historical trends of N fertilizer usage and nitrogen use efficiency (NUE) among regions, countries, and crops. The reasons for these differences must be understood to estimate future N fertilizer requirements. Global nitrogen needs will depend on: (i) changes in cropped cereal area and the associated yield increases required to meet increasing cereal demand from population and income growth, and (ii) changes in NUE at the farm level. Our analysis indicates that the anticipated 38% increase in global cereal demand by 2025 can be met by a 30% increase in N use on cereals, provided that the steady decline in cereal harvest area is halted and the yield response to applied N can be increased by 20%. If losses of cereal cropping area continue at the rate of the past 20 years (-0.33% per year) and NUE cannot be increased substantially, a 60% increase in global N use on cereals would be required to meet cereal demand. Interventions to increase NUE and reduce N losses to the environment must be accomplished at the farm-or field-scale through a combination of improved technologies and carefully crafted local policies that contribute to the adoption of improved N management; uniform regional or national directives are unlikely to be effective at both sustaining yield increases and improving NUE. Examples from several countries show that increases in NUE at rates of 1% per year or more can be achieved if adequate investments are made in research and extension. Failure to arrest the decrease in cereal crop area and to improve NUE in the world's most important agricultural systems will likely cause severe damage to environmental services at local, regional, and global scales due to a large increase in reactive N load in the environment.
Cereal area and nitrogen use efficiency are drivers of future nitrogen fertilizer consumption.
Dobermann, Achim; Cassman, Kenneth G
2005-12-01
At a global scale, cereal yields and fertilizer N consumption have increased in a near-linear fashion during the past 40 years and are highly correlated with one another. However, large differences exist in historical trends of N fertilizer usage and nitrogen use efficiency (NUE) among regions, countries, and crops. The reasons for these differences must be understood to estimate future N fertilizer requirements. Global nitrogen needs will depend on: (i) changes in cropped cereal area and the associated yield increases required to meet increasing cereal demand from population and income growth, and (ii) changes in NUE at the farm level. Our analysis indicates that the anticipated 38% increase in global cereal demand by 2025 can be met by a 30% increase in N use on cereals, provided that the steady decline in cereal harvest area is halted and the yield response to applied N can be increased by 20%. If losses of cereal cropping area continue at the rate of the past 20 years (-0.33% per year) and NUE cannot be increased substantially, a 60% increase in global N use on cereals would be required to meet cereal demand. Interventions to increase NUE and reduce N losses to the environment must be accomplished at the farm- or field-scale through a combination of improved technologies and carefully crafted local policies that contribute to the adoption of improved N management; uniform regional or national directives are unlikey to be effective at both sustaining yield increases and improving NUE. Examples from several countries show that increases in NUE at rates of 1% per year or more can be achieved if adequate investments are made in research and extension. Failure to arrest the decrease in cereal crop area and to improve NUE in the world's most important agricultural systems will likely cause severe damage to environmental services at local, regional, and global scales due to a large increase in reactive N load in the environment.
Metal-Matrix Composites: Status and Prospects
1974-12-01
increased to meet or exceed even the most severe engine requirements. This optimism is based largely on an order-of-magnitude improvement in Charpy test...relationships between coupon tests, subsize structural components, and structural component evaluations 4. The development of su-table design
An investigation of older driver freeway needs and capabilities : summary report
DOT National Transportation Integrated Search
1998-11-01
Limited-access highways and the Interstate highway system have been largely responsible for greatly increasing the mobility and safety enjoyed by the American driving population. However, as the driving population ages, it is not known if the require...
NASA Astrophysics Data System (ADS)
Aaboud, M.; Aad, G.; Abbott, B.; Abdinov, O.; Abeloos, B.; Abidi, S. H.; AbouZeid, O. S.; Abraham, N. L.; Abramowicz, H.; Abreu, H.; Abreu, R.; Abulaiti, Y.; Acharya, B. S.; Adachi, S.; Adamczyk, L.; Adelman, J.; Adersberger, M.; Adye, T.; Affolder, A. A.; Afik, Y.; Agatonovic-Jovin, T.; Agheorghiesei, C.; Aguilar-Saavedra, J. A.; Ahlen, S. P.; Ahmadov, F.; Aielli, G.; Akatsuka, S.; Akerstedt, H.; Åkesson, T. P. A.; Akilli, E.; Akimov, A. V.; Alberghi, G. L.; Albert, J.; Albicocco, P.; Alconada Verzini, M. J.; Alderweireldt, S. C.; Aleksa, M.; Aleksandrov, I. N.; Alexa, C.; Alexander, G.; Alexopoulos, T.; Alhroob, M.; Ali, B.; Aliev, M.; Alimonti, G.; Alison, J.; Alkire, S. P.; Allbrooke, B. M. M.; Allen, B. W.; Allport, P. P.; Aloisio, A.; Alonso, A.; Alonso, F.; Alpigiani, C.; Alshehri, A. A.; Alstaty, M. I.; Alvarez Gonzalez, B.; Álvarez Piqueras, D.; Alviggi, M. G.; Amadio, B. T.; Amaral Coutinho, Y.; Amelung, C.; Amidei, D.; Amor Dos Santos, S. P.; Amoroso, S.; Amundsen, G.; Anastopoulos, C.; Ancu, L. S.; Andari, N.; Andeen, T.; Anders, C. F.; Anders, J. K.; Anderson, K. J.; Andreazza, A.; Andrei, V.; Angelidakis, S.; Angelozzi, I.; Angerami, A.; Anisenkov, A. V.; Anjos, N.; Annovi, A.; Antel, C.; Antonelli, M.; Antonov, A.; Antrim, D. J.; Anulli, F.; Aoki, M.; Aperio Bella, L.; Arabidze, G.; Arai, Y.; Araque, J. P.; Araujo Ferraz, V.; Arce, A. T. H.; Ardell, R. E.; Arduh, F. A.; Arguin, J.-F.; Argyropoulos, S.; Arik, M.; Armbruster, A. J.; Armitage, L. J.; Arnaez, O.; Arnold, H.; Arratia, M.; Arslan, O.; Artamonov, A.; Artoni, G.; Artz, S.; Asai, S.; Asbah, N.; Ashkenazi, A.; Asquith, L.; Assamagan, K.; Astalos, R.; Atkinson, M.; Atlay, N. B.; Augsten, K.; Avolio, G.; Axen, B.; Ayoub, M. K.; Azuelos, G.; Baas, A. E.; Baca, M. J.; Bachacou, H.; Bachas, K.; Backes, M.; Bagnaia, P.; Bahmani, M.; Bahrasemani, H.; Baines, J. T.; Bajic, M.; Baker, O. K.; Bakker, P. J.; Baldin, E. M.; Balek, P.; Balli, F.; Balunas, W. K.; Banas, E.; Bandyopadhyay, A.; Banerjee, Sw.; Bannoura, A. A. E.; Barak, L.; Barberio, E. L.; Barberis, D.; Barbero, M.; Barillari, T.; Barisits, M.-S.; Barkeloo, J. T.; Barklow, T.; Barlow, N.; Barnes, S. L.; Barnett, B. M.; Barnett, R. M.; Barnovska-Blenessy, Z.; Baroncelli, A.; Barone, G.; Barr, A. J.; Barranco Navarro, L.; Barreiro, F.; Barreiro Guimarães da Costa, J.; Bartoldus, R.; Barton, A. E.; Bartos, P.; Basalaev, A.; Bassalat, A.; Bates, R. L.; Batista, S. J.; Batley, J. R.; Battaglia, M.; Bauce, M.; Bauer, F.; Bawa, H. S.; Beacham, J. B.; Beattie, M. D.; Beau, T.; Beauchemin, P. H.; Bechtle, P.; Beck, H. P.; Beck, H. C.; Becker, K.; Becker, M.; Becot, C.; Beddall, A. J.; Beddall, A.; Bednyakov, V. A.; Bedognetti, M.; Bee, C. P.; Beermann, T. A.; Begalli, M.; Begel, M.; Behr, J. K.; Bell, A. S.; Bella, G.; Bellagamba, L.; Bellerive, A.; Bellomo, M.; Belotskiy, K.; Beltramello, O.; Belyaev, N. L.; Benary, O.; Benchekroun, D.; Bender, M.; Benekos, N.; Benhammou, Y.; Benhar Noccioli, E.; Benitez, J.; Benjamin, D. P.; Benoit, M.; Bensinger, J. R.; Bentvelsen, S.; Beresford, L.; Beretta, M.; Berge, D.; Bergeaas Kuutmann, E.; Berger, N.; Bergsten, L. J.; Beringer, J.; Berlendis, S.; Bernard, N. R.; Bernardi, G.; Bernius, C.; Bernlochner, F. U.; Berry, T.; Berta, P.; Bertella, C.; Bertoli, G.; Bertram, I. A.; Bertsche, C.; Besjes, G. J.; Bessidskaia Bylund, O.; Bessner, M.; Besson, N.; Bethani, A.; Bethke, S.; Betti, A.; Bevan, A. J.; Beyer, J.; Bianchi, R. M.; Biebel, O.; Biedermann, D.; Bielski, R.; Bierwagen, K.; Biesuz, N. V.; Biglietti, M.; Billoud, T. R. V.; Bilokon, H.; Bindi, M.; Bingul, A.; Bini, C.; Biondi, S.; Bisanz, T.; Bittrich, C.; Bjergaard, D. M.; Black, J. E.; Black, K. M.; Blair, R. E.; Blazek, T.; Bloch, I.; Blocker, C.; Blue, A.; Blumenschein, U.; Blunier, S.; Bobbink, G. J.; Bobrovnikov, V. S.; Bocchetta, S. S.; Bocci, A.; Bock, C.; Boehler, M.; Boerner, D.; Bogavac, D.; Bogdanchikov, A. G.; Bohm, C.; Boisvert, V.; Bokan, P.; Bold, T.; Boldyrev, A. S.; Bolz, A. E.; Bomben, M.; Bona, M.; Boonekamp, M.; Borisov, A.; Borissov, G.; Bortfeldt, J.; Bortoletto, D.; Bortolotto, V.; Boscherini, D.; Bosman, M.; Bossio Sola, J. D.; Boudreau, J.; Bouhova-Thacker, E. V.; Boumediene, D.; Bourdarios, C.; Boutle, S. K.; Boveia, A.; Boyd, J.; Boyko, I. R.; Bozson, A. J.; Bracinik, J.; Brandt, A.; Brandt, G.; Brandt, O.; Braren, F.; Bratzler, U.; Brau, B.; Brau, J. E.; Breaden Madden, W. D.; Brendlinger, K.; Brennan, A. J.; Brenner, L.; Brenner, R.; Bressler, S.; Briglin, D. L.; Bristow, T. M.; Britton, D.; Britzger, D.; Brochu, F. M.; Brock, I.; Brock, R.; Brooijmans, G.; Brooks, T.; Brooks, W. K.; Brosamer, J.; Brost, E.; Broughton, J. H.; Bruckman de Renstrom, P. A.; Bruncko, D.; Bruni, A.; Bruni, G.; Bruni, L. S.; Bruno, S.; Brunt, BH; Bruschi, M.; Bruscino, N.; Bryant, P.; Bryngemark, L.; Buanes, T.; Buat, Q.; Buchholz, P.; Buckley, A. G.; Budagov, I. A.; Buehrer, F.; Bugge, M. K.; Bulekov, O.; Bullock, D.; Burch, T. J.; Burdin, S.; Burgard, C. D.; Burger, A. M.; Burghgrave, B.; Burka, K.; Burke, S.; Burmeister, I.; Burr, J. T. P.; Büscher, D.; Büscher, V.; Bussey, P.; Butler, J. M.; Buttar, C. M.; Butterworth, J. M.; Butti, P.; Buttinger, W.; Buzatu, A.; Buzykaev, A. R.; Changqiao, C.-Q.; Cabrera Urbán, S.; Caforio, D.; Cai, H.; Cairo, V. M.; Cakir, O.; Calace, N.; Calafiura, P.; Calandri, A.; Calderini, G.; Calfayan, P.; Callea, G.; Caloba, L. P.; Calvente Lopez, S.; Calvet, D.; Calvet, S.; Calvet, T. P.; Camacho Toro, R.; Camarda, S.; Camarri, P.; Cameron, D.; Caminal Armadans, R.; Camincher, C.; Campana, S.; Campanelli, M.; Camplani, A.; Campoverde, A.; Canale, V.; Cano Bret, M.; Cantero, J.; Cao, T.; Capeans Garrido, M. D. M.; Caprini, I.; Caprini, M.; Capua, M.; Carbone, R. M.; Cardarelli, R.; Cardillo, F.; Carli, I.; Carli, T.; Carlino, G.; Carlson, B. T.; Carminati, L.; Carney, R. M. D.; Caron, S.; Carquin, E.; Carrá, S.; Carrillo-Montoya, G. D.; Casadei, D.; Casado, M. P.; Casha, A. F.; Casolino, M.; Casper, D. W.; Castelijn, R.; Castillo Gimenez, V.; Castro, N. F.; Catinaccio, A.; Catmore, J. R.; Cattai, A.; Caudron, J.; Cavaliere, V.; Cavallaro, E.; Cavalli, D.; Cavalli-Sforza, M.; Cavasinni, V.; Celebi, E.; Ceradini, F.; Cerda Alberich, L.; Cerqueira, A. S.; Cerri, A.; Cerrito, L.; Cerutti, F.; Cervelli, A.; Cetin, S. A.; Chafaq, A.; Chakraborty, D.; Chan, S. K.; Chan, W. S.; Chan, Y. L.; Chang, P.; Chapman, J. D.; Charlton, D. G.; Chau, C. C.; Chavez Barajas, C. A.; Che, S.; Cheatham, S.; Chegwidden, A.; Chekanov, S.; Chekulaev, S. V.; Chelkov, G. A.; Chelstowska, M. A.; Chen, C.; Chen, C.; Chen, H.; Chen, J.; Chen, S.; Chen, S.; Chen, X.; Chen, Y.; Cheng, H. C.; Cheng, H. J.; Cheplakov, A.; Cheremushkina, E.; Cherkaoui El Moursli, R.; Cheu, E.; Cheung, K.; Chevalier, L.; Chiarella, V.; Chiarelli, G.; Chiodini, G.; Chisholm, A. S.; Chitan, A.; Chiu, Y. H.; Chizhov, M. V.; Choi, K.; Chomont, A. R.; Chouridou, S.; Chow, Y. S.; Christodoulou, V.; Chu, M. C.; Chudoba, J.; Chuinard, A. J.; Chwastowski, J. J.; Chytka, L.; Ciftci, A. K.; Cinca, D.; Cindro, V.; Cioara, I. A.; Ciocio, A.; Cirotto, F.; Citron, Z. H.; Citterio, M.; Ciubancan, M.; Clark, A.; Clark, B. L.; Clark, M. R.; Clark, P. J.; Clarke, R. N.; Clement, C.; Coadou, Y.; Cobal, M.; Coccaro, A.; Cochran, J.; Colasurdo, L.; Cole, B.; Colijn, A. P.; Collot, J.; Colombo, T.; Conde Muiño, P.; Coniavitis, E.; Connell, S. H.; Connelly, I. A.; Constantinescu, S.; Conti, G.; Conventi, F.; Cooke, M.; Cooper-Sarkar, A. M.; Cormier, F.; Cormier, K. J. R.; Corradi, M.; Corriveau, F.; Cortes-Gonzalez, A.; Costa, G.; Costa, M. J.; Costanzo, D.; Cottin, G.; Cowan, G.; Cox, B. E.; Cranmer, K.; Crawley, S. J.; Creager, R. A.; Cree, G.; Crépé-Renaudin, S.; Crescioli, F.; Cribbs, W. A.; Cristinziani, M.; Croft, V.; Crosetti, G.; Cueto, A.; Cuhadar Donszelmann, T.; Cukierman, A. R.; Cummings, J.; Curatolo, M.; Cúth, J.; Czekierda, S.; Czodrowski, P.; D'amen, G.; D'Auria, S.; D'eramo, L.; D'Onofrio, M.; Da Cunha Sargedas De Sousa, M. J.; Da Via, C.; Dabrowski, W.; Dado, T.; Dai, T.; Dale, O.; Dallaire, F.; Dallapiccola, C.; Dam, M.; Dandoy, J. R.; Daneri, M. F.; Dang, N. P.; Daniells, A. C.; Dann, N. S.; Danninger, M.; Dano Hoffmann, M.; Dao, V.; Darbo, G.; Darmora, S.; Dassoulas, J.; Dattagupta, A.; Daubney, T.; Davey, W.; David, C.; Davidek, T.; Davis, D. R.; Davison, P.; Dawe, E.; Dawson, I.; De, K.; de Asmundis, R.; De Benedetti, A.; De Castro, S.; De Cecco, S.; De Groot, N.; de Jong, P.; De la Torre, H.; De Lorenzi, F.; De Maria, A.; De Pedis, D.; De Salvo, A.; De Sanctis, U.; De Santo, A.; De Vasconcelos Corga, K.; De Vivie De Regie, J. B.; Debbe, R.; Debenedetti, C.; Dedovich, D. V.; Dehghanian, N.; Deigaard, I.; Del Gaudio, M.; Del Peso, J.; Delgove, D.; Deliot, F.; Delitzsch, C. M.; Dell'Acqua, A.; Dell'Asta, L.; Dell'Orso, M.; Della Pietra, M.; della Volpe, D.; Delmastro, M.; Delporte, C.; Delsart, P. A.; DeMarco, D. A.; Demers, S.; Demichev, M.; Demilly, A.; Denisov, S. P.; Denysiuk, D.; Derendarz, D.; Derkaoui, J. E.; Derue, F.; Dervan, P.; Desch, K.; Deterre, C.; Dette, K.; Devesa, M. R.; Deviveiros, P. O.; Dewhurst, A.; Dhaliwal, S.; Di Bello, F. A.; Di Ciaccio, A.; Di Ciaccio, L.; Di Clemente, W. K.; Di Donato, C.; Di Girolamo, A.; Di Girolamo, B.; Di Micco, B.; Di Nardo, R.; Di Petrillo, K. F.; Di Simone, A.; Di Sipio, R.; Di Valentino, D.; Diaconu, C.; Diamond, M.; Dias, F. A.; Diaz, M. A.; Dickinson, J.; Diehl, E. B.; Dietrich, J.; Díez Cornell, S.; Dimitrievska, A.; Dingfelder, J.; Dita, P.; Dita, S.; Dittus, F.; Djama, F.; Djobava, T.; Djuvsland, J. I.; do Vale, M. A. B.; Dobos, D.; Dobre, M.; Dodsworth, D.; Doglioni, C.; Dolejsi, J.; Dolezal, Z.; Donadelli, M.; Donati, S.; Dondero, P.; Donini, J.; Dopke, J.; Doria, A.; Dova, M. T.; Doyle, A. T.; Drechsler, E.; Dris, M.; Du, Y.; Duarte-Campderros, J.; Dubinin, F.; Dubreuil, A.; Duchovni, E.; Duckeck, G.; Ducourthial, A.; Ducu, O. A.; Duda, D.; Dudarev, A.; Dudder, A. Chr.; Duffield, E. M.; Duflot, L.; Dührssen, M.; Dulsen, C.; Dumancic, M.; Dumitriu, A. E.; Duncan, A. K.; Dunford, M.; Duperrin, A.; Duran Yildiz, H.; Düren, M.; Durglishvili, A.; Duschinger, D.; Dutta, B.; Duvnjak, D.; Dyndal, M.; Dziedzic, B. S.; Eckardt, C.; Ecker, K. M.; Edgar, R. C.; Eifert, T.; Eigen, G.; Einsweiler, K.; Ekelof, T.; El Kacimi, M.; El Kosseifi, R.; Ellajosyula, V.; Ellert, M.; Elles, S.; Ellinghaus, F.; Elliot, A. A.; Ellis, N.; Elmsheuser, J.; Elsing, M.; Emeliyanov, D.; Enari, Y.; Ennis, J. S.; Epland, M. B.; Erdmann, J.; Ereditato, A.; Ernst, M.; Errede, S.; Escalier, M.; Escobar, C.; Esposito, B.; Estrada Pastor, O.; Etienvre, A. I.; Etzion, E.; Evans, H.; Ezhilov, A.; Ezzi, M.; Fabbri, F.; Fabbri, L.; Fabiani, V.; Facini, G.; Fakhrutdinov, R. M.; Falciano, S.; Falla, R. J.; Faltova, J.; Fang, Y.; Fanti, M.; Farbin, A.; Farilla, A.; Farina, C.; Farina, E. M.; Farooque, T.; Farrell, S.; Farrington, S. M.; Farthouat, P.; Fassi, F.; Fassnacht, P.; Fassouliotis, D.; Faucci Giannelli, M.; Favareto, A.; Fawcett, W. J.; Fayard, L.; Fedin, O. L.; Fedorko, W.; Feigl, S.; Feligioni, L.; Feng, C.; Feng, E. J.; Fenton, M. J.; Fenyuk, A. B.; Feremenga, L.; Fernandez Martinez, P.; Ferrando, J.; Ferrari, A.; Ferrari, P.; Ferrari, R.; Ferreira de Lima, D. E.; Ferrer, A.; Ferrere, D.; Ferretti, C.; Fiedler, F.; Filipčič, A.; Filipuzzi, M.; Filthaut, F.; Fincke-Keeler, M.; Finelli, K. D.; Fiolhais, M. C. N.; Fiorini, L.; Fischer, A.; Fischer, C.; Fischer, J.; Fisher, W. C.; Flaschel, N.; Fleck, I.; Fleischmann, P.; Fletcher, R. R. M.; Flick, T.; Flierl, B. M.; Flores Castillo, L. R.; Flowerdew, M. J.; Forcolin, G. T.; Formica, A.; Förster, F. A.; Forti, A.; Foster, A. G.; Fournier, D.; Fox, H.; Fracchia, S.; Francavilla, P.; Franchini, M.; Franchino, S.; Francis, D.; Franconi, L.; Franklin, M.; Frate, M.; Fraternali, M.; Freeborn, D.; Fressard-Batraneanu, S. M.; Freund, B.; Froidevaux, D.; Frost, J. A.; Fukunaga, C.; Fusayasu, T.; Fuster, J.; Gabizon, O.; Gabrielli, A.; Gabrielli, A.; Gach, G. P.; Gadatsch, S.; Gadomski, S.; Gagliardi, G.; Gagnon, L. G.; Galea, C.; Galhardo, B.; Gallas, E. J.; Gallop, B. J.; Gallus, P.; Galster, G.; Gan, K. K.; Ganguly, S.; Gao, Y.; Gao, Y. S.; Garay Walls, F. M.; García, C.; García Navarro, J. E.; García Pascual, J. A.; Garcia-Sciveres, M.; Gardner, R. W.; Garelli, N.; Garonne, V.; Gascon Bravo, A.; Gasnikova, K.; Gatti, C.; Gaudiello, A.; Gaudio, G.; Gavrilenko, I. L.; Gay, C.; Gaycken, G.; Gazis, E. N.; Gee, C. N. P.; Geisen, J.; Geisen, M.; Geisler, M. P.; Gellerstedt, K.; Gemme, C.; Genest, M. H.; Geng, C.; Gentile, S.; Gentsos, C.; George, S.; Gerbaudo, D.; Geßner, G.; Ghasemi, S.; Ghneimat, M.; Giacobbe, B.; Giagu, S.; Giangiacomi, N.; Giannetti, P.; Gibson, S. M.; Gignac, M.; Gilchriese, M.; Gillberg, D.; Gilles, G.; Gingrich, D. M.; Giordani, M. P.; Giorgi, F. M.; Giraud, P. F.; Giromini, P.; Giugliarelli, G.; Giugni, D.; Giuli, F.; Giuliani, C.; Giulini, M.; Gjelsten, B. K.; Gkaitatzis, S.; Gkialas, I.; Gkougkousis, E. L.; Gkountoumis, P.; Gladilin, L. K.; Glasman, C.; Glatzer, J.; Glaysher, P. C. F.; Glazov, A.; Goblirsch-Kolb, M.; Godlewski, J.; Goldfarb, S.; Golling, T.; Golubkov, D.; Gomes, A.; Gonçalo, R.; Goncalves Gama, R.; Goncalves Pinto Firmino Da Costa, J.; Gonella, G.; Gonella, L.; Gongadze, A.; Gonski, J. L.; González de la Hoz, S.; Gonzalez-Sevilla, S.; Goossens, L.; Gorbounov, P. A.; Gordon, H. A.; Gorelov, I.; Gorini, B.; Gorini, E.; Gorišek, A.; Goshaw, A. T.; Gössling, C.; Gostkin, M. I.; Gottardo, C. A.; Goudet, C. R.; Goujdami, D.; Goussiou, A. G.; Govender, N.; Gozani, E.; Grabowska-Bold, I.; Gradin, P. O. J.; Gramling, J.; Gramstad, E.; Grancagnolo, S.; Gratchev, V.; Gravila, P. M.; Gray, C.; Gray, H. M.; Greenwood, Z. D.; Grefe, C.; Gregersen, K.; Gregor, I. M.; Grenier, P.; Grevtsov, K.; Griffiths, J.; Grillo, A. A.; Grimm, K.; Grinstein, S.; Gris, Ph.; Grivaz, J.-F.; Groh, S.; Gross, E.; Grosse-Knetter, J.; Grossi, G. C.; Grout, Z. J.; Grummer, A.; Guan, L.; Guan, W.; Guenther, J.; Guescini, F.; Guest, D.; Gueta, O.; Gui, B.; Guido, E.; Guillemin, T.; Guindon, S.; Gul, U.; Gumpert, C.; Guo, J.; Guo, W.; Guo, Y.; Gupta, R.; Gurbuz, S.; Gustavino, G.; Gutelman, B. J.; Gutierrez, P.; Gutierrez Ortiz, N. G.; Gutschow, C.; Guyot, C.; Guzik, M. P.; Gwenlan, C.; Gwilliam, C. B.; Haas, A.; Haber, C.; Hadavand, H. K.; Haddad, N.; Hadef, A.; Hageböck, S.; Hagihara, M.; Hakobyan, H.; Haleem, M.; Haley, J.; Halladjian, G.; Hallewell, G. D.; Hamacher, K.; Hamal, P.; Hamano, K.; Hamilton, A.; Hamity, G. N.; Hamnett, P. G.; Han, L.; Han, S.; Hanagaki, K.; Hanawa, K.; Hance, M.; Handl, D. M.; Haney, B.; Hanke, P.; Hansen, J. B.; Hansen, J. D.; Hansen, M. C.; Hansen, P. H.; Hara, K.; Hard, A. S.; Harenberg, T.; Hariri, F.; Harkusha, S.; Harrison, P. F.; Hartmann, N. M.; Hasegawa, Y.; Hasib, A.; Hassani, S.; Haug, S.; Hauser, R.; Hauswald, L.; Havener, L. B.; Havranek, M.; Hawkes, C. M.; Hawkings, R. J.; Hayakawa, D.; Hayden, D.; Hays, C. P.; Hays, J. M.; Hayward, H. S.; Haywood, S. J.; Head, S. J.; Heck, T.; Hedberg, V.; Heelan, L.; Heer, S.; Heidegger, K. K.; Heim, S.; Heim, T.; Heinemann, B.; Heinrich, J. J.; Heinrich, L.; Heinz, C.; Hejbal, J.; Helary, L.; Held, A.; Hellman, S.; Helsens, C.; Henderson, R. C. W.; Heng, Y.; Henkelmann, S.; Henriques Correia, A. M.; Henrot-Versille, S.; Herbert, G. H.; Herde, H.; Herget, V.; Hernández Jiménez, Y.; Herr, H.; Herten, G.; Hertenberger, R.; Hervas, L.; Herwig, T. C.; Hesketh, G. G.; Hessey, N. P.; Hetherly, J. W.; Higashino, S.; Higón-Rodriguez, E.; Hildebrand, K.; Hill, E.; Hill, J. C.; Hiller, K. H.; Hillier, S. J.; Hils, M.; Hinchliffe, I.; Hirose, M.; Hirschbuehl, D.; Hiti, B.; Hladik, O.; Hlaluku, D. R.; Hoad, X.; Hobbs, J.; Hod, N.; Hodgkinson, M. C.; Hodgson, P.; Hoecker, A.; Hoeferkamp, M. R.; Hoenig, F.; Hohn, D.; Holmes, T. R.; Holzbock, M.; Homann, M.; Honda, S.; Honda, T.; Hong, T. M.; Hooberman, B. H.; Hopkins, W. H.; Horii, Y.; Horton, A. J.; Hostachy, J.-Y.; Hostiuc, A.; Hou, S.; Hoummada, A.; Howarth, J.; Hoya, J.; Hrabovsky, M.; Hrdinka, J.; Hristova, I.; Hrivnac, J.; Hryn'ova, T.; Hrynevich, A.; Hsu, P. J.; Hsu, S.-C.; Hu, Q.; Hu, S.; Huang, Y.; Hubacek, Z.; Hubaut, F.; Huegging, F.; Huffman, T. B.; Hughes, E. W.; Huhtinen, M.; Hunter, R. F. H.; Huo, P.; Huseynov, N.; Huston, J.; Huth, J.; Hyneman, R.; Iacobucci, G.; Iakovidis, G.; Ibragimov, I.; Iconomidou-Fayard, L.; Idrissi, Z.; Iengo, P.; Igonkina, O.; Iizawa, T.; Ikegami, Y.; Ikeno, M.; Ilchenko, Y.; Iliadis, D.; Ilic, N.; Iltzsche, F.; Introzzi, G.; Ioannou, P.; Iodice, M.; Iordanidou, K.; Ippolito, V.; Isacson, M. F.; Ishijima, N.; Ishino, M.; Ishitsuka, M.; Issever, C.; Istin, S.; Ito, F.; Iturbe Ponce, J. M.; Iuppa, R.; Iwasaki, H.; Izen, J. M.; Izzo, V.; Jabbar, S.; Jackson, P.; Jacobs, R. M.; Jain, V.; Jakobi, K. B.; Jakobs, K.; Jakobsen, S.; Jakoubek, T.; Jamin, D. O.; Jana, D. K.; Jansky, R.; Janssen, J.; Janus, M.; Janus, P. A.; Jarlskog, G.; Javadov, N.; Javůrek, T.; Javurkova, M.; Jeanneau, F.; Jeanty, L.; Jejelava, J.; Jelinskas, A.; Jenni, P.; Jeske, C.; Jézéquel, S.; Ji, H.; Jia, J.; Jiang, H.; Jiang, Y.; Jiang, Z.; Jiggins, S.; Jimenez Pena, J.; Jin, S.; Jinaru, A.; Jinnouchi, O.; Jivan, H.; Johansson, P.; Johns, K. A.; Johnson, C. A.; Johnson, W. J.; Jon-And, K.; Jones, R. W. L.; Jones, S. D.; Jones, S.; Jones, T. J.; Jongmanns, J.; Jorge, P. M.; Jovicevic, J.; Ju, X.; Juste Rozas, A.; Köhler, M. K.; Kaczmarska, A.; Kado, M.; Kagan, H.; Kagan, M.; Kahn, S. J.; Kaji, T.; Kajomovitz, E.; Kalderon, C. W.; Kaluza, A.; Kama, S.; Kamenshchikov, A.; Kanaya, N.; Kanjir, L.; Kantserov, V. A.; Kanzaki, J.; Kaplan, B.; Kaplan, L. S.; Kar, D.; Karakostas, K.; Karastathis, N.; Kareem, M. J.; Karentzos, E.; Karpov, S. N.; Karpova, Z. M.; Karthik, K.; Kartvelishvili, V.; Karyukhin, A. N.; Kasahara, K.; Kashif, L.; Kass, R. D.; Kastanas, A.; Kataoka, Y.; Kato, C.; Katre, A.; Katzy, J.; Kawade, K.; Kawagoe, K.; Kawamoto, T.; Kawamura, G.; Kay, E. F.; Kazanin, V. F.; Keeler, R.; Kehoe, R.; Keller, J. S.; Kellermann, E.; Kempster, J. J.; Kendrick, J.; Keoshkerian, H.; Kepka, O.; Kerševan, B. P.; Kersten, S.; Keyes, R. A.; Khader, M.; Khalil-zada, F.; Khanov, A.; Kharlamov, A. G.; Kharlamova, T.; Khodinov, A.; Khoo, T. J.; Khovanskiy, V.; Khramov, E.; Khubua, J.; Kido, S.; Kilby, C. R.; Kim, H. Y.; Kim, S. H.; Kim, Y. K.; Kimura, N.; Kind, O. M.; King, B. T.; Kirchmeier, D.; Kirk, J.; Kiryunin, A. E.; Kishimoto, T.; Kisielewska, D.; Kitali, V.; Kivernyk, O.; Kladiva, E.; Klapdor-Kleingrothaus, T.; Klein, M. H.; Klein, M.; Klein, U.; Kleinknecht, K.; Klimek, P.; Klimentov, A.; Klingenberg, R.; Klingl, T.; Klioutchnikova, T.; Klitzner, F. F.; Kluge, E.-E.; Kluit, P.; Kluth, S.; Kneringer, E.; Knoops, E. B. F. G.; Knue, A.; Kobayashi, A.; Kobayashi, D.; Kobayashi, T.; Kobel, M.; Kocian, M.; Kodys, P.; Koffas, T.; Koffeman, E.; Köhler, N. M.; Koi, T.; Kolb, M.; Koletsou, I.; Kondo, T.; Kondrashova, N.; Köneke, K.; König, A. C.; Kono, T.; Konoplich, R.; Konstantinidis, N.; Konya, B.; Kopeliansky, R.; Koperny, S.; Kopp, A. K.; Korcyl, K.; Kordas, K.; Korn, A.; Korol, A. A.; Korolkov, I.; Korolkova, E. V.; Kortner, O.; Kortner, S.; Kosek, T.; Kostyukhin, V. V.; Kotwal, A.; Koulouris, A.; Kourkoumeli-Charalampidi, A.; Kourkoumelis, C.; Kourlitis, E.; Kouskoura, V.; Kowalewska, A. B.; Kowalewski, R.; Kowalski, T. Z.; Kozakai, C.; Kozanecki, W.; Kozhin, A. S.; Kramarenko, V. A.; Kramberger, G.; Krasnopevtsev, D.; Krasny, M. W.; Krasznahorkay, A.; Krauss, D.; Kremer, J. A.; Kretzschmar, J.; Kreutzfeldt, K.; Krieger, P.; Krizka, K.; Kroeninger, K.; Kroha, H.; Kroll, J.; Kroll, J.; Kroseberg, J.; Krstic, J.; Kruchonak, U.; Krüger, H.; Krumnack, N.; Kruse, M. C.; Kubota, T.; Kucuk, H.; Kuday, S.; Kuechler, J. T.; Kuehn, S.; Kugel, A.; Kuger, F.; Kuhl, T.; Kukhtin, V.; Kukla, R.; Kulchitsky, Y.; Kuleshov, S.; Kulinich, Y. P.; Kuna, M.; Kunigo, T.; Kupco, A.; Kupfer, T.; Kuprash, O.; Kurashige, H.; Kurchaninov, L. L.; Kurochkin, Y. A.; Kurth, M. G.; Kuwertz, E. S.; Kuze, M.; Kvita, J.; Kwan, T.; Kyriazopoulos, D.; La Rosa, A.; La Rosa Navarro, J. L.; La Rotonda, L.; La Ruffa, F.; Lacasta, C.; Lacava, F.; Lacey, J.; Lack, D. P. J.; Lacker, H.; Lacour, D.; Ladygin, E.; Lafaye, R.; Laforge, B.; Lagouri, T.; Lai, S.; Lammers, S.; Lampl, W.; Lançon, E.; Landgraf, U.; Landon, M. P. J.; Lanfermann, M. C.; Lang, V. S.; Lange, J. C.; Langenberg, R. J.; Lankford, A. J.; Lanni, F.; Lantzsch, K.; Lanza, A.; Lapertosa, A.; Laplace, S.; Laporte, J. F.; Lari, T.; Lasagni Manghi, F.; Lassnig, M.; Lau, T. S.; Laurelli, P.; Lavrijsen, W.; Law, A. T.; Laycock, P.; Lazovich, T.; Lazzaroni, M.; Le, B.; Le Dortz, O.; Le Guirriec, E.; Le Quilleuc, E. P.; LeBlanc, M.; LeCompte, T.; Ledroit-Guillon, F.; Lee, C. A.; Lee, G. R.; Lee, S. C.; Lee, L.; Lefebvre, B.; Lefebvre, G.; Lefebvre, M.; Legger, F.; Leggett, C.; Lehmann Miotto, G.; Lei, X.; Leight, W. A.; Leite, M. A. L.; Leitner, R.; Lellouch, D.; Lemmer, B.; Leney, K. J. C.; Lenz, T.; Lenzi, B.; Leone, R.; Leone, S.; Leonidopoulos, C.; Lerner, G.; Leroy, C.; Les, R.; Lesage, A. A. J.; Lester, C. G.; Levchenko, M.; Levêque, J.; Levin, D.; Levinson, L. J.; Levy, M.; Lewis, D.; Li, B.; Li, H.; Li, L.; Li, Q.; Li, Q.; Li, S.; Li, X.; Li, Y.; Liang, Z.; Liberti, B.; Liblong, A.; Lie, K.; Liebal, J.; Liebig, W.; Limosani, A.; Lin, C. Y.; Lin, K.; Lin, S. C.; Lin, T. H.; Linck, R. A.; Lindquist, B. E.; Lionti, A. E.; Lipeles, E.; Lipniacka, A.; Lisovyi, M.; Liss, T. M.; Lister, A.; Litke, A. M.; Liu, B.; Liu, H.; Liu, H.; Liu, J. K. K.; Liu, J.; Liu, J. B.; Liu, K.; Liu, L.; Liu, M.; Liu, Y. L.; Liu, Y.; Livan, M.; Lleres, A.; Llorente Merino, J.; Lloyd, S. L.; Lo, C. Y.; Lo Sterzo, F.; Lobodzinska, E. M.; Loch, P.; Loebinger, F. K.; Loesle, A.; Loew, K. M.; Lohse, T.; Lohwasser, K.; Lokajicek, M.; Long, B. A.; Long, J. D.; Long, R. E.; Longo, L.; Looper, K. A.; Lopez, J. A.; Lopez Paz, I.; Lopez Solis, A.; Lorenz, J.; Lorenzo Martinez, N.; Losada, M.; Lösel, P. J.; Lou, X.; Lounis, A.; Love, J.; Love, P. A.; Lu, H.; Lu, N.; Lu, Y. J.; Lubatti, H. J.; Luci, C.; Lucotte, A.; Luedtke, C.; Luehring, F.; Lukas, W.; Luminari, L.; Lundberg, O.; Lund-Jensen, B.; Lutz, M. S.; Luzi, P. M.; Lynn, D.; Lysak, R.; Lytken, E.; Lyu, F.; Lyubushkin, V.; Ma, H.; Ma, L. L.; Ma, Y.; Maccarrone, G.; Macchiolo, A.; Macdonald, C. M.; Maček, B.; Machado Miguens, J.; Madaffari, D.; Madar, R.; Mader, W. F.; Madsen, A.; Madysa, N.; Maeda, J.; Maeland, S.; Maeno, T.; Maevskiy, A. S.; Magerl, V.; Maiani, C.; Maidantchik, C.; Maier, T.; Maio, A.; Majersky, O.; Majewski, S.; Makida, Y.; Makovec, N.; Malaescu, B.; Malecki, Pa.; Maleev, V. P.; Malek, F.; Mallik, U.; Malon, D.; Malone, C.; Maltezos, S.; Malyukov, S.; Mamuzic, J.; Mancini, G.; Mandić, I.; Maneira, J.; Manhaes de Andrade Filho, L.; Manjarres Ramos, J.; Mankinen, K. H.; Mann, A.; Manousos, A.; Mansoulie, B.; Mansour, J. D.; Mantifel, R.; Mantoani, M.; Manzoni, S.; Mapelli, L.; Marceca, G.; March, L.; Marchese, L.; Marchiori, G.; Marcisovsky, M.; Marin Tobon, C. A.; Marjanovic, M.; Marley, D. E.; Marroquim, F.; Marsden, S. P.; Marshall, Z.; Martensson, M. U. F.; Marti-Garcia, S.; Martin, C. B.; Martin, T. A.; Martin, V. J.; Martin dit Latour, B.; Martinez, M.; Martinez Outschoorn, V. I.; Martin-Haugh, S.; Martoiu, V. S.; Martyniuk, A. C.; Marzin, A.; Masetti, L.; Mashimo, T.; Mashinistov, R.; Masik, J.; Maslennikov, A. L.; Mason, L. H.; Massa, L.; Mastrandrea, P.; Mastroberardino, A.; Masubuchi, T.; Mättig, P.; Maurer, J.; Maxfield, S. J.; Maximov, D. A.; Mazini, R.; Maznas, I.; Mazza, S. M.; Mc Fadden, N. C.; Mc Goldrick, G.; Mc Kee, S. P.; McCarn, A.; McCarthy, R. L.; McCarthy, T. G.; McClymont, L. I.; McDonald, E. F.; Mcfayden, J. A.; Mchedlidze, G.; McMahon, S. J.; McNamara, P. C.; McNicol, C. J.; McPherson, R. A.; Meehan, S.; Megy, T. J.; Mehlhase, S.; Mehta, A.; Meideck, T.; Meier, K.; Meirose, B.; Melini, D.; Mellado Garcia, B. R.; Mellenthin, J. D.; Melo, M.; Meloni, F.; Melzer, A.; Menary, S. B.; Meng, L.; Meng, X. T.; Mengarelli, A.; Menke, S.; Meoni, E.; Mergelmeyer, S.; Merlassino, C.; Mermod, P.; Merola, L.; Meroni, C.; Merritt, F. S.; Messina, A.; Metcalfe, J.; Mete, A. S.; Meyer, C.; Meyer, J.-P.; Meyer, J.; Meyer Zu Theenhausen, H.; Miano, F.; Middleton, R. P.; Miglioranzi, S.; Mijović, L.; Mikenberg, G.; Mikestikova, M.; Mikuž, M.; Milesi, M.; Milic, A.; Millar, D. A.; Miller, D. W.; Mills, C.; Milov, A.; Milstead, D. A.; Minaenko, A. A.; Minami, Y.; Minashvili, I. A.; Mincer, A. I.; Mindur, B.; Mineev, M.; Minegishi, Y.; Ming, Y.; Mir, L. M.; Mirto, A.; Mistry, K. P.; Mitani, T.; Mitrevski, J.; Mitsou, V. A.; Miucci, A.; Miyagawa, P. S.; Mizukami, A.; Mjörnmark, J. U.; Mkrtchyan, T.; Mlynarikova, M.; Moa, T.; Mochizuki, K.; Mogg, P.; Mohapatra, S.; Molander, S.; Moles-Valls, R.; Mondragon, M. C.; Mönig, K.; Monk, J.; Monnier, E.; Montalbano, A.; Montejo Berlingen, J.; Monticelli, F.; Monzani, S.; Moore, R. W.; Morange, N.; Moreno, D.; Moreno Llácer, M.; Morettini, P.; Morgenstern, S.; Mori, D.; Mori, T.; Morii, M.; Morinaga, M.; Morisbak, V.; Morley, A. K.; Mornacchi, G.; Morris, J. D.; Morvaj, L.; Moschovakos, P.; Mosidze, M.; Moss, H. J.; Moss, J.; Motohashi, K.; Mount, R.; Mountricha, E.; Moyse, E. J. W.; Muanza, S.; Mueller, F.; Mueller, J.; Mueller, R. S. P.; Muenstermann, D.; Mullen, P.; Mullier, G. A.; Munoz Sanchez, F. J.; Murray, W. J.; Musheghyan, H.; Muškinja, M.; Myagkov, A. G.; Myska, M.; Nachman, B. P.; Nackenhorst, O.; Nagai, K.; Nagai, R.; Nagano, K.; Nagasaka, Y.; Nagata, K.; Nagel, M.; Nagy, E.; Nairz, A. M.; Nakahama, Y.; Nakamura, K.; Nakamura, T.; Nakano, I.; Naranjo Garcia, R. F.; Narayan, R.; Narrias Villar, D. I.; Naryshkin, I.; Naumann, T.; Navarro, G.; Nayyar, R.; Neal, H. A.; Nechaeva, P. Yu.; Neep, T. J.; Negri, A.; Negrini, M.; Nektarijevic, S.; Nellist, C.; Nelson, A.; Nelson, M. E.; Nemecek, S.; Nemethy, P.; Nessi, M.; Neubauer, M. S.; Neumann, M.; Newman, P. R.; Ng, T. Y.; Ng, Y. S.; Nguyen Manh, T.; Nickerson, R. B.; Nicolaidou, R.; Nielsen, J.; Nikiforou, N.; Nikolaenko, V.; Nikolic-Audit, I.; Nikolopoulos, K.; Nilsson, P.; Ninomiya, Y.; Nisati, A.; Nishu, N.; Nisius, R.; Nitsche, I.; Nitta, T.; Nobe, T.; Noguchi, Y.; Nomachi, M.; Nomidis, I.; Nomura, M. A.; Nooney, T.; Nordberg, M.; Norjoharuddeen, N.; Novgorodova, O.; Nozaki, M.; Nozka, L.; Ntekas, K.; Nurse, E.; Nuti, F.; O'connor, K.; O'Neil, D. C.; O'Rourke, A. A.; O'Shea, V.; Oakham, F. G.; Oberlack, H.; Obermann, T.; Ocariz, J.; Ochi, A.; Ochoa, I.; Ochoa-Ricoux, J. P.; Oda, S.; Odaka, S.; Oh, A.; Oh, S. H.; Ohm, C. C.; Ohman, H.; Oide, H.; Okawa, H.; Okumura, Y.; Okuyama, T.; Olariu, A.; Oleiro Seabra, L. F.; Olivares Pino, S. A.; Oliveira Damazio, D.; Olsson, M. J. R.; Olszewski, A.; Olszowska, J.; Onofre, A.; Onogi, K.; Onyisi, P. U. E.; Oppen, H.; Oreglia, M. J.; Oren, Y.; Orestano, D.; Orlando, N.; Orr, R. S.; Osculati, B.; Ospanov, R.; Otero y Garzon, G.; Otono, H.; Ouchrif, M.; Ould-Saada, F.; Ouraou, A.; Oussoren, K. P.; Ouyang, Q.; Owen, M.; Owen, R. E.; Ozcan, V. E.; Ozturk, N.; Pachal, K.; Pacheco Pages, A.; Pacheco Rodriguez, L.; Padilla Aranda, C.; Pagan Griso, S.; Paganini, M.; Paige, F.; Palacino, G.; Palazzo, S.; Palestini, S.; Palka, M.; Pallin, D.; Panagiotopoulou, E. St.; Panagoulias, I.; Pandini, C. E.; Panduro Vazquez, J. G.; Pani, P.; Panitkin, S.; Pantea, D.; Paolozzi, L.; Papadopoulou, Th. D.; Papageorgiou, K.; Paramonov, A.; Paredes Hernandez, D.; Parker, A. J.; Parker, M. A.; Parker, K. A.; Parodi, F.; Parsons, J. A.; Parzefall, U.; Pascuzzi, V. R.; Pasner, J. M.; Pasqualucci, E.; Passaggio, S.; Pastore, Fr.; Pataraia, S.; Pater, J. R.; Pauly, T.; Pearson, B.; Pedraza Lopez, S.; Pedro, R.; Peleganchuk, S. V.; Penc, O.; Peng, C.; Peng, H.; Penwell, J.; Peralva, B. S.; Perego, M. M.; Perepelitsa, D. V.; Peri, F.; Perini, L.; Pernegger, H.; Perrella, S.; Peschke, R.; Peshekhonov, V. D.; Peters, K.; Peters, R. F. Y.; Petersen, B. A.; Petersen, T. C.; Petit, E.; Petridis, A.; Petridou, C.; Petroff, P.; Petrolo, E.; Petrov, M.; Petrucci, F.; Pettersson, N. E.; Peyaud, A.; Pezoa, R.; Phillips, F. H.; Phillips, P. W.; Piacquadio, G.; Pianori, E.; Picazio, A.; Pickering, M. A.; Piegaia, R.; Pilcher, J. E.; Pilkington, A. D.; Pinamonti, M.; Pinfold, J. L.; Pirumov, H.; Pitt, M.; Plazak, L.; Pleier, M.-A.; Pleskot, V.; Plotnikova, E.; Pluth, D.; Podberezko, P.; Poettgen, R.; Poggi, R.; Poggioli, L.; Pogrebnyak, I.; Pohl, D.; Pokharel, I.; Polesello, G.; Poley, A.; Policicchio, A.; Polifka, R.; Polini, A.; Pollard, C. S.; Polychronakos, V.; Pommès, K.; Ponomarenko, D.; Pontecorvo, L.; Popeneciu, G. A.; Portillo Quintero, D. M.; Pospisil, S.; Potamianos, K.; Potrap, I. N.; Potter, C. J.; Potti, H.; Poulsen, T.; Poveda, J.; Pozo Astigarraga, M. E.; Pralavorio, P.; Pranko, A.; Prell, S.; Price, D.; Primavera, M.; Prince, S.; Proklova, N.; Prokofiev, K.; Prokoshin, F.; Protopopescu, S.; Proudfoot, J.; Przybycien, M.; Puri, A.; Puzo, P.; Qian, J.; Qin, G.; Qin, Y.; Quadt, A.; Queitsch-Maitland, M.; Quilty, D.; Raddum, S.; Radeka, V.; Radescu, V.; Radhakrishnan, S. K.; Radloff, P.; Rados, P.; Ragusa, F.; Rahal, G.; Raine, J. A.; Rajagopalan, S.; Rangel-Smith, C.; Rashid, T.; Raspopov, S.; Ratti, M. G.; Rauch, D. M.; Rauscher, F.; Rave, S.; Ravinovich, I.; Rawling, J. H.; Raymond, M.; Read, A. L.; Readioff, N. P.; Reale, M.; Rebuzzi, D. M.; Redelbach, A.; Redlinger, G.; Reece, R.; Reed, R. G.; Reeves, K.; Rehnisch, L.; Reichert, J.; Reiss, A.; Rembser, C.; Ren, H.; Rescigno, M.; Resconi, S.; Resseguie, E. D.; Rettie, S.; Reynolds, E.; Rezanova, O. L.; Reznicek, P.; Rezvani, R.; Richter, R.; Richter, S.; Richter-Was, E.; Ricken, O.; Ridel, M.; Rieck, P.; Riegel, C. J.; Rieger, J.; Rifki, O.; Rijssenbeek, M.; Rimoldi, A.; Rimoldi, M.; Rinaldi, L.; Ripellino, G.; Ristić, B.; Ritsch, E.; Riu, I.; Rizatdinova, F.; Rizvi, E.; Rizzi, C.; Roberts, R. T.; Robertson, S. H.; Robichaud-Veronneau, A.; Robinson, D.; Robinson, J. E. M.; Robson, A.; Rocco, E.; Roda, C.; Rodina, Y.; Rodriguez Bosca, S.; Rodriguez Perez, A.; Rodriguez Rodriguez, D.; Roe, S.; Rogan, C. S.; Røhne, O.; Roloff, J.; Romaniouk, A.; Romano, M.; Romano Saez, S. M.; Romero Adam, E.; Rompotis, N.; Ronzani, M.; Roos, L.; Rosati, S.; Rosbach, K.; Rose, P.; Rosien, N.-A.; Rossi, E.; Rossi, L. P.; Rosten, J. H. N.; Rosten, R.; Rotaru, M.; Rothberg, J.; Rousseau, D.; Roy, D.; Rozanov, A.; Rozen, Y.; Ruan, X.; Rubbo, F.; Rühr, F.; Ruiz-Martinez, A.; Rurikova, Z.; Rusakovich, N. A.; Russell, H. L.; Rutherfoord, J. P.; Ruthmann, N.; Rüttinger, E. M.; Ryabov, Y. F.; Rybar, M.; Rybkin, G.; Ryu, S.; Ryzhov, A.; Rzehorz, G. F.; Saavedra, A. F.; Sabato, G.; Sacerdoti, S.; Sadrozinski, H. F.-W.; Sadykov, R.; Safai Tehrani, F.; Saha, P.; Sahinsoy, M.; Saimpert, M.; Saito, M.; Saito, T.; Sakamoto, H.; Sakurai, Y.; Salamanna, G.; Salazar Loyola, J. E.; Salek, D.; Sales De Bruin, P. H.; Salihagic, D.; Salnikov, A.; Salt, J.; Salvatore, D.; Salvatore, F.; Salvucci, A.; Salzburger, A.; Sammel, D.; Sampsonidis, D.; Sampsonidou, D.; Sánchez, J.; Sanchez Martinez, V.; Sanchez Pineda, A.; Sandaker, H.; Sandbach, R. L.; Sander, C. O.; Sandhoff, M.; Sandoval, C.; Sankey, D. P. C.; Sannino, M.; Sano, Y.; Sansoni, A.; Santoni, C.; Santos, H.; Santoyo Castillo, I.; Sapronov, A.; Saraiva, J. G.; Sarrazin, B.; Sasaki, O.; Sato, K.; Sauvan, E.; Savage, G.; Savard, P.; Savic, N.; Sawyer, C.; Sawyer, L.; Saxon, J.; Sbarra, C.; Sbrizzi, A.; Scanlon, T.; Scannicchio, D. A.; Schaarschmidt, J.; Schacht, P.; Schachtner, B. M.; Schaefer, D.; Schaefer, L.; Schaefer, R.; Schaeffer, J.; Schaepe, S.; Schaetzel, S.; Schäfer, U.; Schaffer, A. C.; Schaile, D.; Schamberger, R. D.; Schegelsky, V. A.; Scheirich, D.; Schenck, F.; Schernau, M.; Schiavi, C.; Schier, S.; Schildgen, L. K.; Schillo, C.; Schioppa, M.; Schlenker, S.; Schmidt-Sommerfeld, K. R.; Schmieden, K.; Schmitt, C.; Schmitt, S.; Schmitz, S.; Schnoor, U.; Schoeffel, L.; Schoening, A.; Schoenrock, B. D.; Schopf, E.; Schott, M.; Schouwenberg, J. F. P.; Schovancova, J.; Schramm, S.; Schuh, N.; Schulte, A.; Schultens, M. J.; Schultz-Coulon, H.-C.; Schulz, H.; Schumacher, M.; Schumm, B. A.; Schune, Ph.; Schwartzman, A.; Schwarz, T. A.; Schweiger, H.; Schwemling, Ph.; Schwienhorst, R.; Schwindling, J.; Sciandra, A.; Sciolla, G.; Scornajenghi, M.; Scuri, F.; Scutti, F.; Searcy, J.; Seema, P.; Seidel, S. C.; Seiden, A.; Seixas, J. M.; Sekhniaidze, G.; Sekhon, K.; Sekula, S. J.; Semprini-Cesari, N.; Senkin, S.; Serfon, C.; Serin, L.; Serkin, L.; Sessa, M.; Seuster, R.; Severini, H.; Šfiligoj, T.; Sforza, F.; Sfyrla, A.; Shabalina, E.; Shaikh, N. W.; Shan, L. Y.; Shang, R.; Shank, J. T.; Shapiro, M.; Shatalov, P. B.; Shaw, K.; Shaw, S. M.; Shcherbakova, A.; Shehu, C. Y.; Shen, Y.; Sherafati, N.; Sherman, A. D.; Sherwood, P.; Shi, L.; Shimizu, S.; Shimmin, C. O.; Shimojima, M.; Shipsey, I. P. J.; Shirabe, S.; Shiyakova, M.; Shlomi, J.; Shmeleva, A.; Shoaleh Saadi, D.; Shochet, M. J.; Shojaii, S.; Shope, D. R.; Shrestha, S.; Shulga, E.; Shupe, M. A.; Sicho, P.; Sickles, A. M.; Sidebo, P. E.; Sideras Haddad, E.; Sidiropoulou, O.; Sidoti, A.; Siegert, F.; Sijacki, Dj.; Silva, J.; Silverstein, S. B.; Simak, V.; Simic, L.; Simion, S.; Simioni, E.; Simmons, B.; Simon, M.; Sinervo, P.; Sinev, N. B.; Sioli, M.; Siragusa, G.; Siral, I.; Sivoklokov, S. Yu.; Sjölin, J.; Skinner, M. B.; Skubic, P.; Slater, M.; Slavicek, T.; Slawinska, M.; Sliwa, K.; Slovak, R.; Smakhtin, V.; Smart, B. H.; Smiesko, J.; Smirnov, N.; Smirnov, S. Yu.; Smirnov, Y.; Smirnova, L. N.; Smirnova, O.; Smith, J. W.; Smith, M. N. K.; Smith, R. W.; Smizanska, M.; Smolek, K.; Snesarev, A. A.; Snyder, I. M.; Snyder, S.; Sobie, R.; Socher, F.; Soffer, A.; Søgaard, A.; Soh, D. A.; Sokhrannyi, G.; Solans Sanchez, C. A.; Solar, M.; Soldatov, E. Yu.; Soldevila, U.; Solodkov, A. A.; Soloshenko, A.; Solovyanov, O. V.; Solovyev, V.; Sommer, P.; Son, H.; Sopczak, A.; Sosa, D.; Sotiropoulou, C. L.; Sottocornola, S.; Soualah, R.; Soukharev, A. M.; South, D.; Sowden, B. C.; Spagnolo, S.; Spalla, M.; Spangenberg, M.; Spanò, F.; Sperlich, D.; Spettel, F.; Spieker, T. M.; Spighi, R.; Spigo, G.; Spiller, L. A.; Spousta, M.; Denis, R. D. St.; Stabile, A.; Stamen, R.; Stamm, S.; Stanecka, E.; Stanek, R. W.; Stanescu, C.; Stanitzki, M. M.; Stapf, B. S.; Stapnes, S.; Starchenko, E. A.; Stark, G. H.; Stark, J.; Stark, S. H.; Staroba, P.; Starovoitov, P.; Stärz, S.; Staszewski, R.; Stegler, M.; Steinberg, P.; Stelzer, B.; Stelzer, H. J.; Stelzer-Chilton, O.; Stenzel, H.; Stevenson, T. J.; Stewart, G. A.; Stockton, M. C.; Stoebe, M.; Stoicea, G.; Stolte, P.; Stonjek, S.; Stradling, A. R.; Straessner, A.; Stramaglia, M. E.; Strandberg, J.; Strandberg, S.; Strauss, M.; Strizenec, P.; Ströhmer, R.; Strom, D. M.; Stroynowski, R.; Strubig, A.; Stucci, S. A.; Stugu, B.; Styles, N. A.; Su, D.; Su, J.; Suchek, S.; Sugaya, Y.; Suk, M.; Sulin, V. V.; Sultan, DMS; Sultansoy, S.; Sumida, T.; Sun, S.; Sun, X.; Suruliz, K.; Suster, C. J. E.; Sutton, M. R.; Suzuki, S.; Svatos, M.; Swiatlowski, M.; Swift, S. P.; Sykora, I.; Sykora, T.; Ta, D.; Tackmann, K.; Taenzer, J.; Taffard, A.; Tafirout, R.; Tahirovic, E.; Taiblum, N.; Takai, H.; Takashima, R.; Takasugi, E. H.; Takeda, K.; Takeshita, T.; Takubo, Y.; Talby, M.; Talyshev, A. A.; Tanaka, J.; Tanaka, M.; Tanaka, R.; Tanaka, S.; Tanioka, R.; Tannenwald, B. B.; Tapia Araya, S.; Tapprogge, S.; Tarem, S.; Tartarelli, G. F.; Tas, P.; Tasevsky, M.; Tashiro, T.; Tassi, E.; Tavares Delgado, A.; Tayalati, Y.; Taylor, A. C.; Taylor, A. J.; Taylor, G. N.; Taylor, P. T. E.; Taylor, W.; Teixeira-Dias, P.; Temple, D.; Ten Kate, H.; Teng, P. K.; Teoh, J. J.; Tepel, F.; Terada, S.; Terashi, K.; Terron, J.; Terzo, S.; Testa, M.; Teuscher, R. J.; Thais, S. J.; Theveneaux-Pelzer, T.; Thiele, F.; Thomas, J. P.; Thomas-Wilsker, J.; Thompson, P. D.; Thompson, A. S.; Thomsen, L. A.; Thomson, E.; Tian, Y.; Tibbetts, M. J.; Ticse Torres, R. E.; Tikhomirov, V. O.; Tikhonov, Yu. A.; Timoshenko, S.; Tipton, P.; Tisserant, S.; Todome, K.; Todorova-Nova, S.; Todt, S.; Tojo, J.; Tokár, S.; Tokushuku, K.; Tolley, E.; Tomlinson, L.; Tomoto, M.; Tompkins, L.; Toms, K.; Tong, B.; Tornambe, P.; Torrence, E.; Torres, H.; Torró Pastor, E.; Toth, J.; Touchard, F.; Tovey, D. R.; Treado, C. J.; Trefzger, T.; Tresoldi, F.; Tricoli, A.; Trigger, I. M.; Trincaz-Duvoid, S.; Tripiana, M. F.; Trischuk, W.; Trocmé, B.; Trofymov, A.; Troncon, C.; Trottier-McDonald, M.; Trovatelli, M.; Truong, L.; Trzebinski, M.; Trzupek, A.; Tsang, K. W.; Tseng, J. C.-L.; Tsiareshka, P. V.; Tsirintanis, N.; Tsiskaridze, S.; Tsiskaridze, V.; Tskhadadze, E. G.; Tsukerman, I. I.; Tsulaia, V.; Tsuno, S.; Tsybychev, D.; Tu, Y.; Tudorache, A.; Tudorache, V.; Tulbure, T. T.; Tuna, A. N.; Turchikhin, S.; Turgeman, D.; Turk Cakir, I.; Turra, R.; Tuts, P. M.; Ucchielli, G.; Ueda, I.; Ughetto, M.; Ukegawa, F.; Unal, G.; Undrus, A.; Unel, G.; Ungaro, F. C.; Unno, Y.; Uno, K.; Unverdorben, C.; Urban, J.; Urquijo, P.; Urrejola, P.; Usai, G.; Usui, J.; Vacavant, L.; Vacek, V.; Vachon, B.; Vadla, K. O. H.; Vaidya, A.; Valderanis, C.; Valdes Santurio, E.; Valente, M.; Valentinetti, S.; Valero, A.; Valéry, L.; Valkar, S.; Vallier, A.; Valls Ferrer, J. A.; Van Den Wollenberg, W.; van der Graaf, H.; van Gemmeren, P.; Van Nieuwkoop, J.; van Vulpen, I.; van Woerden, M. C.; Vanadia, M.; Vandelli, W.; Vaniachine, A.; Vankov, P.; Vardanyan, G.; Vari, R.; Varnes, E. W.; Varni, C.; Varol, T.; Varouchas, D.; Vartapetian, A.; Varvell, K. E.; Vasquez, J. G.; Vasquez, G. A.; Vazeille, F.; Vazquez Furelos, D.; Vazquez Schroeder, T.; Veatch, J.; Veeraraghavan, V.; Veloce, L. M.; Veloso, F.; Veneziano, S.; Ventura, A.; Venturi, M.; Venturi, N.; Venturini, A.; Vercesi, V.; Verducci, M.; Verkerke, W.; Vermeulen, A. T.; Vermeulen, J. C.; Vetterli, M. C.; Viaux Maira, N.; Viazlo, O.; Vichou, I.; Vickey, T.; Vickey Boeriu, O. E.; Viehhauser, G. H. A.; Viel, S.; Vigani, L.; Villa, M.; Villaplana Perez, M.; Vilucchi, E.; Vincter, M. G.; Vinogradov, V. B.; Vishwakarma, A.; Vittori, C.; Vivarelli, I.; Vlachos, S.; Vogel, M.; Vokac, P.; Volpi, G.; von der Schmitt, H.; von Toerne, E.; Vorobel, V.; Vorobev, K.; Vos, M.; Voss, R.; Vossebeld, J. H.; Vranjes, N.; Vranjes Milosavljevic, M.; Vrba, V.; Vreeswijk, M.; Vuillermet, R.; Vukotic, I.; Wagner, P.; Wagner, W.; Wagner-Kuhr, J.; Wahlberg, H.; Wahrmund, S.; Wakamiya, K.; Walder, J.; Walker, R.; Walkowiak, W.; Wallangen, V.; Wang, C.; Wang, C.; Wang, F.; Wang, H.; Wang, H.; Wang, J.; Wang, J.; Wang, Q.; Wang, R.-J.; Wang, R.; Wang, S. M.; Wang, T.; Wang, W.; Wang, W.; Wang, Z.; Wanotayaroj, C.; Warburton, A.; Ward, C. P.; Wardrope, D. R.; Washbrook, A.; Watkins, P. M.; Watson, A. T.; Watson, M. F.; Watts, G.; Watts, S.; Waugh, B. M.; Webb, A. F.; Webb, S.; Weber, M. S.; Weber, S. M.; Weber, S. W.; Weber, S. A.; Webster, J. S.; Weidberg, A. R.; Weinert, B.; Weingarten, J.; Weirich, M.; Weiser, C.; Weits, H.; Wells, P. S.; Wenaus, T.; Wengler, T.; Wenig, S.; Wermes, N.; Werner, M. D.; Werner, P.; Wessels, M.; Weston, T. D.; Whalen, K.; Whallon, N. L.; Wharton, A. M.; White, A. S.; White, A.; White, M. J.; White, R.; Whiteson, D.; Whitmore, B. W.; Wickens, F. J.; Wiedenmann, W.; Wielers, M.; Wiglesworth, C.; Wiik-Fuchs, L. A. M.; Wildauer, A.; Wilk, F.; Wilkens, H. G.; Williams, H. H.; Williams, S.; Willis, C.; Willocq, S.; Wilson, J. A.; Wingerter-Seez, I.; Winkels, E.; Winklmeier, F.; Winston, O. J.; Winter, B. T.; Wittgen, M.; Wobisch, M.; Wolf, A.; Wolf, T. M. H.; Wolff, R.; Wolter, M. W.; Wolters, H.; Wong, V. W. S.; Woods, N. L.; Worm, S. D.; Wosiek, B. K.; Wotschack, J.; Wozniak, K. W.; Wu, M.; Wu, S. L.; Wu, X.; Wu, Y.; Wyatt, T. R.; Wynne, B. M.; Xella, S.; Xi, Z.; Xia, L.; Xu, D.; Xu, L.; Xu, T.; Xu, W.; Yabsley, B.; Yacoob, S.; Yamaguchi, D.; Yamaguchi, Y.; Yamamoto, A.; Yamamoto, S.; Yamanaka, T.; Yamane, F.; Yamatani, M.; Yamazaki, T.; Yamazaki, Y.; Yan, Z.; Yang, H.; Yang, H.; Yang, Y.; Yang, Z.; Yao, W.-M.; Yap, Y. C.; Yasu, Y.; Yatsenko, E.; Yau Wong, K. H.; Ye, J.; Ye, S.; Yeletskikh, I.; Yigitbasi, E.; Yildirim, E.; Yorita, K.; Yoshihara, K.; Young, C.; Young, C. J. S.; Yu, J.; Yu, J.; Yuen, S. P. Y.; Yusuff, I.; Zabinski, B.; Zacharis, G.; Zaidan, R.; Zaitsev, A. M.; Zakharchuk, N.; Zalieckas, J.; Zaman, A.; Zambito, S.; Zanzi, D.; Zeitnitz, C.; Zemaityte, G.; Zemla, A.; Zeng, J. C.; Zeng, Q.; Zenin, O.; Ženiš, T.; Zerwas, D.; Zhang, D.; Zhang, D.; Zhang, F.; Zhang, G.; Zhang, H.; Zhang, J.; Zhang, L.; Zhang, L.; Zhang, M.; Zhang, P.; Zhang, R.; Zhang, R.; Zhang, X.; Zhang, Y.; Zhang, Z.; Zhao, X.; Zhao, Y.; Zhao, Z.; Zhemchugov, A.; Zhou, B.; Zhou, C.; Zhou, L.; Zhou, M.; Zhou, M.; Zhou, N.; Zhou, Y.; Zhu, C. G.; Zhu, H.; Zhu, J.; Zhu, Y.; Zhuang, X.; Zhukov, K.; Zibell, A.; Zieminska, D.; Zimine, N. I.; Zimmermann, C.; Zimmermann, S.; Zinonos, Z.; Zinser, M.; Ziolkowski, M.; Živković, L.; Zobernig, G.; Zoccoli, A.; Zou, R.; zur Nedden, M.; Zwalinski, L.
2018-01-01
Results of a search for new phenomena in final states with an energetic jet and large missing transverse momentum are reported. The search uses proton-proton collision data corresponding to an integrated luminosity of 36.1 fb-1 at a centre-of-mass energy of 13 TeV collected in 2015 and 2016 with the ATLAS detector at the Large Hadron Collider. Events are required to have at least one jet with a transverse momentum above 250 GeV and no leptons ( e or μ). Several signal regions are considered with increasing requirements on the missing transverse momentum above 250 GeV. Good agreement is observed between the number of events in data and Standard Model predictions. The results are translated into exclusion limits in models with pair-produced weakly interacting dark-matter candidates, large extra spatial dimensions, and supersymmetric particles in several compressed scenarios. [Figure not available: see fulltext.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aaboud, M.; Aad, G.; Abbott, B.
Results of a search for new phenomena in final states with an energetic jet and large missing transverse momentum are reported. The search uses proton-proton collision data corresponding to an integrated luminosity of 36.1 fb -1 at a centre-of-mass energy of 13 TeV collected in 2015 and 2016 with the ATLAS detector at the Large Hadron Collider. Events are required to have at least one jet with a transverse momentum above 250 GeV and no leptons (e or μ). Several signal regions are considered with increasing requirements on the missing transverse momentum above 250 GeV. Good agreement is observed betweenmore » the number of events in data and Standard Model predictions. In conclusion, the results are translated into exclusion limits in models with pair-produced weakly interacting dark-matter candidates, large extra spatial dimensions, and supersymmetric particles in several compressed scenarios.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aaboud, M.; Aad, G.; Abbott, B.
Results of a search for new phenomena in final states with an energetic jet and large missing transverse momentum are reported. The search uses proton-proton collision data corresponding to an integrated luminosity of 36.1 fb –1 at a centre-of-mass energy of 13 TeV collected in 2015 and 2016 with the ATLAS detector at the Large Hadron Collider. Events are required to have at least one jet with a transverse momentum above 250 GeV and no leptons (e or μ). Several signal regions are considered with increasing requirements on the missing transverse momentum above 250 GeV. Good agreement is observed betweenmore » the number of events in data and Standard Model predictions. The results are translated into exclusion limits in models with pair-produced weakly interacting dark-matter candidates, large extra spatial dimensions, and supersymmetric particles in several compressed scenarios.« less
Aaboud, M.; Aad, G.; Abbott, B.; ...
2018-01-25
Results of a search for new phenomena in final states with an energetic jet and large missing transverse momentum are reported. The search uses proton-proton collision data corresponding to an integrated luminosity of 36.1 fb –1 at a centre-of-mass energy of 13 TeV collected in 2015 and 2016 with the ATLAS detector at the Large Hadron Collider. Events are required to have at least one jet with a transverse momentum above 250 GeV and no leptons (e or μ). Several signal regions are considered with increasing requirements on the missing transverse momentum above 250 GeV. Good agreement is observed betweenmore » the number of events in data and Standard Model predictions. The results are translated into exclusion limits in models with pair-produced weakly interacting dark-matter candidates, large extra spatial dimensions, and supersymmetric particles in several compressed scenarios.« less
Aaboud, M.; Aad, G.; Abbott, B.; ...
2018-01-25
Results of a search for new phenomena in final states with an energetic jet and large missing transverse momentum are reported. The search uses proton-proton collision data corresponding to an integrated luminosity of 36.1 fb -1 at a centre-of-mass energy of 13 TeV collected in 2015 and 2016 with the ATLAS detector at the Large Hadron Collider. Events are required to have at least one jet with a transverse momentum above 250 GeV and no leptons (e or μ). Several signal regions are considered with increasing requirements on the missing transverse momentum above 250 GeV. Good agreement is observed betweenmore » the number of events in data and Standard Model predictions. In conclusion, the results are translated into exclusion limits in models with pair-produced weakly interacting dark-matter candidates, large extra spatial dimensions, and supersymmetric particles in several compressed scenarios.« less
Sensemaking in a Value Based Context for Large Scale Complex Engineered Systems
NASA Astrophysics Data System (ADS)
Sikkandar Basha, Nazareen
The design and the development of Large-Scale Complex Engineered Systems (LSCES) requires the involvement of multiple teams and numerous levels of the organization and interactions with large numbers of people and interdisciplinary departments. Traditionally, requirements-driven Systems Engineering (SE) is used in the design and development of these LSCES. The requirements are used to capture the preferences of the stakeholder for the LSCES. Due to the complexity of the system, multiple levels of interactions are required to elicit the requirements of the system within the organization. Since LSCES involves people and interactions between the teams and interdisciplinary departments, it should be socio-technical in nature. The elicitation of the requirements of most large-scale system projects are subjected to creep in time and cost due to the uncertainty and ambiguity of requirements during the design and development. In an organization structure, the cost and time overrun can occur at any level and iterate back and forth thus increasing the cost and time. To avoid such creep past researches have shown that rigorous approaches such as value based designing can be used to control it. But before the rigorous approaches can be used, the decision maker should have a proper understanding of requirements creep and the state of the system when the creep occurs. Sensemaking is used to understand the state of system when the creep occurs and provide a guidance to decision maker. This research proposes the use of the Cynefin framework, sensemaking framework which can be used in the design and development of LSCES. It can aide in understanding the system and decision making to minimize the value gap due to requirements creep by eliminating ambiguity which occurs during design and development. A sample hierarchical organization is used to demonstrate the state of the system at the occurrence of requirements creep in terms of cost and time using the Cynefin framework. These trials are continued for different requirements and at different sub-system level. The results obtained show that the Cynefin framework can be used to improve the value of the system and can be used for predictive analysis. The decision makers can use these findings and use rigorous approaches and improve the design of Large Scale Complex Engineered Systems.
Kumar, Saravana
2013-01-01
The higher education sector is undergoing tremendous change, driven by complex driving forces including financial, administrative, and organisational and stakeholder expectations. It is in this challenging environment, educators are required to maintain and improve the quality of teaching and learning outcomes while contending with increasing class sizes. Despite mixed evidence on the effectiveness of large classes on student outcomes, large classes continue to play an important part in higher education. While large classes pose numerous challenges, they also provide opportunities for innovative solutions. This paper provides an overview of these challenges and highlights opportunities for innovative solutions.
NASA Astrophysics Data System (ADS)
Henkel, Daniela; Eisenhauer, Anton
2017-04-01
During the last decades, the number of large research projects has increased and therewith the requirement for multidisciplinary, multisectoral collaboration. Such complex and large-scale projects pose new competencies to form, manage, and use large, diverse teams as a competitive advantage. For complex projects the effort is magnified because multiple large international research consortia involving academic and non-academic partners, including big industries, NGOs, private and public bodies, all with cultural differences, individually discrepant expectations on teamwork and differences in the collaboration between national and multi-national administrations and research organisations, challenge the organisation and management of such multi-partner research consortia. How many partners are needed to establish and conduct collaboration with a multidisciplinary and multisectoral approach? How much personnel effort and what kinds of management techniques are required for such projects. This presentation identifies advantages and challenges of large research projects based on the experiences made in the context of an Innovative Training Network (ITN) project within Marie Skłodowska-Curie Actions of the European HORIZON 2020 program. Possible strategies are discussed to circumvent and avoid conflicts already at the beginning of the project.
Combustor technology for future small gas turbine aircraft
NASA Technical Reports Server (NTRS)
Lyons, Valerie J.; Niedzwiecki, Richard W.
1993-01-01
Future engine cycles proposed for advanced small gas turbine engines will increase the severity of the operating conditions of the combustor. These cycles call for increased overall engine pressure ratios which increase combustor inlet pressure and temperature. Further, the temperature rise through the combustor and the corresponding exit temperature also increase. Future combustor technology needs for small gas turbine engines is described. New fuel injectors with large turndown ratios which produce uniform circumferential and radial temperature patterns will be required. Uniform burning will be of greater importance because hot gas temperatures will approach turbine material limits. The higher combustion temperatures and increased radiation at high pressures will put a greater heat load on the combustor liners. At the same time, less cooling air will be available as more of the air will be used for combustion. Thus, improved cooling concepts and/or materials requiring little or no direct cooling will be required. Although presently there are no requirements for emissions levels from small gas turbine engines, regulation is expected in the near future. This will require the development of low emission combustors. In particular, nitrogen oxides will increase substantially if new technologies limiting their formation are not evolved and implemented. For example, staged combustion employing lean, premixed/prevaporized, lean direct injection, or rich burn-quick quench-lean burn concepts could replace conventional single stage combustors.
High Throughput Screening of Toxicity Pathways Perturbed by Environmental Chemicals
Toxicology, a field largely unchanged over the past several decades, is undergoing a significant transformation driven by a number of forces – the increasing number of chemicals needing assessment, changing legal requirements, advances in biology and computer science, and concern...
DOT National Transportation Integrated Search
2010-04-01
Distribution centers (DC) have become more common in Texas over the past decade. As : major generators of large truck traffic, DCs can increase design and maintenance requirements of : Texas highway facilities. This handbook contains guidelines for u...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aaboud, M.; Aad, G.; Abbott, B.
We report results of a search for new phenomena in final states with an energetic jet and large missing transverse momentum. The search uses proton-proton collision data corresponding to an integrated luminosity of 3.2 fb more » $-$1 at $$\\sqrt{s}$$ = 13 TeV collected in 2015 with the ATLAS detector at the Large Hadron Collider. Events are required to have at least one jet with a transverse momentum above 250 GeV and no leptons. Several signal regions are considered with increasing missing-transverse-momentum requirements between E$$miss\\atop{T}$$ > 250 GeV and E$$miss\\atop{T}$$ > 700 GeV . Good agreement is observed between the number of events in data and Standard Model predictions. The results are translated into exclusion limits in models with large extra spatial dimensions, pair production of weakly interacting dark-matter candidates, and the production of supersymmetric particles in several compressed scenarios.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aaboud, M.; Aad, G.; Abbott, B.
Results of a search for new phenomena in final states with an energetic jet and large missing transverse momentum are reported. The search uses proton-proton collision data corresponding to an integrated luminosity of 3.2 fb –1 at √s=13 TeV collected in 2015 with the ATLAS detector at the Large Hadron Collider. Events are required to have at least one jet with a transverse momentum above 250 GeV and no leptons. Several signal regions are considered with increasing missing-transverse-momentum requirements between E miss T > 250 GeV and E miss T > 700 GeV. Good agreement is observed between the numbermore » of events in data and Standard Model predictions. Here, the results are translated into exclusion limits in models with large extra spatial dimensions, pair production of weakly interacting dark-matter candidates, and the production of supersymmetric particles in several compressed scenarios.« less
Brown, Adrian P; Borgs, Christian; Randall, Sean M; Schnell, Rainer
2017-06-08
Integrating medical data using databases from different sources by record linkage is a powerful technique increasingly used in medical research. Under many jurisdictions, unique personal identifiers needed for linking the records are unavailable. Since sensitive attributes, such as names, have to be used instead, privacy regulations usually demand encrypting these identifiers. The corresponding set of techniques for privacy-preserving record linkage (PPRL) has received widespread attention. One recent method is based on Bloom filters. Due to superior resilience against cryptographic attacks, composite Bloom filters (cryptographic long-term keys, CLKs) are considered best practice for privacy in PPRL. Real-world performance of these techniques using large-scale data is unknown up to now. Using a large subset of Australian hospital admission data, we tested the performance of an innovative PPRL technique (CLKs using multibit trees) against a gold-standard derived from clear-text probabilistic record linkage. Linkage time and linkage quality (recall, precision and F-measure) were evaluated. Clear text probabilistic linkage resulted in marginally higher precision and recall than CLKs. PPRL required more computing time but 5 million records could still be de-duplicated within one day. However, the PPRL approach required fine tuning of parameters. We argue that increased privacy of PPRL comes with the price of small losses in precision and recall and a large increase in computational burden and setup time. These costs seem to be acceptable in most applied settings, but they have to be considered in the decision to apply PPRL. Further research on the optimal automatic choice of parameters is needed.
NASA Technical Reports Server (NTRS)
Freeman, Hugh B.
1935-01-01
Tests were made in the N.A.C.A. 20-foot wind tunnel on: (1) a wing, of 6.5-foot span, 5.5-foot chord, and 30 percent maximum thickness, fitted with large end plates and (2) a 16-foot span 2.67-foot chord wing of 15 percent maximum thickness to determine the increase in lift obtainable by removing the boundary layer and the power required for the blower. The results of the tests on the stub wing appeared more favorable than previous small-scale tests and indicated that: (1) the suction method was considerably superior to the pressure method, (2) single slots were more effective than multiple slots (where the same pressure was applied to all slots), the slot efficiency increased rapidly for increasing slot widths up to 2 percent of the wing chord and remained practically constant for all larger widths tested, (3) suction pressure and power requirements were quite low (a computation for a light airplane showed that a lift coefficient of 3.0 could be obtained with a suction as low as 2.3 times the dynamic pressure and a power expenditure less than 3 percent of the rated engine power), and (4) the volume of air required to be drawn off was quite high (approximately 0.5 cubic feet per second per unit wing area for an airplane landing at 40 miles per hour with a lift coefficient of 3,0), indicating that considerable duct area must be provided in order to prevent flow losses inside the wing and insure uniform distribution of suction along the span. The results from the tests of the large-span wing were less favorable than those on the stub wing. The reasons for this were, probably: (1) the uneven distribution of suction along the span, (2) the flow losses inside the wing, (3) the small radius of curvature of the leading edge of the wing section, and (4) the low Reynolds Number of these tests, which was about one half that of the stub wing. The results showed a large increase in the maximum lift coefficient with an increase in Reynolds Number in the range of the tests. The results of drag tests showed that the profile drag of the wing was reduced and the L/D ratio was increased throughout the range of lift coefficients corresponding to take-off and climb but that the minimum drag was increased. The slot arrangement that is best for low drag is not the same, however, as that for maximum lift.
Pushing the limits of radiofrequency (RF) neuronal telemetry
Yousefi, Tara; Diaz, Rodolfo E.
2015-01-01
In a previous report it was shown that the channel capacity of an in vivo communication link using microscopic antennas at radiofrequency is severely limited by the requirement not to damage the tissue surrounding the antennas. For dipole-like antennas the strong electric field dissipates too much power into body tissues. Loop-type antennas have a strong magnetic near field and so dissipate much less power into the surrounding tissues but they require such a large current that the antenna temperature is raised to the thermal damage threshold of the tissue. The only solution was increasing the antenna size into hundreds of microns, which makes reporting on an individual neuron impossible. However, recently demonstrated true magnetic antennas offer an alternative not covered in the previous report. The near field of these antennas is dominated by the magnetic field yet they don’t require large currents. Thus they combine the best characteristics of dipoles and loops. By calculating the coupling between identical magnetic antennas inside a model of the body medium we show an increase in the power transfer of up to 8 orders of magnitude higher than could be realized with the loops and dipoles, making the microscopic RF in-vivo transmitting antenna possible. PMID:26035824
Copyright protection of remote sensing imagery by means of digital watermarking
NASA Astrophysics Data System (ADS)
Barni, Mauro; Bartolini, Franco; Cappellini, Vito; Magli, Enrico; Olmo, Gabriella; Zanini, R.
2001-12-01
The demand for remote sensing data has increased dramatically mainly due to the large number of possible applications capable to exploit remotely sensed data and images. As in many other fields, along with the increase of market potential and product diffusion, the need arises for some sort of protection of the image products from unauthorized use. Such a need is a very crucial one even because the Internet and other public/private networks have become preferred and effective means of data exchange. An important issue arising when dealing with digital image distribution is copyright protection. Such a problem has been largely addressed by resorting to watermarking technology. Before applying watermarking techniques developed for multimedia applications to remote sensing applications, it is important that the requirements imposed by remote sensing imagery are carefully analyzed to investigate whether they are compatible with existing watermarking techniques. On the basis of these motivations, the contribution of this work is twofold: (1) assessment of the requirements imposed by the characteristics of remotely sensed images on watermark-based copyright protection; (2) discussion of a case study where the performance of two popular, state-of-the-art watermarking techniques are evaluated by the light of the requirements at the previous point.
ERIC Educational Resources Information Center
Duggan, Mark; Hayford, Tamara
2013-01-01
From 1991 to 2009, the fraction of Medicaid recipients enrolled in HMOs and other forms of Medicaid managed care (MMC) increased from 11 percent to 71 percent. This increase was largely driven by state and local mandates that required most Medicaid recipients to enroll in an MMC plan. Theoretically, it is ambiguous whether the shift from…
Coordinated Management of Academic Health Centers.
Balser, Jeffrey R; Stead, William W
2017-01-01
Academic health centers (AHCs) are the nation's primary resource for healthcare discovery, innovation, and training. US healthcare revenue growth has declined sharply since 2009, and is forecast to remain well below historic levels for the foreseeable future. As the cost of education and research at nearly all AHCs is heavily subsidized through large transfers from clinical care margins, our institutions face a mounting crisis. Choices centering on how to increase the cost-effectiveness of the AHC enterprise require unprecedented levels of alignment to preserve an environment that nurtures creativity. Management processes require governance models that clarify decision rights while harnessing the talents and the intellectual capital of a large, diverse enterprise to nimbly address unfamiliar organizational challenges. This paper describes key leadership tactics aimed at propelling AHCs along this journey - one that requires from all leaders a commitment to resilience, optimism, and willingness to embrace change.
NASA Astrophysics Data System (ADS)
Bradshaw, A. M.; Reuter, B.; Hamacher, T.
2015-08-01
The energy transformation process beginning to take place in many countries as a response to climate change will reduce substantially the consumption of fossil fuels, but at the same time cause a large increase in the demand for other raw materials. Whereas it is difficult to estimate the quantities of, for example, iron, copper and aluminium required, the situation is somewhat simpler for the rare elements that might be needed in a sustainable energy economy based largely on photovoltaic sources, wind and possibly nuclear fusion. We consider briefly each of these technologies and discuss the supply risks associated with the rare elements required, if they were to be used in the quantities that might be required for a global energy transformation process. In passing, we point out the need in resource studies to define the terms "rare", "scarce" and "critical" and to use them in a consistent way.
Alton, Lesley A; Portugal, Steven J; White, Craig R
2013-02-01
Air-breathing fish of the Anabantoidei group meet their metabolic requirements for oxygen through both aerial and aquatic gas exchange. Siamese fighting fish Betta splendens are anabantoids that frequently engage in aggressive male-male interactions which cause significant increases in metabolic rate and oxygen requirements. These interactions involve opercular flaring behaviour that is thought to limit aquatic oxygen uptake, and combines with the increase in metabolic rate to cause an increase in air-breathing behaviour. Air-breathing events interrupt display behaviour and increase risk of predation, raising the question of how Siamese fighting fish manage their oxygen requirements during agonistic encounters. Using open-flow respirometry, we measured rate of oxygen consumption in displaying fish to determine if males increase oxygen uptake per breath to minimise visits to the surface, or increase their reliance on aquatic oxygen uptake. We found that the increased oxygen requirements of Siamese fighting fish during display behaviour were met by increased oxygen uptake from the air with no significant changes in aquatic oxygen uptake. The increased aerial oxygen uptake was achieved almost entirely by an increase in air-breathing frequency. We conclude that limitations imposed by the reduced gill surface area of air-breathing fish restrict the ability of Siamese fighting fish to increase aquatic uptake, and limitations of the air-breathing organ of anabantoids largely restrict their capacity to increase oxygen uptake per breath. The resulting need to increase surfacing frequency during metabolically demanding agonistic encounters has presumably contributed to the evolution of the stereotyped surfacing behaviour seen during male-male interactions, during which one of the fish will lead the other to the surface, and each will take a breath of air. Copyright © 2012. Published by Elsevier Inc.
Moore, Craig S; Horsfield, Carl J; Saunderson, John R; Beavis, Andrew W
2015-01-01
Objective: The purpose of this study was to develop size-based radiotherapy kilovoltage cone beam CT (CBCT) protocols for the pelvis. Methods: Image noise was measured in an elliptical phantom of varying size for a range of exposure factors. Based on a previously defined “small pelvis” reference patient and CBCT protocol, appropriate exposure factors for small, medium, large and extra-large patients were derived which approximate the image noise behaviour observed on a Philips CT scanner (Philips Medical Systems, Best, Netherlands) with automatic exposure control (AEC). Selection criteria, based on maximum tube current–time product per rotation selected during the radiotherapy treatment planning scan, were derived based on an audit of patient size. Results: It has been demonstrated that 110 kVp yields acceptable image noise for reduced patient dose in pelvic CBCT scans of small, medium and large patients, when compared with manufacturer's default settings (125 kVp). Conversely, extra-large patients require increased exposure factors to give acceptable images. 57% of patients in the local population now receive much lower radiation doses, whereas 13% require higher doses (but now yield acceptable images). Conclusion: The implementation of size-based exposure protocols has significantly reduced radiation dose to the majority of patients with no negative impact on image quality. Increased doses are required on the largest patients to give adequate image quality. Advances in knowledge: The development of size-based CBCT protocols that use the planning CT scan (with AEC) to determine which protocol is appropriate ensures adequate image quality whilst minimizing patient radiation dose. PMID:26419892
Large-angle cosmic microwave background anisotropies in an open universe
NASA Technical Reports Server (NTRS)
Kamionkowski, Marc; Spergel, David N.
1994-01-01
If the universe is open, scales larger than the curvature scale may be probed by observation of large-angle fluctuations in the cosmic microwave background (CMB). We consider primordial adiabatic perturbations and discuss power spectra that are power laws in volume, wavelength, and eigenvalue of the Laplace operator. Such spectra may have arisen if, for example, the universe underwent a period of `frustated' inflation. The resulting large-angle anisotropies of the CMB are computed. The amplitude generally increases as Omega is decreased but decreases as h is increased. Interestingly enough, for all three Ansaetze, anisotropies on angular scales larger than the curvature scale are suppressed relative to the anisotropies on scales smaller than the curvature scale, but cosmic variance makes discrimination between various models difficult. Models with 0.2 approximately less than Omega h approximately less than 0.3 appear compatible with CMB fluctuations detected by Cosmic Background Explorer Satellite (COBE) and the Tenerife experiment and with the amplitude and spectrum of fluctuations of galaxy counts in the APM, CfA, and 1.2 Jy IRAS surveys. COBE normalization for these models yields sigma(sub 8) approximately = 0.5 - 0.7. Models with smaller values of Omega h when normalized to COBE require bias factors in excess of 2 to be compatible with the observed galaxy counts on the 8/h Mpc scale. Requiring that the age of the universe exceed 10 Gyr implies that Omega approximately greater than 0.25, while requiring that from the last-scattering term in the Sachs-Wolfe formula, large-angle anisotropies come primarily from the decay of potential fluctuations at z approximately less than 1/Omega. Thus, if the universe is open, COBE has been detecting temperature fluctuations produced at moderate redshift rather than at z approximately 1300.
Advanced aerosense display interfaces
NASA Astrophysics Data System (ADS)
Hopper, Darrel G.; Meyer, Frederick M.
1998-09-01
High-resolution display technologies are being developed to meet the ever-increasing demand for realistic detail. The requirement for evermore visual information exceeds the capacity of fielded aerospace display interfaces. In this paper we begin an exploration of display interfaces and evolving aerospace requirements. Current and evolving standards for avionics, commercial, and flat panel displays are summarized and compared to near term goals for military and aerospace applications. Aerospace and military applications prior to 2005 up to UXGA and digital HDTV resolution can be met by using commercial interface standard developments. Advanced aerospace requirements require yet higher resolutions (2560 X 2048 color pixels, 5120 X 4096 color pixels at 85 Hz, etc.) and necessitate the initiation of discussion herein of an 'ultra digital interface standard (UDIS)' which includes 'smart interface' features such as large memory and blazingly fast resizing microcomputer. Interface capacity, IT, increased about 105 from 1973 to 1998; 102 more is needed for UDIS.
Engineering Education for Agricultural and Rural Development in Africa
ERIC Educational Resources Information Center
Adewumi, B. A.
2008-01-01
Agricultural Engineering has transformed agricultural practices from subsistence level to medium and large-scale production via mechanisation in the developed nations. This has reduced the labour force requirements in agriculture; increased production levels and efficiency, product shelf life and product quality; and resulted into…
USE OF ALTERED MICROORGANISMS FOR FIELD BIODEGRADATION OF HAZARDOUS MATERIALS
The large amount of hazardous waste generated and disposed of has given rise to environmental conditions requiring remedial treatment. he use of landfills has traditionally been a cost-effective means to dispose of waste. owever, increased costs of transportation and decreasing n...
Large increase in fracture resistance of stishovite with crack extension less than one micrometer
Yoshida, Kimiko; Wakai, Fumihiro; Nishiyama, Norimasa; Sekine, Risako; Shinoda, Yutaka; Akatsu, Takashi; Nagoshi, Takashi; Sone, Masato
2015-01-01
The development of strong, tough, and damage-tolerant ceramics requires nano/microstructure design to utilize toughening mechanisms operating at different length scales. The toughening mechanisms so far known are effective in micro-scale, then, they require the crack extension of more than a few micrometers to increase the fracture resistance. Here, we developed a micro-mechanical test method using micro-cantilever beam specimens to determine the very early part of resistance-curve of nanocrystalline SiO2 stishovite, which exhibited fracture-induced amorphization. We revealed that this novel toughening mechanism was effective even at length scale of nanometer due to narrow transformation zone width of a few tens of nanometers and large dilatational strain (from 60 to 95%) associated with the transition of crystal to amorphous state. This testing method will be a powerful tool to search for toughening mechanisms that may operate at nanoscale for attaining both reliability and strength of structural materials. PMID:26051871
Mathematics and mallard management
Cowardin, L.M.; Johnson, D.H.
1979-01-01
Waterfowl managers can effectively use simple population models to aid in making management decisions. We present a basic model of the change in population size as related to survival and recruitment. A management technique designed to increase survival of mallards (Anas platyrhynchos) by limiting harvest on the Chippewa National Forest, Minnesota, is used to illustrate the application of models in decision making. The analysis suggests that the management technique would be of limited effectiveness. In a 2nd example, the change in mallard population in central North Dakota is related to implementing programs to create dense nesting cover with or without supplementary predator control. The analysis suggests that large tracts of land would be required to achieve a hypothetical management objective of increasing harvest by 50% while maintaining a stable population. Less land would be required if predator reduction were used in combination with cover management, but questions about effectiveness and ecological implications of large scale predator reduction remain unresolved. The use of models as a guide to planning research responsive to the needs of management is illustrated.
Pregestational diabetes with extreme insulin resistance: use of U-500 insulin in pregnancy.
Zuckerwise, Lisa C; Werner, Erika F; Pettker, Christian M; McMahon-Brown, Erin K; Thung, Stephen F; Han, Christina S
2012-08-01
Increased insulin requirements in pregnancy can hinder attainment of glycemic control in diabetic patients. U-500 insulin is a concentrated form of regular insulin that can be a valuable tool in the treatment of patients with severe insulin resistance. A 24-year-old woman with pregestational diabetes mellitus experienced increasing insulin requirements during pregnancy, peaking at 650 units daily. The frequent, large-volume injections of standard-concentration insulin were poorly tolerated by the patient and resulted in nonadherence. She subsequently achieved glycemic control on thrice-daily U-500 insulin. Pregnancy exacerbates insulin resistance in diabetic patients, and these patients may require high doses of insulin. U-500 insulin is an effective alternative for patients with severe insulin resistance and should be considered for pregnant women with difficulty achieving glycemic control.
Maresh, J L; Adachi, T; Takahashi, A; Naito, Y; Crocker, D E; Horning, M; Williams, T M; Costa, D P
2015-01-01
The energy requirements of free-ranging marine mammals are challenging to measure due to cryptic and far-ranging feeding habits, but are important to quantify given the potential impacts of high-level predators on ecosystems. Given their large body size and carnivorous lifestyle, we would predict that northern elephant seals (Mirounga angustirostris) have elevated field metabolic rates (FMRs) that require high prey intake rates, especially during pregnancy. Disturbance associated with climate change or human activity is predicted to further elevate energy requirements due to an increase in locomotor costs required to accommodate a reduction in prey or time available to forage. In this study, we determined the FMRs, total energy requirements, and energy budgets of adult, female northern elephant seals. We also examined the impact of increased locomotor costs on foraging success in this species. Body size, time spent at sea and reproductive status strongly influenced FMR. During the short foraging migration, FMR averaged 90.1 (SE = 1.7) kJ kg(-1)d(-1) - only 36 % greater than predicted basal metabolic rate. During the long migration, when seals were pregnant, FMRs averaged 69.4 (±3.0) kJ kg(-1)d(-1) - values approaching those predicted to be necessary to support basal metabolism in mammals of this size. Low FMRs in pregnant seals were driven by hypometabolism coupled with a positive feedback loop between improving body condition and reduced flipper stroking frequency. In contrast, three additional seals carrying large, non-streamlined instrumentation saw a four-fold increase in energy partitioned toward locomotion, resulting in elevated FMRs and only half the mass gain of normally-swimming study animals. These results highlight the importance of keeping locomotion costs low for successful foraging in this species. In preparation for lactation and two fasting periods with high demands on energy reserves, migrating elephant seals utilize an economical foraging strategy whereby energy savings from reduced locomotion costs are shuttled towards somatic growth and fetal gestation. Remarkably, the energy requirements of this species, particularly during pregnancy, are 70-80 % lower than expected for mammalian carnivores, approaching or even falling below values predicted to be necessary to support basal metabolism in mammals of this size.
High School Physical Education Requirements and Youth Body Weight: New Evidence from the YRBS.
Sabia, Joseph J; Nguyen, Thanh Tam; Rosenberg, Oren
2017-10-01
Previous research has found that high school physical education (PE) requirements are largely ineffective at reducing youth body weight. However, these studies were forced to rely on cross-state variation in PE requirements to identify their impacts, raising concerns that estimated policy effects may be confounded by state-level unobservables. Using data from the State and National Youth Risk Behavior Surveys and exploiting recent changes in state high school PE laws, we re-examine the effect of PE requirements on body weight. Our estimates show that a one-semester increase in PE requirements is associated with a 10 to 13% increase in minutes per week spent physically active in PE classes, but with no change in net vigorous exercise and little change in youth body weight. We conclude that substitution of in-school for outside-of-school physical activity and small resultant net energy expenditures can explain the absence of body weight effects. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.
Accuracy assessment with complex sampling designs
Raymond L. Czaplewski
2010-01-01
A reliable accuracy assessment of remotely sensed geospatial data requires a sufficiently large probability sample of expensive reference data. Complex sampling designs reduce cost or increase precision, especially with regional, continental and global projects. The General Restriction (GR) Estimator and the Recursive Restriction (RR) Estimator separate a complex...
Intrusion-Tolerant Replication under Attack
ERIC Educational Resources Information Center
Kirsch, Jonathan
2010-01-01
Much of our critical infrastructure is controlled by large software systems whose participants are distributed across the Internet. As our dependence on these critical systems continues to grow, it becomes increasingly important that they meet strict availability and performance requirements, even in the face of malicious attacks, including those…
Characterizing body temperature and activity changes at the onset of estrus in replacement gilts
USDA-ARS?s Scientific Manuscript database
Accurate estrus detection can improve sow conception rates and increase swine production efficiency. Unfortunately, current estrus detection practices based on individual animal behavior may be inefficient due to large sow populations at commercial farms and the associated labor required. Therefore,...
The molecular basis for tree growth habit in Prunus persica (peach)
USDA-ARS?s Scientific Manuscript database
The large size and spreading growth habit of trees requires excessive labor, land space, and pesticides. Genetically improving tree shapes so they can be planted at higher density and/or more readily adapted to mechanization would increase productivity and be more environmentally friendly. Current...
SPACE FOR AUDIO-VISUAL LARGE GROUP INSTRUCTION.
ERIC Educational Resources Information Center
GAUSEWITZ, CARL H.
WITH AN INCREASING INTEREST IN AND UTILIZATION OF AUDIO-VISUAL MEDIA IN EDUCATION FACILITIES, IT IS IMPORTANT THAT STANDARDS ARE ESTABLISHED FOR ESTIMATING THE SPACE REQUIRED FOR VIEWING THESE VARIOUS MEDIA. THIS MONOGRAPH SUGGESTS SUCH STANDARDS FOR VIEWING AREAS, VIEWING ANGLES, SEATING PATTERNS, SCREEN CHARACTERISTICS AND EQUIPMENT PERFORMANCES…
Strategic Planning Tools for Large-Scale Technology-Based Assessments
ERIC Educational Resources Information Center
Koomen, Marten; Zoanetti, Nathan
2018-01-01
Education systems are increasingly being called upon to implement new technology-based assessment systems that generate efficiencies, better meet changing stakeholder expectations, or fulfil new assessment purposes. These assessment systems require coordinated organisational effort to implement and can be expensive in time, skill and other…
Quinley, J C; Baker, T D
1986-07-01
Historically, the Agency for International Development (AIDS) health budget has been closely tied to overall development spending. A large increase in the international health appropriations in 1984 broke this pattern. Investigation shows that active grass roots organizing and congressional lobbying are the most likely responsible factors in the increase. Maintenance and expansion of this success will require increased recognition of and participation in these activities by individuals and organizations involved in international health.
Acceleration techniques for dependability simulation. M.S. Thesis
NASA Technical Reports Server (NTRS)
Barnette, James David
1995-01-01
As computer systems increase in complexity, the need to project system performance from the earliest design and development stages increases. We have to employ simulation for detailed dependability studies of large systems. However, as the complexity of the simulation model increases, the time required to obtain statistically significant results also increases. This paper discusses an approach that is application independent and can be readily applied to any process-based simulation model. Topics include background on classical discrete event simulation and techniques for random variate generation and statistics gathering to support simulation.
Psychiatric Emergencies in the Elderly.
Sikka, Veronica; Kalra, S; Galwankar, Sagar; Sagar, Galwankar
2015-11-01
With the increasing life expectancy, the geriatric population has been increasing over the past few decades. By the year 2050, it is projected to compose more than a fifth of the entire population, representing a 147% increase in this age group. There has been a steady increase in the number of medical and psychiatric disorders, and a large percentage of geriatric patients are now presenting to the emergency department with such disorders. The management of our progressively complex geriatric patient population will require an integrative team approach involving emergency medicine, psychiatry, and hospitalist medicine. Published by Elsevier Inc.
Møller, Anders Pape; Nielsen, Jan Tøttrup
2015-11-01
Many animals build extravagant nests that exceed the size required for successful reproduction. Large nests may signal the parenting ability of nest builders suggesting that nests may have a signaling function. In particular, many raptors build very large nests for their body size. We studied nest size in the goshawk Accipiter gentilis, which is a top predator throughout most of the Nearctic. Both males and females build nests, and males provision their females and offspring with food. Nest volume in the goshawk is almost three-fold larger than predicted from their body size. Nest size in the goshawk is highly variable and may reach more than 600 kg for a bird that weighs ca. 1 kg. While 8.5% of nests fell down, smaller nests fell down more often than large nests. There was a hump-shaped relationship between nest volume and female age, with a decline in nest volume late in life, as expected for senescence. Clutch size increased with nest volume. Nest volume increased during 1977-2014 in an accelerating fashion, linked to increasing spring temperature during April, when goshawks build and start reproduction. These findings are consistent with nest size being a reliable signal of parental ability, with large nest size signaling superior parenting ability and senescence, and also indicating climate warming.
Scale-Up: Improving Large Enrollment Physics Courses
NASA Astrophysics Data System (ADS)
Beichner, Robert
1999-11-01
The Student-Centered Activities for Large Enrollment University Physics (SCALE-UP) project is working to establish a learning environment that will promote increased conceptual understanding, improved problem-solving performance, and greater student satisfaction, while still maintaining class sizes of approximately 100. We are also addressing the new ABET engineering accreditation requirements for inquiry-based learning along with communication and team-oriented skills development. Results of studies of our latest classroom design, plans for future classroom space, and the current iteration of instructional materials will be discussed.
Effects of large herbivores on grassland arthropod diversity.
van Klink, R; van der Plas, F; van Noordwijk, C G E Toos; WallisDeVries, M F; Olff, H
2015-05-01
Both arthropods and large grazing herbivores are important components and drivers of biodiversity in grassland ecosystems, but a synthesis of how arthropod diversity is affected by large herbivores has been largely missing. To fill this gap, we conducted a literature search, which yielded 141 studies on this topic of which 24 simultaneously investigated plant and arthropod diversity. Using the data from these 24 studies, we compared the responses of plant and arthropod diversity to an increase in grazing intensity. This quantitative assessment showed no overall significant effect of increasing grazing intensity on plant diversity, while arthropod diversity was generally negatively affected. To understand these negative effects, we explored the mechanisms by which large herbivores affect arthropod communities: direct effects, changes in vegetation structure, changes in plant community composition, changes in soil conditions, and cascading effects within the arthropod interaction web. We identify three main factors determining the effects of large herbivores on arthropod diversity: (i) unintentional predation and increased disturbance, (ii) decreases in total resource abundance for arthropods (biomass) and (iii) changes in plant diversity, vegetation structure and abiotic conditions. In general, heterogeneity in vegetation structure and abiotic conditions increases at intermediate grazing intensity, but declines at both low and high grazing intensity. We conclude that large herbivores can only increase arthropod diversity if they cause an increase in (a)biotic heterogeneity, and then only if this increase is large enough to compensate for the loss of total resource abundance and the increased mortality rate. This is expected to occur only at low herbivore densities or with spatio-temporal variation in herbivore densities. As we demonstrate that arthropod diversity is often more negatively affected by grazing than plant diversity, we strongly recommend considering the specific requirements of arthropods when applying grazing management and to include arthropods in monitoring schemes. Conservation strategies aiming at maximizing heterogeneity, including regulation of herbivore densities (through human interventions or top-down control), maintenance of different types of management in close proximity and rotational grazing regimes, are the most promising options to conserve arthropod diversity. © 2014 The Authors. Biological Reviews published by John Wiley & Sons Ltd on behalf of Cambridge Philosophical Society.
Effects of large herbivores on grassland arthropod diversity
van Klink, R; van der Plas, F; van Noordwijk, C G E (Toos); WallisDeVries, M F; Olff, H
2015-01-01
Both arthropods and large grazing herbivores are important components and drivers of biodiversity in grassland ecosystems, but a synthesis of how arthropod diversity is affected by large herbivores has been largely missing. To fill this gap, we conducted a literature search, which yielded 141 studies on this topic of which 24 simultaneously investigated plant and arthropod diversity. Using the data from these 24 studies, we compared the responses of plant and arthropod diversity to an increase in grazing intensity. This quantitative assessment showed no overall significant effect of increasing grazing intensity on plant diversity, while arthropod diversity was generally negatively affected. To understand these negative effects, we explored the mechanisms by which large herbivores affect arthropod communities: direct effects, changes in vegetation structure, changes in plant community composition, changes in soil conditions, and cascading effects within the arthropod interaction web. We identify three main factors determining the effects of large herbivores on arthropod diversity: (i) unintentional predation and increased disturbance, (ii) decreases in total resource abundance for arthropods (biomass) and (iii) changes in plant diversity, vegetation structure and abiotic conditions. In general, heterogeneity in vegetation structure and abiotic conditions increases at intermediate grazing intensity, but declines at both low and high grazing intensity. We conclude that large herbivores can only increase arthropod diversity if they cause an increase in (a)biotic heterogeneity, and then only if this increase is large enough to compensate for the loss of total resource abundance and the increased mortality rate. This is expected to occur only at low herbivore densities or with spatio-temporal variation in herbivore densities. As we demonstrate that arthropod diversity is often more negatively affected by grazing than plant diversity, we strongly recommend considering the specific requirements of arthropods when applying grazing management and to include arthropods in monitoring schemes. Conservation strategies aiming at maximizing heterogeneity, including regulation of herbivore densities (through human interventions or top-down control), maintenance of different types of management in close proximity and rotational grazing regimes, are the most promising options to conserve arthropod diversity. PMID:24837856
Aaboud, M.; Aad, G.; Abbott, B.; ...
2016-08-22
Results of a search for new phenomena in final states with an energetic jet and large missing transverse momentum are reported. The search uses proton-proton collision data corresponding to an integrated luminosity of 3.2 fb –1 at √s=13 TeV collected in 2015 with the ATLAS detector at the Large Hadron Collider. Events are required to have at least one jet with a transverse momentum above 250 GeV and no leptons. Several signal regions are considered with increasing missing-transverse-momentum requirements between E miss T > 250 GeV and E miss T > 700 GeV. Good agreement is observed between the numbermore » of events in data and Standard Model predictions. Here, the results are translated into exclusion limits in models with large extra spatial dimensions, pair production of weakly interacting dark-matter candidates, and the production of supersymmetric particles in several compressed scenarios.« less
Flying in a flock comes at a cost in pigeons.
Usherwood, James R; Stavrou, Marinos; Lowe, John C; Roskilly, Kyle; Wilson, Alan M
2011-06-22
Flying birds often form flocks, with social, navigational and anti-predator implications. Further, flying in a flock can result in aerodynamic benefits, thus reducing power requirements, as demonstrated by a reduction in heart rate and wingbeat frequency in pelicans flying in a V-formation. But how general is an aerodynamic power reduction due to group-flight? V-formation flocks are limited to moderately steady flight in relatively large birds, and may represent a special case. What are the aerodynamic consequences of flying in the more usual 'cluster' flock? Here we use data from innovative back-mounted Global Positioning System (GPS) and 6-degrees-of-freedom inertial sensors to show that pigeons (1) maintain powered, banked turns like aircraft, imposing dorsal accelerations of up to 2g, effectively doubling body weight and quadrupling induced power requirements; (2) increase flap frequency with increases in all conventional aerodynamic power requirements; and (3) increase flap frequency when flying near, particularly behind, other birds. Therefore, unlike V-formation pelicans, pigeons do not gain an aerodynamic advantage from flying in a flock. Indeed, the increased flap frequency, whether due to direct aerodynamic interactions or requirements for increased stability or control, suggests a considerable energetic cost to flight in a tight cluster flock.
Flying in a flock comes at a cost in pigeons
Usherwood, James R.; Stavrou, Marinos; Lowe, John C.; Roskilly, Kyle; Wilson, Alan M.
2011-01-01
Flying birds often form flocks, with social1, navigational2 and anti-predator3 implications. Further, flying in a flock can result in aerodynamic benefits, thus reducing power requirements4, as demonstrated by a reduction in heart rate and wingbeat frequency in pelicans flying in a V-formation5. But how general is an aerodynamic power reduction due to group-flight? V-formation flocks are limited to moderately steady flight in relatively large birds, and may represent a special case. What are the aerodynamic consequences of flying in the more usual ‘cluster’ 6,7 flock? Here, we use data from innovative back-mounted GPS and 6 degree of freedom inertial sensors to show that pigeons 1) maintain powered, banked turns like aircraft, imposing dorsal accelerations of up to 2g, effectively doubling body weight and quadrupling induced power requirements; 2) increase flap frequency with increases in all conventional aerodynamic power requirements; and 3) increase flap frequency when flying near, particularly behind, other birds. Therefore, unlike V-formation pelicans, pigeons do not gain an aerodynamic advantage from flying in a flock; indeed, the increased flap frequency – whether due to direct aerodynamic interactions or requirements for increased stability or control – suggests a considerable energetic cost to flight in a tight cluster flock. PMID:21697946
Stereoselective aminoacylation of RNA
NASA Technical Reports Server (NTRS)
Usher, D. A.; Needels, M. C.; Brenner, T.
1986-01-01
Prebiotic chemistry is faced with a major problem: how could a controlled and selective reaction occur, when there is present in the same solution a large number of alternative possible coreactants? This problem is solved in the modern cell by the presence of enzymes, which are not only highly efficient and controllable catalysts, but which also can impose on their substrates a precise structural requirement. However, enzymes are the result of billions of years of evolution, and we cannot invoke them as prebiotic catalysts. One approach to solving this problem in the prebiotic context is to make use of template-directed reactions. These reactions increase the number of structural requirements that must be simultaneously present in a molecule for it to be able to react, and thereby increase the selectivity of the reaction. They also can give a large increase in the rate of a reaction, if the template constrains two potential coreactants to lie close together. A third benefit is that information that is present in the template molecule can be passed on to the product molecules. If the earliest organisms were based on proteins and nucleic acids, then the investigation of peptide synthesis on an oligonucleotide template is highly relevant to the study of the origin of life.
Robles, Brenda; Wood, Michelle; Kimmons, Joel; Kuo, Tony
2013-03-01
National, state, and local institutions that procure, distribute, sell, and/or serve food to employees, students, and the public are increasingly capitalizing on existing operational infrastructures to create healthier food environments. Integration of healthy nutrition standards and other recommended practices [e.g., energy (kilocalories) postings at point-of-purchase, portion size restrictions, product placement guidelines, and signage] into new or renewing food service and vending contracts codifies an institution's commitment to increasing the availability of healthful food options in their food service venues and vending machines. These procurement requirements, in turn, have the potential to positively influence consumers' food-purchasing behaviors. Although these strategies are becoming increasingly popular, much remains unknown about their context, the processes required to implement them effectively, and the factors that facilitate their sustainability, especially in such broad and diverse settings as schools, county government facilities, and cities. To contribute to this gap in information, we reviewed and compared nutrition standards and other best practices implemented recently in a large school district, in a large county government, and across 10 municipalities in Los Angeles County. We report lessons learned from these efforts.
Promoting active learning using audience response system in large bioscience classes.
Efstathiou, Nikolaos; Bailey, Cara
2012-01-01
This paper considers the challenges of bioscience teaching and learning in pre-registration nurse education. Effective learning requires active student participation which is problematic when teaching large groups of students. New technologies, such as the audience response system (ARS), have been introduced to increase student participation and support them in the understanding of complex bioscience concepts. Within one university department, an evaluation was undertaken to identify the perceptions of pre-registration nurse students on the use of ARS in the teaching and learning of bioscience. Our findings concur with others that ARS increases student participation and aids in identifying misconceptions and in correcting them. Students found ARS very useful and wanted ARS to be used in additional modules too. Although ARS did not seem to motivate students to study adequately before attending the relevant sessions, it increased discussion among students and awareness of their level of knowledge compared to their peers. Further research is required to identify the effectiveness of ARS in the teaching and learning of bioscience and its impact on the performance of the students in their final assessments. Copyright © 2011 Elsevier Ltd. All rights reserved.
Robles, Brenda; Wood, Michelle; Kimmons, Joel; Kuo, Tony
2013-01-01
National, state, and local institutions that procure, distribute, sell, and/or serve food to employees, students, and the public are increasingly capitalizing on existing operational infrastructures to create healthier food environments. Integration of healthy nutrition standards and other recommended practices [e.g., energy (kilocalories) postings at point-of-purchase, portion size restrictions, product placement guidelines, and signage] into new or renewing food service and vending contracts codifies an institution’s commitment to increasing the availability of healthful food options in their food service venues and vending machines. These procurement requirements, in turn, have the potential to positively influence consumers’ food-purchasing behaviors. Although these strategies are becoming increasingly popular, much remains unknown about their context, the processes required to implement them effectively, and the factors that facilitate their sustainability, especially in such broad and diverse settings as schools, county government facilities, and cities. To contribute to this gap in information, we reviewed and compared nutrition standards and other best practices implemented recently in a large school district, in a large county government, and across 10 municipalities in Los Angeles County. We report lessons learned from these efforts. PMID:23493535
Islam, M. R.; Garcia, S. C.; Clark, C. E. F.; Kerrisk, K. L.
2015-01-01
To maintain a predominantly pasture-based system, the large herd milked by automatic milking rotary would be required to walk significant distances. Walking distances of greater than 1-km are associated with an increased incidence of undesirably long milking intervals and reduced milk yield. Complementary forages can be incorporated into pasture-based systems to lift total home grown feed in a given area, thus potentially ‘concentrating’ feed closer to the dairy. The aim of this modelling study was to investigate the total land area required and associated walking distance for large automatic milking system (AMS) herds when incorporating complementary forage rotations (CFR) into the system. Thirty-six scenarios consisting of 3 AMS herds (400, 600, 800 cows), 2 levels of pasture utilisation (current AMS utilisation of 15.0 t dry matter [DM]/ha, termed as moderate; optimum pasture utilisation of 19.7 t DM/ha, termed as high) and 6 rates of replacement of each of these pastures by grazeable CFR (0%, 10%, 20%, 30%, 40%, 50%) were investigated. Results showed that AMS cows were required to walk greater than 1-km when the farm area was greater than 86 ha. Insufficient pasture could be produced within a 1 km distance (i.e. 86 ha land) with home-grown feed (HGF) providing 43%, 29%, and 22% of the metabolisable energy (ME) required by 400, 600, and 800 cows, respectively from pastures. Introduction of pasture (moderate): CFR in AMS at a ratio of 80:20 can feed a 400 cow AMS herd, and can supply 42% and 31% of the ME requirements for 600 and 800 cows, respectively with pasture (moderate): CFR at 50:50 levels. In contrast to moderate pasture, 400 cows can be managed on high pasture utilisation (provided 57% of the total ME requirements). However, similar to the scenarios conducted with moderate pasture, there was insufficient feed produced within 1-km distance of the dairy for 600 or 800 cows. An 800 cow herd required 140 and 130 ha on moderate and high pasture-based AMS system, respectively with the introduction of pasture: CFR at a ratio of 50:50. Given the impact of increasing land area past 86 ha on walking distance, cow numbers could be increased by purchasing feed from off the milking platform and/or using the land outside 1-km distance for conserved feed. However, this warrants further investigations into risk analyses of different management options including development of an innovative system to manage large herds in an AMS farming system. PMID:25925068
The cost of a large-scale hollow fibre MBR.
Verrecht, Bart; Maere, Thomas; Nopens, Ingmar; Brepols, Christoph; Judd, Simon
2010-10-01
A cost sensitivity analysis was carried out for a full-scale hollow fibre membrane bioreactor to quantify the effect of design choices and operational parameters on cost. Different options were subjected to a long term dynamic influent profile and evaluated using ASM1 for effluent quality, aeration requirements and sludge production. The results were used to calculate a net present value (NPV), incorporating both capital expenditure (capex), based on costs obtained from equipment manufacturers and full-scale plants, and operating expenditure (opex), accounting for energy demand, sludge production and chemical cleaning costs. Results show that the amount of contingency built in to cope with changes in feedwater flow has a large impact on NPV. Deviation from a constant daily flow increases NPV as mean plant utilisation decreases. Conversely, adding a buffer tank reduces NPV, since less membrane surface is required when average plant utilisation increases. Membrane cost and lifetime is decisive in determining NPV: an increased membrane replacement interval from 5 to 10 years reduces NPV by 19%. Operation at higher SRT increases the NPV, since the reduced costs for sludge treatment are offset by correspondingly higher aeration costs at higher MLSS levels, though the analysis is very sensitive to sludge treatment costs. A higher sustainable flux demands greater membrane aeration, but the subsequent opex increase is offset by the reduced membrane area and the corresponding lower capex. Copyright © 2010 Elsevier Ltd. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pugmire, David; Kress, James; Choi, Jong
Data driven science is becoming increasingly more common, complex, and is placing tremendous stresses on visualization and analysis frameworks. Data sources producing 10GB per second (and more) are becoming increasingly commonplace in both simulation, sensor and experimental sciences. These data sources, which are often distributed around the world, must be analyzed by teams of scientists that are also distributed. Enabling scientists to view, query and interact with such large volumes of data in near-real-time requires a rich fusion of visualization and analysis techniques, middleware and workflow systems. Here, this paper discusses initial research into visualization and analysis of distributed datamore » workflows that enables scientists to make near-real-time decisions of large volumes of time varying data.« less
Aerodynamic flight control to increase payload capability of future launch vehicles
NASA Technical Reports Server (NTRS)
Cochran, John E., Jr.
1995-01-01
The development of new launch vehicles will require that designers use innovative approaches to achieve greater performance in terms of pay load capability. The objective of the work performed under this delivery order was to provide technical assistance to the Contract Officer's Technical Representative (COTR) in the development of ideas and concepts for increasing the payload capability of launch vehicles by incorporating aerodynamic controls. Although aerodynamic controls, such as moveable fins, are currently used on relatively small missiles, the evolution of large launch vehicles has been moving away from aerodynamic control. The COTR reasoned that a closer investigation of the use of aerodynamic controls on large vehicles was warranted.
NASA Astrophysics Data System (ADS)
McFall, Steve
1994-03-01
With the increase in business automation and the widespread availability and low cost of computer systems, law enforcement agencies have seen a corresponding increase in criminal acts involving computers. The examination of computer evidence is a new field of forensic science with numerous opportunities for research and development. Research is needed to develop new software utilities to examine computer storage media, expert systems capable of finding criminal activity in large amounts of data, and to find methods of recovering data from chemically and physically damaged computer storage media. In addition, defeating encryption and password protection of computer files is also a topic requiring more research and development.
Intelligent redundant actuation system requirements and preliminary system design
NASA Technical Reports Server (NTRS)
Defeo, P.; Geiger, L. J.; Harris, J.
1985-01-01
Several redundant actuation system configurations were designed and demonstrated to satisfy the stringent operational requirements of advanced flight control systems. However, this has been accomplished largely through brute force hardware redundancy, resulting in significantly increased computational requirements on the flight control computers which perform the failure analysis and reconfiguration management. Modern technology now provides powerful, low-cost microprocessors which are effective in performing failure isolation and configuration management at the local actuator level. One such concept, called an Intelligent Redundant Actuation System (IRAS), significantly reduces the flight control computer requirements and performs the local tasks more comprehensively than previously feasible. The requirements and preliminary design of an experimental laboratory system capable of demonstrating the concept and sufficiently flexible to explore a variety of configurations are discussed.
Atmospheric considerations regarding the impact of heat dissipation from a nuclear energy center
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rotty, R.M.; Bauman, H.; Bennett, L.L.
1976-05-01
Potential changes in climate resulting from a large nuclear energy center are discussed. On a global scale, no noticeable changes are likely, but on both a regional and a local scale, changes can be expected. Depending on the cooling system employed, the amount of fog may increase, the amount and distribution of precipitation will change, and the frequency or location of severe storms may change. Very large heat releases over small surface areas can result in greater atmospheric instability; a large number of closely spaced natural-draft cooling towers have this disadvantage. On the other hand, employment of natural-draft towers makesmore » an increase in the occurrence of ground fog unlikely. The analysis suggests that the cooling towers for a large nuclear energy center should be located in clusters of four with at least 2.5-mile spacing between the clusters. This is equivalent to the requirement of one acre of land surface per each two megawatts of heat being rejected.« less
An efficient parallel-processing method for transposing large matrices in place.
Portnoff, M R
1999-01-01
We have developed an efficient algorithm for transposing large matrices in place. The algorithm is efficient because data are accessed either sequentially in blocks or randomly within blocks small enough to fit in cache, and because the same indexing calculations are shared among identical procedures operating on independent subsets of the data. This inherent parallelism makes the method well suited for a multiprocessor computing environment. The algorithm is easy to implement because the same two procedures are applied to the data in various groupings to carry out the complete transpose operation. Using only a single processor, we have demonstrated nearly an order of magnitude increase in speed over the previously published algorithm by Gate and Twigg for transposing a large rectangular matrix in place. With multiple processors operating in parallel, the processing speed increases almost linearly with the number of processors. A simplified version of the algorithm for square matrices is presented as well as an extension for matrices large enough to require virtual memory.
Johansen, J L; Pratchett, M S; Messmer, V; Coker, D J; Tobin, A J; Hoey, A S
2015-09-08
Increased ocean temperature due to climate change is raising metabolic demands and energy requirements of marine ectotherms. If productivity of marine systems and fisheries are to persist, individual species must compensate for this demand through increasing energy acquisition or decreasing energy expenditure. Here we reveal that the most important coral reef fishery species in the Indo-west Pacific, the large predatory coral trout Plectropomus leopardus (Serranidae), can behaviourally adjust food intake to maintain body-condition under elevated temperatures, and acclimate over time to consume larger meals. However, these increased energetic demands are unlikely to be met by adequate production at lower trophic levels, as smaller prey species are often the first to decline in response to climate-induced loss of live coral and structural complexity. Consequently, ubiquitous increases in energy consumption due to climate change will increase top-down competition for a dwindling biomass of prey, potentially distorting entire food webs and associated fisheries.
Large predatory coral trout species unlikely to meet increasing energetic demands in a warming ocean
Johansen, J.L.; Pratchett, M.S.; Messmer, V.; Coker, D.J.; Tobin, A.J.; Hoey, A.S.
2015-01-01
Increased ocean temperature due to climate change is raising metabolic demands and energy requirements of marine ectotherms. If productivity of marine systems and fisheries are to persist, individual species must compensate for this demand through increasing energy acquisition or decreasing energy expenditure. Here we reveal that the most important coral reef fishery species in the Indo-west Pacific, the large predatory coral trout Plectropomus leopardus (Serranidae), can behaviourally adjust food intake to maintain body-condition under elevated temperatures, and acclimate over time to consume larger meals. However, these increased energetic demands are unlikely to be met by adequate production at lower trophic levels, as smaller prey species are often the first to decline in response to climate-induced loss of live coral and structural complexity. Consequently, ubiquitous increases in energy consumption due to climate change will increase top-down competition for a dwindling biomass of prey, potentially distorting entire food webs and associated fisheries. PMID:26345733
Capstone Portfolios and Geography Student Learning Outcomes
ERIC Educational Resources Information Center
Mossa, Joann
2014-01-01
Due to increasing demands regarding student learning outcomes and accreditation, a capstone portfolio was added to assess critical thinking and communication skills of geography majors at a large public university in the USA. The portfolio guidelines were designed to be adaptable to a flexible curriculum where about half of the requirements within…
Student Engagement in Neo-Liberal Times: What Is Missing?
ERIC Educational Resources Information Center
Zepke, Nick
2018-01-01
Quality teaching is increasingly prioritized in higher education. One reason is that government funding requires students to succeed in their studies and be ready for employment. In response, educators throughout the Western world have generated large quantities of evidence-based, practical, often uncritical research about what works to improve…
Industrial Arts and Technology.
ERIC Educational Resources Information Center
Baird, David A.
Increased emphasis on academic and mathematical skills in high school courses such as "technology education" appeal only to the above average and motivated students, leaving a large majority of less-able students with a distorted view of future jobs, which do not, in fact, require such an academic approach. At the same time, industrial arts…
Increasing Immunization Compliance
ERIC Educational Resources Information Center
Toole, Kimberly; Perry, Cynthia S.
2004-01-01
School nurses often have the responsibility to ensure that students meet all immunization requirements for school entry and school attendance. In large inner-city school districts, many obstacles exist which make this task daunting and often result in lengthy absences and exclusions for students. It is critical that school nurses find creative and…
Service Blueprinting: Transforming the Student Experience
ERIC Educational Resources Information Center
Bitner, Mary Jo; Ostrom, Amy L.; Burkhard, Kevin A.
2012-01-01
There is much discussion today about the need to transform higher education for the benefit of students, employers, and society at large. Experts and researchers list the numerous challenges: low student retention and graduation rates, the increasing cost of higher education, and concerns that graduates don't possess the skills required to compete…
Proximity Displays for Access Control
ERIC Educational Resources Information Center
Vaniea, Kami
2012-01-01
Managing access to shared digital information, such as photographs and documents. is difficult for end users who are accumulating an increasingly large and diverse collection of data that they want to share with others. Current policy-management solutions require a user to proactively seek out and open a separate policy-management interface when…
Anxiety in Language Testing: The APTIS Case
ERIC Educational Resources Information Center
Valencia Robles, Jeannette de Fátima
2017-01-01
The requirement of holding a diploma which certifies proficiency level in a foreign language is constantly increasing in academic and working environments. Computer-based testing has become a prevailing tendency for these and other educational purposes. Each year large numbers of students take online language tests everywhere in the world. In…
Knockout of an outer membrane protein operon of anaplasma marginale by transposon mutagenesis
USDA-ARS?s Scientific Manuscript database
Large amounts of data generated by genomics, transcriptomics and proteomics technologies have increased our understanding of the biology of Anaplasma marginale. However, these data have also led to new assumptions that require testing, ideally through classic genetic mutation. One example is the def...
Comments on the MIT Assessment of the Mars One Plan
NASA Technical Reports Server (NTRS)
Jones, Harry W.
2015-01-01
The MIT assessment of the Mars One mission plan reveals design assumptions that would cause significant difficulties. Growing crops in the crew chamber produces excessive oxygen levels. The assumed in-situ resource utilization (ISRU) equipment has too low a Technology Readiness Level (TRL). The required spare parts cause a large and increasing launch mass logistics burden. The assumed International Space Station (ISS) Environmental Control and Life Support (ECLS) technologies were developed for microgravity and therefore are not suitable for Mars gravity. Growing food requires more mass than sending food from Earth. The large number of spares is due to the relatively low reliability of ECLS and the low TRL of ISRU. The Mars One habitat design is similar to past concepts but does not incorporate current knowledge. The MIT architecture analysis tool for long-term settlements on the Martian surface includes an ECLS system simulation, an ISRU sizing model, and an analysis of required spares. The MIT tool showed the need for separate crop and crew chambers, the large spare parts logistics, that crops require more mass than Earth food, and that more spares are needed if reliability is lower. That ISRU has low TRL and ISS ECLS was designed for microgravity are well known. Interestingly, the results produced by the architecture analysis tool - separate crop chamber, large spares mass, large crop chamber mass, and low reliability requiring more spares - were also well known. A common approach to ECLS architecture analysis is to build a complex model that is intended to be all-inclusive and is hoped will help solve all design problems. Such models can struggle to replicate obvious and well-known results and are often unable to answer unanticipated new questions. A better approach would be to survey the literature for background knowledge and then directly analyze the important problems.
Performance Investigations of a Large Centrifugal Compressor from an Experimental Turbojet Engine
NASA Technical Reports Server (NTRS)
Ginsburg, Ambrose; Creagh, John W. R.; Ritter, William K.
1948-01-01
An investigation was conducted on a large centrifugal compressor from an experimental turbojet engine to determine the performance of the compressor and to obtain fundamental information on the aerodynamic problems associated with large centrifugal-type compressors. The results of the research conducted on the compressor indicated that the compressor would not meet the desired engine-design air-flow requirements (78 lb/sec) because of an air-flow restriction in the vaned collector (diffuser). Revision of the vaned collector resulted in an increased air-flow capacity over the speed range and showed improved matching of the impeller and diffuser components. At maximum flow, the original compressor utilized approximately 90 percent of the available geometric throat area at the vaned-collector inlet and the revised compressor utilized approximately 94 percent, regardless of impeller speed. The ratio of the maximum weight flows of the revised and original compressors were less than the ratio of effective critical throat areas of the two compressors because of the large pressure losses in the impeller near the impeller inelt and the difference increased with an increase in impeller speed. In order to further increase the pressure ratio and maximum weight flow of the compressor, the impeller must be modified to eliminate the pressure losses therein.
Large optical glass blanks for the ELT generation
NASA Astrophysics Data System (ADS)
Jedamzik, Ralf; Petzold, Uwe; Dietrich, Volker; Wittmer, Volker; Rexius, Olga
2016-07-01
The upcoming extremely large telescope projects like the E-ELT, TMT or GMT telescopes require not only large amount of mirror blank substrates but have also sophisticated instrument setups. Common instrument components are atmospheric dispersion correctors that compensate for the varying atmospheric path length depending on the telescope inclination angle. These elements consist usually of optical glass blanks that have to be large due to the increased size of the focal beam of the extremely large telescopes. SCHOTT has a long experience in producing and delivering large optical glass blanks for astronomical applications up to 1 m and in homogeneity grades up to H3 quality in the past. The most common optical glass available in large formats is SCHOTT N-BK7. But other glass types like F2 or LLF1 can also be produced in formats up to 1 m. The extremely large telescope projects partly demand atmospheric dispersion components even in sizes beyond 1m up to a range of 1.5 m diameter. The production of such large homogeneous optical glass banks requires tight control of all process steps. To cover this demand in the future SCHOTT initiated a research project to improve the large optical blank production process steps from melting to annealing and measurement. Large optical glass blanks are measured in several sub-apertures that cover the total clear aperture of the application. With SCHOTT's new stitching software it is now possible to combine individual sub-aperture measurements to a total homogeneity map of the blank. In this presentation first results will be demonstrated.
Improving efficiency of polystyrene concrete production with composite binders
NASA Astrophysics Data System (ADS)
Lesovik, R. V.; Ageeva, M. S.; Lesovik, G. A.; Sopin, D. M.; Kazlitina, O. V.; Mitrokhina, A. A.
2018-03-01
According to leading marketing researchers, the construction market in Russia and CIS will continue growing at a rapid rate; this applies not only to a large-scale major construction, but to a construction of single-family houses and small-scale industrial facilities as well. Due to this, there are increased requirements for heat insulation of the building enclosures and a significant demand for efficient walling materials with high thermal performance. All these developments led to higher requirements imposed on the equipment that produces such materials.
A simple rule for the evolution of contingent cooperation in large groups
Schonmann, Roberto H.; Boyd, Robert
2016-01-01
Humans cooperate in large groups of unrelated individuals, and many authors have argued that such cooperation is sustained by contingent reward and punishment. However, such sanctioning systems can also stabilize a wide range of behaviours, including mutually deleterious behaviours. Moreover, it is very likely that large-scale cooperation is derived in the human lineage. Thus, understanding the evolution of mutually beneficial cooperative behaviour requires knowledge of when strategies that support such behaviour can increase when rare. Here, we derive a simple formula that gives the relatedness necessary for contingent cooperation in n-person iterated games to increase when rare. This rule applies to a wide range of pay-off functions and assumes that the strategies supporting cooperation are based on the presence of a threshold fraction of cooperators. This rule suggests that modest levels of relatedness are sufficient for invasion by strategies that make cooperation contingent on previous cooperation by a small fraction of group members. In contrast, only high levels of relatedness allow the invasion by strategies that require near universal cooperation. In order to derive this formula, we introduce a novel methodology for studying evolution in group structured populations including local and global group-size regulation and fluctuations in group size. PMID:26729938
Sawers, Andrew; Hahn, Michael E
2013-08-01
Motor learning strategies that increase practice difficulty and the size of movement errors are thought to facilitate motor learning. In contrast to this, gradual training minimizes movement errors and reduces practice difficulty by incrementally introducing task requirements, yet remains as effective as sudden training and its large movement errors for learning novel reaching tasks. While attractive as a locomotor rehabilitation strategy, it remains unknown whether the efficacy of gradual training extends to learning locomotor tasks and their unique requirements. The influence of gradual vs. sudden training on learning a locomotor task, asymmetric split belt treadmill walking, was examined by assessing whole body sagittal plane kinematics during 24 hour retention and transfer performance following either gradual or sudden training. Despite less difficult and less specific practice for the gradual cohort on day 1, gradual training resulted in equivalent motor learning of the novel locomotor task as sudden training when assessed by retention and transfer a day later. This suggests that large movement errors and increased practice difficulty may not be necessary for learning novel locomotor tasks. Further, gradual training may present a viable locomotor rehabilitation strategy avoiding large movement errors that could limit or impair improvements in locomotor performance. Copyright © 2013 Elsevier B.V. All rights reserved.
Energy Storage Requirements for Achieving 50% Penetration of Solar Photovoltaic Energy in California
DOE Office of Scientific and Technical Information (OSTI.GOV)
Denholm, Paul; Margolis, Robert
2016-09-01
We estimate the storage required to enable PV penetration up to 50% in California (with renewable penetration over 66%), and we quantify the complex relationships among storage, PV penetration, grid flexibility, and PV costs due to increased curtailment. We find that the storage needed depends strongly on the amount of other flexibility resources deployed. With very low-cost PV (three cents per kilowatt-hour) and a highly flexible electric power system, about 19 gigawatts of energy storage could enable 50% PV penetration with a marginal net PV levelized cost of energy (LCOE) comparable to the variable costs of future combined-cycle gas generatorsmore » under carbon constraints. This system requires extensive use of flexible generation, transmission, demand response, and electrifying one quarter of the vehicle fleet in California with largely optimized charging. A less flexible system, or more expensive PV would require significantly greater amounts of storage. The amount of storage needed to support very large amounts of PV might fit within a least-cost framework driven by declining storage costs and reduced storage-duration needs due to high PV penetration.« less
Energy Storage Requirements for Achieving 50% Solar Photovoltaic Energy Penetration in California
DOE Office of Scientific and Technical Information (OSTI.GOV)
Denholm, Paul; Margolis, Robert
2016-08-01
We estimate the storage required to enable PV penetration up to 50% in California (with renewable penetration over 66%), and we quantify the complex relationships among storage, PV penetration, grid flexibility, and PV costs due to increased curtailment. We find that the storage needed depends strongly on the amount of other flexibility resources deployed. With very low-cost PV (three cents per kilowatt-hour) and a highly flexible electric power system, about 19 gigawatts of energy storage could enable 50% PV penetration with a marginal net PV levelized cost of energy (LCOE) comparable to the variable costs of future combined-cycle gas generatorsmore » under carbon constraints. This system requires extensive use of flexible generation, transmission, demand response, and electrifying one quarter of the vehicle fleet in California with largely optimized charging. A less flexible system, or more expensive PV would require significantly greater amounts of storage. The amount of storage needed to support very large amounts of PV might fit within a least-cost framework driven by declining storage costs and reduced storage-duration needs due to high PV penetration.« less
Watermarking-based protection of remote sensing images: requirements and possible solutions
NASA Astrophysics Data System (ADS)
Barni, Mauro; Bartolini, Franco; Cappellini, Vito; Magli, Enrico; Olmo, Gabriella
2001-12-01
Earth observation missions have recently attracted ag rowing interest form the scientific and industrial communities, mainly due to the large number of possible applications capable to exploit remotely sensed data and images. Along with the increase of market potential, the need arises for the protection of the image products from non-authorized use. Such a need is a very crucial one even because the Internet and other public/private networks have become preferred means of data exchange. A crucial issue arising when dealing with digital image distribution is copyright protection. Such a problem has been largely addressed by resorting to watermarking technology. A question that obviously arises is whether the requirements imposed by remote sensing imagery are compatible with existing watermarking techniques. On the basis of these motivations, the contribution of this work is twofold: i) assessment of the requirements imposed by the characteristics of remotely sensed images on watermark-based copyright protection ii) analysis of the state-of-the-art, and performance evaluation of existing algorithms in terms of the requirements at the previous point.
Calibration method for a large-scale structured light measurement system.
Wang, Peng; Wang, Jianmei; Xu, Jing; Guan, Yong; Zhang, Guanglie; Chen, Ken
2017-05-10
The structured light method is an effective non-contact measurement approach. The calibration greatly affects the measurement precision of structured light systems. To construct a large-scale structured light system with high accuracy, a large-scale and precise calibration gauge is always required, which leads to an increased cost. To this end, in this paper, a calibration method with a planar mirror is proposed to reduce the calibration gauge size and cost. An out-of-focus camera calibration method is also proposed to overcome the defocusing problem caused by the shortened distance during the calibration procedure. The experimental results verify the accuracy of the proposed calibration method.
Accommodation-based liquid crystal adaptive optics system for large ocular aberration correction.
Mu, Quanquan; Cao, Zhaoliang; Li, Chao; Jiang, Baoguang; Hu, Lifa; Xuan, Li
2008-12-15
According to ocular aberration property and liquid crystal (LC) corrector characteristics, we calculated the minimum pixel demand of the LC corrector used for compensating large ocular aberrations. Then, an accommodation based optical configuration was introduced to reduce the demand. Based on this an adaptive optics (AO) retinal imaging system was built. Subjects with different defocus and astigmatism were tested to prove this. For myopia lower than 5D it performs well. When myopia is as large as 8D the accommodation error increased to nearly 3D, which requires the LC corrector to have 667 x 667 pixels to get a well-corrected image.
Affordable and Lightweight High-Resolution X-ray Optics for Astronomical Missions
NASA Technical Reports Server (NTRS)
Zhang, W. W.; Biskach, M. P.; Bly, V. T.; Carter, J. M.; Chan, K. W.; Gaskin, J. A.; Hong, M.; Hohl, B. R.; Jones, W. D.; Kolodziejczak, J. J.
2014-01-01
Future x-ray astronomical missions require x-ray mirror assemblies that provide both high angular resolution and large photon collecting area. In addition, as x-ray astronomy undertakes more sensitive sky surveys, a large field of view is becoming increasingly important as well. Since implementation of these requirements must be carried out in broad political and economical contexts, any technology that meets these performance requirements must also be financially affordable and can be implemented on a reasonable schedule. In this paper we report on progress of an x-ray optics development program that has been designed to address all of these requirements. The program adopts the segmented optical design, thereby is capable of making both small and large mirror assemblies for missions of any size. This program has five technical elements: (1) fabrication of mirror substrates, (2) coating, (3) alignment, (4) bonding, and (5) mirror module systems engineering and testing. In the past year we have made progress in each of these five areas, advancing the angular resolution of mirror modules from 10.8 arc-seconds half-power diameter reported (HPD) a year ago to 8.3 arc-seconds now. These mirror modules have been subjected to and passed all environmental tests, including vibration, acoustic, and thermal vacuum. As such this technology is ready for implementing a mission that requires a 10-arc-second mirror assembly. Further development in the next two years would make it ready for a mission requiring a 5-arc-second mirror assembly. We expect that, by the end of this decade, this technology would enable the x-ray astrophysical community to compete effectively for a major x-ray mission in the 2020s that would require one or more 1-arc-second mirror assemblies for imaging, spectroscopic, timing, and survey studies.
Approaches to sheltered-water oil spills
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jacobs, M.A.; Waldron, D.M.
Technology has produced more effective and efficient oil removal equipment for on-water cleanup in the past five years. Much of the innovation has been to increase recovery capacity to meet the planning volumes required to government regulations. However, more than 95 percent of the spills are relatively small. Large equipment, often requiring large platforms, is not very useful and is difficult/expensive to operate on small spills. In addition, damage from spills results when oil impacts shorelines. The emphasis on spill response should address the ability of the equipment to remove oil in a nearshore environment. Clean Seas has been attemptingmore » to address this need since the Avila Pipeline spill in 1992, in which a 180 barrel spill resulted in about $18 million damage/cleanup cost.« less
Gabriel, Lucinda E K; Webb, Steve A R
2013-10-01
Influenza pandemics occur intermittently and represent an existential global infectious diseases threat. The purpose of this review is to describe clinical and research preparedness for future pandemics. Pandemic influenza typically results in large numbers of individuals with life-threatening pneumonia requiring treatment in ICUs. Clinical preparedness of ICUs relates to planning to provide increased 'surge' capacity to meet increased demand and requires consideration of staffing, equipment and consumables, bed-space availability and management systems. Research preparedness is also necessary, as timely clinical research has the potential to change the trajectory of a pandemic. The clinical research response during the 2009 H1N1 influenza pandemic was suboptimal. Better planning is necessary to optimize both clinical and research responses to future pandemics.
NASA Astrophysics Data System (ADS)
Widiyanto, Sigit; Setyawan, Aris Budi; Tarigan, Avinanta; Sussanto, Herry
2016-02-01
The increase of the number of business impact on the increasing service requirements for companies and Small Medium Enterprises (SMEs) in submitting their license request. The service system that is needed must be able to accommodate a large number of documents, various institutions, and time limitations of applicant. In addition, it is also required distributed applications which is able to be integrated each other. Service oriented application fits perfectly developed along client-server application which has been developed by the Government to digitalize submitted data. RESTful architecture and MVC framework are embedded in developing application. As a result, the application proves its capability in solving security, transaction speed, and data accuracy issues.
NASA Technical Reports Server (NTRS)
1975-01-01
User technology requirements are identified in relation to needed technology advancement for future space missions in the areas of navigation, guidance, and control. Emphasis is placed on: reduction of mission support cost by 50% through autonomous operation, a ten-fold increase in mission output through improved pointing and control, and a hundred-fold increase in human productivity in space through large-scale teleoperator applications.
Water limited agriculture in Africa: Climate change sensitivity of large scale land investments
NASA Astrophysics Data System (ADS)
Rulli, M. C.; D'Odorico, P.; Chiarelli, D. D.; Davis, K. F.
2015-12-01
The past few decades have seen unprecedented changes in the global agricultural system with a dramatic increase in the rates of food production fueled by an escalating demand for food calories, as a result of demographic growth, dietary changes, and - more recently - new bioenergy policies. Food prices have become consistently higher and increasingly volatile with dramatic spikes in 2007-08 and 2010-11. The confluence of these factors has heightened demand for land and brought a wave of land investment to the developing world: some of the more affluent countries are trying to secure land rights in areas suitable for agriculture. According to some estimates, to date, roughly 38 million hectares have been acquired worldwide by large scale investors, 16 million of which in Africa. More than 85% of large scale land acquisitions in Africa are by foreign investors. Many land deals are motivated not only by the need for fertile land but for the water resources required for crop production. Despite some recent assessments of the water appropriation associated with large scale land investments, their impact on the water resources of the target countries under present conditions and climate change scenarios remains poorly understood. Here we investigate irrigation water requirements by various crops planted in the acquired land as an indicator of the pressure likely placed by land investors on ("blue") water resources of target regions in Africa and evaluate the sensitivity to climate changes scenarios.
Noise Reduction Potential of Large, Over-the-Wing Mounted, Advanced Turbofan Engines
NASA Technical Reports Server (NTRS)
Berton, Jeffrey J.
2000-01-01
As we look to the future, increasingly stringent civilian aviation noise regulations will require the design and manufacture of extremely quiet commercial aircraft. Indeed, the noise goal for NASA's Aeronautics Enterprise calls for technologies that will help to provide a 20 EPNdB reduction relative to today's levels by the year 2022. Further, the large fan diameters of modem, increasingly higher bypass ratio engines pose a significant packaging and aircraft installation challenge. One design approach that addresses both of these challenges is to mount the engines above the wing. In addition to allowing the performance trend towards large, ultra high bypass ratio cycles to continue, this over-the-wing design is believed to offer noise shielding benefits to observers on the ground. This paper describes the analytical certification noise predictions of a notional, long haul, commercial quadjet transport with advanced, high bypass engines mounted above the wing.
Shilov, Ignat V; Seymour, Sean L; Patel, Alpesh A; Loboda, Alex; Tang, Wilfred H; Keating, Sean P; Hunter, Christie L; Nuwaysir, Lydia M; Schaeffer, Daniel A
2007-09-01
The Paragon Algorithm, a novel database search engine for the identification of peptides from tandem mass spectrometry data, is presented. Sequence Temperature Values are computed using a sequence tag algorithm, allowing the degree of implication by an MS/MS spectrum of each region of a database to be determined on a continuum. Counter to conventional approaches, features such as modifications, substitutions, and cleavage events are modeled with probabilities rather than by discrete user-controlled settings to consider or not consider a feature. The use of feature probabilities in conjunction with Sequence Temperature Values allows for a very large increase in the effective search space with only a very small increase in the actual number of hypotheses that must be scored. The algorithm has a new kind of user interface that removes the user expertise requirement, presenting control settings in the language of the laboratory that are translated to optimal algorithmic settings. To validate this new algorithm, a comparison with Mascot is presented for a series of analogous searches to explore the relative impact of increasing search space probed with Mascot by relaxing the tryptic digestion conformance requirements from trypsin to semitrypsin to no enzyme and with the Paragon Algorithm using its Rapid mode and Thorough mode with and without tryptic specificity. Although they performed similarly for small search space, dramatic differences were observed in large search space. With the Paragon Algorithm, hundreds of biological and artifact modifications, all possible substitutions, and all levels of conformance to the expected digestion pattern can be searched in a single search step, yet the typical cost in search time is only 2-5 times that of conventional small search space. Despite this large increase in effective search space, there is no drastic loss of discrimination that typically accompanies the exploration of large search space.
Spatial and temporal changes in cumulative human impacts on the world's ocean.
Halpern, Benjamin S; Frazier, Melanie; Potapenko, John; Casey, Kenneth S; Koenig, Kellee; Longo, Catherine; Lowndes, Julia Stewart; Rockwood, R Cotton; Selig, Elizabeth R; Selkoe, Kimberly A; Walbridge, Shaun
2015-07-14
Human pressures on the ocean are thought to be increasing globally, yet we know little about their patterns of cumulative change, which pressures are most responsible for change, and which places are experiencing the greatest increases. Managers and policymakers require such information to make strategic decisions and monitor progress towards management objectives. Here we calculate and map recent change over 5 years in cumulative impacts to marine ecosystems globally from fishing, climate change, and ocean- and land-based stressors. Nearly 66% of the ocean and 77% of national jurisdictions show increased human impact, driven mostly by climate change pressures. Five percent of the ocean is heavily impacted with increasing pressures, requiring management attention. Ten percent has very low impact with decreasing pressures. Our results provide large-scale guidance about where to prioritize management efforts and affirm the importance of addressing climate change to maintain and improve the condition of marine ecosystems.
Spatial and temporal changes in cumulative human impacts on the world's ocean
Halpern, Benjamin S.; Frazier, Melanie; Potapenko, John; Casey, Kenneth S.; Koenig, Kellee; Longo, Catherine; Lowndes, Julia Stewart; Rockwood, R. Cotton; Selig, Elizabeth R.; Selkoe, Kimberly A.; Walbridge, Shaun
2015-01-01
Human pressures on the ocean are thought to be increasing globally, yet we know little about their patterns of cumulative change, which pressures are most responsible for change, and which places are experiencing the greatest increases. Managers and policymakers require such information to make strategic decisions and monitor progress towards management objectives. Here we calculate and map recent change over 5 years in cumulative impacts to marine ecosystems globally from fishing, climate change, and ocean- and land-based stressors. Nearly 66% of the ocean and 77% of national jurisdictions show increased human impact, driven mostly by climate change pressures. Five percent of the ocean is heavily impacted with increasing pressures, requiring management attention. Ten percent has very low impact with decreasing pressures. Our results provide large-scale guidance about where to prioritize management efforts and affirm the importance of addressing climate change to maintain and improve the condition of marine ecosystems. PMID:26172980
Bionimbus: a cloud for managing, analyzing and sharing large genomics datasets
Heath, Allison P; Greenway, Matthew; Powell, Raymond; Spring, Jonathan; Suarez, Rafael; Hanley, David; Bandlamudi, Chai; McNerney, Megan E; White, Kevin P; Grossman, Robert L
2014-01-01
Background As large genomics and phenotypic datasets are becoming more common, it is increasingly difficult for most researchers to access, manage, and analyze them. One possible approach is to provide the research community with several petabyte-scale cloud-based computing platforms containing these data, along with tools and resources to analyze it. Methods Bionimbus is an open source cloud-computing platform that is based primarily upon OpenStack, which manages on-demand virtual machines that provide the required computational resources, and GlusterFS, which is a high-performance clustered file system. Bionimbus also includes Tukey, which is a portal, and associated middleware that provides a single entry point and a single sign on for the various Bionimbus resources; and Yates, which automates the installation, configuration, and maintenance of the software infrastructure required. Results Bionimbus is used by a variety of projects to process genomics and phenotypic data. For example, it is used by an acute myeloid leukemia resequencing project at the University of Chicago. The project requires several computational pipelines, including pipelines for quality control, alignment, variant calling, and annotation. For each sample, the alignment step requires eight CPUs for about 12 h. BAM file sizes ranged from 5 GB to 10 GB for each sample. Conclusions Most members of the research community have difficulty downloading large genomics datasets and obtaining sufficient storage and computer resources to manage and analyze the data. Cloud computing platforms, such as Bionimbus, with data commons that contain large genomics datasets, are one choice for broadening access to research data in genomics. PMID:24464852
NASA Astrophysics Data System (ADS)
Fader, Marianela; Shi, Sinan; von Bloh, Werner; Bondeau, Alberte; Cramer, Wolfgang
2016-04-01
Irrigation in the Mediterranean is of vital importance for food security, employment and economic development. Our research shows that, at present, Mediterranean region could save 35% of water by implementing more efficient irrigation and conveyance systems. Some countries like Syria, Egypt and Turkey have higher saving potentials than others. Currently some crops, especially sugar cane and agricultural trees, consume in average more irrigation water per hectare than annual crops (1). Also under climate change, more efficient irrigation is of vital importance for counteracting increases in irrigation water requirements. The Mediterranean area as a whole might face an increase in gross irrigation requirements between 4% and 18% from climate change alone by the end of the century if irrigation systems and conveyance are not improved. Population growth increases these numbers to 22% and 74%, respectively, affecting mainly the Southern and Eastern Mediterranean. However, improved irrigation technologies and conveyance systems have large water saving potentials, especially in the Eastern Mediterranean, and may be able to compensate to some degree the increases due to climate change and population growth. Both subregions would need around 35% more water than today if they could afford some degree of modernization of irrigation and conveyance systems and benefit from the CO2-fertilization effect (1). However, in some scenarios (in this case as combinations of climate change, irrigation technology, influence of population growth and CO2-fertilization effect) water scarcity may constrain the supply of the irrigation water needed in future in Algeria, Libya, Israel, Jordan, Lebanon, Syria, Serbia, Morocco, Tunisia and Spain (1). In this study, vegetation growth, phenology, agricultural production and irrigation water requirements and withdrawal were simulated with the process-based ecohydrological and agro-ecosystem model LPJmL ("Lund-Potsdam-Jena managed Land") after a large development that comprised the improved representation of Mediterranean crops (2). References: (1) Fader, M., von Bloh, W., Shi, S., Bondeau, A., Cramer, W. (2015) : Mediterranean irrigation under climate change : More efficient irrigation needed to compensate increases in irrigation water requirements. HESSD 12, 8459-8504. (2) Fader, M., von Bloh, W., Shi, S., Bondeau, A., Cramer, W. (2015) : Modelling Mediterranean agro-ecosystems by including agricultural trees in the LPJmL model. Geosci. Model Dev., 8, 3545-3561, 2015.
Etonitazene as a reinforcer: oral intake of etonitazene by rhesus monkeys.
Carroll, M E; Meisch, R A
1978-12-08
Drinking of etonitazene HCI was studied in three rhesus monkeys during daily 3-h sessions. As the drug concentration was increased, the number of liquid deliveries decreased, and etonitazene intake (microgram/kg body weight) increased. As fixed-ratio (FR) requirements were increased, rate of responding increased, and liquid deliveries slightly decreased. When water was substituted for the drug, there was a large increase in responding for several sessions, followed by a slow decline to low rates. When etonitazene was reintroduced, responding abruptly increased to previous drug levels. These data suggest that etonitazene can serve as a positive reinforcer when taken orally by rhesus monkeys.
A Node Linkage Approach for Sequential Pattern Mining
Navarro, Osvaldo; Cumplido, René; Villaseñor-Pineda, Luis; Feregrino-Uribe, Claudia; Carrasco-Ochoa, Jesús Ariel
2014-01-01
Sequential Pattern Mining is a widely addressed problem in data mining, with applications such as analyzing Web usage, examining purchase behavior, and text mining, among others. Nevertheless, with the dramatic increase in data volume, the current approaches prove inefficient when dealing with large input datasets, a large number of different symbols and low minimum supports. In this paper, we propose a new sequential pattern mining algorithm, which follows a pattern-growth scheme to discover sequential patterns. Unlike most pattern growth algorithms, our approach does not build a data structure to represent the input dataset, but instead accesses the required sequences through pseudo-projection databases, achieving better runtime and reducing memory requirements. Our algorithm traverses the search space in a depth-first fashion and only preserves in memory a pattern node linkage and the pseudo-projections required for the branch being explored at the time. Experimental results show that our new approach, the Node Linkage Depth-First Traversal algorithm (NLDFT), has better performance and scalability in comparison with state of the art algorithms. PMID:24933123
Novikov, I; Fund, N; Freedman, L S
2010-01-15
Different methods for the calculation of sample size for simple logistic regression (LR) with one normally distributed continuous covariate give different results. Sometimes the difference can be large. Furthermore, some methods require the user to specify the prevalence of cases when the covariate equals its population mean, rather than the more natural population prevalence. We focus on two commonly used methods and show through simulations that the power for a given sample size may differ substantially from the nominal value for one method, especially when the covariate effect is large, while the other method performs poorly if the user provides the population prevalence instead of the required parameter. We propose a modification of the method of Hsieh et al. that requires specification of the population prevalence and that employs Schouten's sample size formula for a t-test with unequal variances and group sizes. This approach appears to increase the accuracy of the sample size estimates for LR with one continuous covariate.
Mass storage technology in networks
NASA Astrophysics Data System (ADS)
Ishii, Katsunori; Takeda, Toru; Itao, Kiyoshi; Kaneko, Reizo
1990-08-01
Trends and features of mass storage subsystems in network are surveyed and their key technologies spotlighted. Storage subsystems are becoming increasingly important in new network systems in which communications and data processing are systematically combined. These systems require a new class of high-performance mass-information storage in order to effectively utilize their processing power. The requirements of high transfer rates, high transactional rates and large storage capacities, coupled with high functionality, fault tolerance and flexibility in configuration, are major challenges in storage subsystems. Recent progress in optical disk technology has resulted in improved performance of on-line external memories to optical disk drives, which are competing with mid-range magnetic disks. Optical disks are more effective than magnetic disks in using low-traffic random-access file storing multimedia data that requires large capacity, such as in archive use and in information distribution use by ROM disks. Finally, it demonstrates image coded document file servers for local area network use that employ 130mm rewritable magneto-optical disk subsystems.
Repeated seizures induce long-term increase in hippocampal benzodiazepine receptors.
McNamara, J O; Peper, A M; Patrone, V
1980-01-01
Repeated seizures, whether induced by kindling or electroshock, caused a long-lasting (at least 24 hr) increase of [3H]diazepam binding in hippocampal membranes of Sprague-Dawley rats. Scatchard analyses demonstrated that increased numbers of binding sites accounted for the increase. Neither repeated hypoxia nor repeated administration of electrical current without inducing seizures caused an increase of [3H]diazepam binding. Regardless of the method used for seizure induction, the response was graded in that large numbers of seizures were required to induce significant increases, whereas fewer seizures induced only slight increases. We suggest that the receptor increases imply a heightened response to benzodiazepines and more powerful hippocampal recurrent inhibition. PMID:6930682
Lumb, Andrew B; Nair, Sindhu
2010-03-01
Breathing increased fractional oxygen concentration (FiO2) is recommended for the treatment of tissue ischaemia. The theoretical benefits of increasing FiO2 on tissue oxygenation were evaluated using standard physiological equations. Assuming constant oxygen consumption by tissues throughout the length of a capillary, the oxygen content at 20 arbitrary points along a capillary was calculated. Using mathematical representations of the haemoglobin dissociation curve and an iterative approach to include the dissolved oxygen component of oxygen content, the oxygen partial pressure (PO2) profile along a capillary was estimated. High FiO2 concentrations cause large increases in PO2 at the arteriolar end of capillaries but these large PO2 values, caused by the extra dissolved oxygen, rapidly decline along the capillary. At the venular end of the capillary (the area of tissue most likely to be hypoxic), breathing oxygen causes only a modest improvement in PO2. Increasing FiO2 to treat tissue hypoxia has clear benefits, but a multimodal approach to management is required.
Marran, K J; Davey, B; Lang, A; Segal, D G
2013-04-10
Postprandial glucose excursions contribute significantly to average blood glucose, glycaemic variability and cardiovascular risk. Carbohydrate counting is a method of insulin dosing that balances carbohydrate load to insulin dose using a fixed ratio. Many patients and current insulin pumps calculate insulin delivery for meals based on a linear carbohydrate-to-insulin relationship. It is our hypothesis that a non-linear relationship exists between the amounts of carbohydrate consumed and the insulin required to cover it. To document blood glucose exposure in response to increasing carbohydrate loads on fixed carbohydrate-to-insulin ratios. Five type 1 diabetic subjects receiving insulin pump therapy with good control were recruited. Morning basal rates and carbohydrate- to-insulin ratios were optimised. A Medtronic glucose sensor was used for 5 days to collect data for area-under-the-curve (AUC) analysis, during which standardised meals of increasing carbohydrate loads were consumed. Increasing carbohydrate loads using a fixed carbohydrate-to-insulin ratio resulted in increasing glucose AUC. The relationship was found to be exponential rather than linear. Late postprandial hypoglycaemia followed carbohydrate loads of >60 g and this was often followed by rebound hyperglycaemia that lasted >6 hours. A non-linear relationship exists between carbohydrates consumed and the insulin required to cover them. This has implications for control of postprandial blood sugars, especially when consuming large carbohydrate loads. Further studies are required to look at the optimal ratios, duration and type of insulin boluses required to cover increasing carbohydrate loads.
NASA Astrophysics Data System (ADS)
Fader, M.; Shi, S.; von Bloh, W.; Bondeau, A.; Cramer, W.
2015-08-01
Irrigation in the Mediterranean is of vital importance for food security, employment and economic development. This study systematically assesses how climate change and increases in atmospheric CO2 concentrations may affect irrigation requirements in the Mediterranean region by 2080-2090. Future demographic change and technological improvements in irrigation systems are accounted for, as is the spread of climate forcing, warming levels and potential realization of the CO2-fertilization effect. Vegetation growth, phenology, agricultural production and irrigation water requirements and withdrawal were simulated with the process-based ecohydrological and agro-ecosystem model LPJmL after a large development that comprised the improved representation of Mediterranean crops. At present the Mediterranean region could save 35 % of water by implementing more efficient irrigation and conveyance systems. Some countries like Syria, Egypt and Turkey have higher saving potentials than others. Currently some crops, especially sugar cane and agricultural trees, consume in average more irrigation water per hectare than annual crops. Different crops show different magnitude of changes in net irrigation requirements due to climate change, being the increases most pronounced in agricultural trees. The Mediterranean area as a whole might face an increase in gross irrigation requirements between 4 and 18 % from climate change alone if irrigation systems and conveyance are not improved (2 °C global warming combined with full CO2-fertilization effect, and 5 °C global warming combined with no CO2-fertilization effect, respectively). Population growth increases these numbers to 22 and 74 %, respectively, affecting mainly the Southern and Eastern Mediterranean. However, improved irrigation technologies and conveyance systems have large water saving potentials, especially in the Eastern Mediterranean, and may be able to compensate to some degree the increases due to climate change and population growth. Both subregions would need around 35 % more water than today if they could afford some degree of modernization of irrigation and conveyance systems and benefit from the CO2-fertilization effect. Nevertheless, water scarcity might pose further challenges to the agricultural sector: Algeria, Libya, Israel, Jordan, Lebanon, Syria, Serbia, Morocco, Tunisia and Spain have a high risk of not being able to sustainably meet future irrigation water requirements in some scenarios. The results presented in this study point to the necessity of performing further research on climate-friendly agro-ecosystems in order to assess, on the one side, their degree of resilience to climate shocks, and on the other side, their adaptation potential when confronted with higher temperatures and changes in water availability.
Probing failure susceptibilities of earthquake faults using small-quake tidal correlations.
Brinkman, Braden A W; LeBlanc, Michael; Ben-Zion, Yehuda; Uhl, Jonathan T; Dahmen, Karin A
2015-01-27
Mitigating the devastating economic and humanitarian impact of large earthquakes requires signals for forecasting seismic events. Daily tide stresses were previously thought to be insufficient for use as such a signal. Recently, however, they have been found to correlate significantly with small earthquakes, just before large earthquakes occur. Here we present a simple earthquake model to investigate whether correlations between daily tidal stresses and small earthquakes provide information about the likelihood of impending large earthquakes. The model predicts that intervals of significant correlations between small earthquakes and ongoing low-amplitude periodic stresses indicate increased fault susceptibility to large earthquake generation. The results agree with the recent observations of large earthquakes preceded by time periods of significant correlations between smaller events and daily tide stresses. We anticipate that incorporating experimentally determined parameters and fault-specific details into the model may provide new tools for extracting improved probabilities of impending large earthquakes.
Empirical relationships between tree fall and landscape-level amounts of logging and fire
Blanchard, Wade; Blair, David; McBurney, Lachlan; Stein, John; Banks, Sam C.
2018-01-01
Large old trees are critically important keystone structures in forest ecosystems globally. Populations of these trees are also in rapid decline in many forest ecosystems, making it important to quantify the factors that influence their dynamics at different spatial scales. Large old trees often occur in forest landscapes also subject to fire and logging. However, the effects on the risk of collapse of large old trees of the amount of logging and fire in the surrounding landscape are not well understood. Using an 18-year study in the Mountain Ash (Eucalyptus regnans) forests of the Central Highlands of Victoria, we quantify relationships between the probability of collapse of large old hollow-bearing trees at a site and the amount of logging and the amount of fire in the surrounding landscape. We found the probability of collapse increased with an increasing amount of logged forest in the surrounding landscape. It also increased with a greater amount of burned area in the surrounding landscape, particularly for trees in highly advanced stages of decay. The most likely explanation for elevated tree fall with an increasing amount of logged or burned areas in the surrounding landscape is change in wind movement patterns associated with cutblocks or burned areas. Previous studies show that large old hollow-bearing trees are already at high risk of collapse in our study area. New analyses presented here indicate that additional logging operations in the surrounding landscape will further elevate that risk. Current logging prescriptions require the protection of large old hollow-bearing trees on cutblocks. We suggest that efforts to reduce the probability of collapse of large old hollow-bearing trees on unlogged sites will demand careful landscape planning to limit the amount of timber harvesting in the surrounding landscape. PMID:29474487
Empirical relationships between tree fall and landscape-level amounts of logging and fire.
Lindenmayer, David B; Blanchard, Wade; Blair, David; McBurney, Lachlan; Stein, John; Banks, Sam C
2018-01-01
Large old trees are critically important keystone structures in forest ecosystems globally. Populations of these trees are also in rapid decline in many forest ecosystems, making it important to quantify the factors that influence their dynamics at different spatial scales. Large old trees often occur in forest landscapes also subject to fire and logging. However, the effects on the risk of collapse of large old trees of the amount of logging and fire in the surrounding landscape are not well understood. Using an 18-year study in the Mountain Ash (Eucalyptus regnans) forests of the Central Highlands of Victoria, we quantify relationships between the probability of collapse of large old hollow-bearing trees at a site and the amount of logging and the amount of fire in the surrounding landscape. We found the probability of collapse increased with an increasing amount of logged forest in the surrounding landscape. It also increased with a greater amount of burned area in the surrounding landscape, particularly for trees in highly advanced stages of decay. The most likely explanation for elevated tree fall with an increasing amount of logged or burned areas in the surrounding landscape is change in wind movement patterns associated with cutblocks or burned areas. Previous studies show that large old hollow-bearing trees are already at high risk of collapse in our study area. New analyses presented here indicate that additional logging operations in the surrounding landscape will further elevate that risk. Current logging prescriptions require the protection of large old hollow-bearing trees on cutblocks. We suggest that efforts to reduce the probability of collapse of large old hollow-bearing trees on unlogged sites will demand careful landscape planning to limit the amount of timber harvesting in the surrounding landscape.
Critical review: Uncharted waters? The future of the electricity-water nexus.
Sanders, Kelly T
2015-01-06
Electricity generation often requires large amounts of water, most notably for cooling thermoelectric power generators and moving hydroelectric turbines. This so-called "electricity-water nexus" has received increasing attention in recent years by governments, nongovernmental organizations, industry, and academics, especially in light of increasing water stress in many regions around the world. Although many analyses have attempted to project the future water requirements of electricity generation, projections vary considerably due to differences in temporal and spatial boundaries, modeling frameworks, and scenario definitions. This manuscript is intended to provide a critical review of recent publications that address the future water requirements of electricity production and define the factors that will moderate the water requirements of the electric grid moving forward to inform future research. The five variables identified include changes in (1) fuel consumption patterns, (2) cooling technology preferences, (3) environmental regulations, (4) ambient climate conditions, and (5) electric grid characteristics. These five factors are analyzed to provide guidance for future research related to the electricity-water nexus.
The Use of Environmental Test Facilities for Purposes Beyond Their Original Design
NASA Technical Reports Server (NTRS)
Fisher, Terry C.; Marner, W. J.
2000-01-01
Increasing demands from space flight project offices are requiring environmental testing facilities to become more versatile with increased capabilities. At the same time, maintaining a cost-effective approach to test operations has driven efforts to use these facilities for purposes beyond their original design. This paper presents an overview of the Jet Propulsion Laboratory's efforts to provide JPL's space flight projects with test facilities to meet unique test requirements and to serve the needs of selected outside customers. The large number of recent Mars Missions, including the Mars Pathfinder project, have required testing of components and systems in a Martian surface environment in facilities originally designed for deep space testing. The unique problems associated with performing these tests are discussed, along with practical solutions. Other unique test requirements are discussed including the use of space simulation chambers for testing high altitude balloon gondolas and the use of vacuum chambers for system level test firing of an ion propulsion engine.
Analysis of complications from abdominoplasty: a review of 206 cases at a university hospital.
Neaman, Keith C; Hansen, Juliana E
2007-03-01
The number of abdominoplasties performed in the United States has been steadily increasing over the past decade. A large proportion of these patients are bariatric patients who remain obese despite prior weight-reduction surgery. This study was done to review the experience of patients undergoing abdominoplasty at a university hospital. A retrospective chart review of 206 consecutive patients was performed. The overall complication rate was 37.4%. Major complications [hematoma requiring surgical intervention, seroma requiring aspiration or surgical drainage, cellulitis or abscess requiring hospitalization and intravenous (IV) antibiotics, deep vein thrombosis (DVT), and pulmonary embolism (PE)] occurred in 16% of patients. The rate of minor complications (hematoma or seroma requiring no intervention, epidermolysis, small-wound dehiscence, neuropathic pain, and minor cellulitis) was 26.7%. Obese patients had a significantly increased risk of developing major complications as compared with nonobese patients (53.4% versus 28.8%, P = 0.001). An in-depth analysis of all complications and risk factors was done.
The Reach up Early Childhood Parenting Program: Origins, Content, and Implementation
ERIC Educational Resources Information Center
Walker, Susan P.; Chang, Susan M.; Smith, Joanne A.; Baker-Henningham, Helen
2018-01-01
Nurturing care in early childhood requires responsive interactions and opportunities to learn; however, there are few large-scale programs in low- and middle-income countries that support parents' ability to provide responsive care and activities that help children learn. The Reach Up training program was developed to increase capacity of…
Rice Water use efficiency and yield under continuous and intermittent irrigation
USDA-ARS?s Scientific Manuscript database
In the Brazilian state of Rio Grande do Sul, rice (Oryza sativa L.) is predominantly grown using continuous fl ood irrigation, which requires large quantities of fresh water. Due to increasing scarcity and demand for water, modern agricultural systems need to produce more food with less water. Th e ...
Stephanie J. Wessell-Kelly; Deanna H. Olson
2013-01-01
Increasing global demands on forest resources are driving large-scale shifts toward plantation forestry. Simultaneously balancing resource extraction and ecological sustainability objectives in plantation forests requires the incorporation of innovative silvicultural strategies such as leave islands (green-tree retention clusters). Our primary research goal was to...
Optimize Resources and Help Reduce Cost of Ownership with Dell[TM] Systems Management
ERIC Educational Resources Information Center
Technology & Learning, 2008
2008-01-01
Maintaining secure, convenient administration of the PC system environment can be a significant drain on resources. Deskside visits can greatly increase the cost of supporting a large number of computers. Even simple tasks, such as tracking inventory or updating software, quickly become expensive when they require physically visiting every…
Fostering Eroticism in Science Education to Promote Erotic Generosities for the Ocean-Other
ERIC Educational Resources Information Center
Luther, Rachel
2013-01-01
Despite the increase in marine science curriculum in secondary schools, marine science is not generally required curricula and has been largely deemphasized or ignored in relation to earth science, biology, chemistry, and physics. I call for the integration and implementation of marine science more fully in secondary science education through…
Automated Acquisition, Cataloging, and Circulation in a Large Research Library.
ERIC Educational Resources Information Center
Boylan, Merle N.; And Others
This report describes automated procedures now in use for book acquisition, and book and document cataloging and circulation, in the library at Lawrence Radiation Laboratory, Livermore. The purpose of the automation is to increase frequency and accuracy of record updatings, decrease the time required to maintain records, improve the formats of the…
Bridge Programs in Illinois: Summaries, Outcomes, and Cross-Site Findings
ERIC Educational Resources Information Center
Bragg, D.; Harmon, T.; Kirby, C.; Kim, S.
2010-01-01
An increasing number of jobs in today's workforce require postsecondary education, yet large numbers of workers lack the essential skills and credentials to fill these jobs. The result is that many workers remain underemployed, reaching a ceiling early in their working careers. In 2007, the Joyce Foundation launched the Shifting Gears initiative…
Postfire seeding for erosion control: Effectiveness and impacts on native plant communities
Jan L. Beyers
2004-01-01
Abstract: Large, high-severity wildfires remove vegetation cover and expose mineral soil, ofen causing erosion and runoff during postfire rain events to increase dramatically. Land-management agencies in the United States are required to assess site conditions after wildfire and, where necessary, implement emergency watershed rehabilitation measures to help...
Incentivizing Multiple Revisions Improves Student Writing without Increasing Instructor Workload
ERIC Educational Resources Information Center
Stellmack, Mark A.; Sandidge, Rita R.; Sippl, Amy L.; Miller, Danneka J.
2015-01-01
Previous research has shown that when students are required to submit a draft and a revision of their writing, large proportions of students do not improve across drafts. We implemented a writing assignment in which students were permitted to submit up to four optional drafts. To encourage substantive revisions, students were awarded additional…
ERIC Educational Resources Information Center
Austin-Hickey, Rachel
2013-01-01
Community colleges play an important role in the accessibility of higher education to the American population and developmental coursework is of vital importance to college degree attainment. The large demand for student remediation in math requires optimal commitment of developmental math faculty members. Increased organizational commitment has…
Industries in the United States generate large volumes of non-hazardous wastes, sludges, by-products, and spent materials that require disposal or other end-of-life management. Solid-waste management stakeholders are increasingly employing or researching methods for beneficial r...
Building to Scale: An Analysis of Web-Based Services in CIC (Big Ten) Libraries.
ERIC Educational Resources Information Center
Dewey, Barbara I.
Advancing library services in large universities requires creative approaches for "building to scale." This is the case for CIC, Committee on Institutional Cooperation (Big Ten), libraries whose home institutions serve thousands of students, faculty, staff, and others. Developing virtual Web-based services is an increasingly viable…
ERIC Educational Resources Information Center
Sabalauskas, Kara L.; Ortolani, Charles L.; McCall, Matthew J.
2014-01-01
Child welfare providers are increasingly required to demonstrate that strengths-based, evidence-informed practices are central to their intervention methodology. This case study describes how a large child welfare agency instituted cognitive behavioural therapy (CBT) as the core component of a strength-based practice model with the goal of…
Computer-Based Assessment of Complex Problem Solving: Concept, Implementation, and Application
ERIC Educational Resources Information Center
Greiff, Samuel; Wustenberg, Sascha; Holt, Daniel V.; Goldhammer, Frank; Funke, Joachim
2013-01-01
Complex Problem Solving (CPS) skills are essential to successfully deal with environments that change dynamically and involve a large number of interconnected and partially unknown causal influences. The increasing importance of such skills in the 21st century requires appropriate assessment and intervention methods, which in turn rely on adequate…
ERIC Educational Resources Information Center
Parker, Loran Carleton; Gleichsner, Alyssa M.; Adedokun, Omolola A.; Forney, James
2016-01-01
Transformation of research in all biological fields necessitates the design, analysis and, interpretation of large data sets. Preparing students with the requisite skills in experimental design, statistical analysis, and interpretation, and mathematical reasoning will require both curricular reform and faculty who are willing and able to integrate…
ERIC Educational Resources Information Center
Matthews, Kelly E.; Adams, Peter; Goos, Merrilyn
2010-01-01
Modern biological sciences require practitioners to have increasing levels of knowledge, competence, and skills in mathematics and programming. A recent review of the science curriculum at the University of Queensland, a large, research-intensive institution in Australia, resulted in the development of a more quantitatively rigorous undergraduate…
Federal Register 2010, 2011, 2012, 2013, 2014
2012-11-20
... trading and clearing of foreign exchange swaps and foreign exchange forwards would create systemic risk... clearing and exchange trading requirements on the foreign exchange market would increase systemic risk by... argue that the exemption would create a large regulatory loophole that could exacerbate systemic risk...
Affective Experiences of International and Home Students during the Information Search Process
ERIC Educational Resources Information Center
Haley, Adele Nicole; Clough, Paul
2017-01-01
An increasing number of students are studying abroad requiring that they interact with information in languages other than their mother tongue. The UK in particular has seen a large growth in international students within Higher Education. These nonnative English speaking students present a distinct user group for university information services,…
USDA-ARS?s Scientific Manuscript database
Potato breeding cycles typically last 6-7 years because of the modest seed multiplication rate and large number of traits required of new varieties. Genomic selection has the potential to increase genetic gain per unit of time, through higher accuracy and/or a shorter cycle. Both possibilities were ...
Student Thoughts and Perceptions on Curriculum Reform
ERIC Educational Resources Information Center
VanderJagt, Douglas D.
2013-01-01
The purpose of this qualitative case study was to examine how students experience and respond to Michigan's increased graduation requirements. The study was conducted in a large, suburban high school that instituted a change to a trimester system in response to the state mandate. A criterion-based sample of 16 students, both college bound and…
NASA Astrophysics Data System (ADS)
Gur, David; Rockette, Howard E.; Sumkin, Jules H.; Hoy, Ronald J.; Feist, John H.; Thaete, F. Leland; King, Jill L.; Slasky, B. S.; Miketic, Linda M.; Straub, William H.
1991-07-01
In a series of large ROC studies, the authors analyzed the time radiologists took to diagnose PA chest images as a function of observer performance indices (Az), display environments, and difficulty of cases. Board-certified radiologists interpreted at least 600 images each for the presence or absence of one or more of the following abnormalities: interstitial disease, nodule, and pneumothorax. Results indicated that there exists a large inter- reader variability in the time required to diagnose PA chest images. There is no correlation between a reader's specific median reading time and his/her performance. Time generally increases as the number of abnormalities on a single image increases and for cases with subtle abnormalities. Results also indicated that, in general, the longer the time for interpretation of a specific case (within reader), the further the observer's confidence ratings were from the truth. These findings were found to hold true regardless of the display mode. These results may have implications with regards to the appropriate methodology that should be used for imaging systems evaluations and for measurements of productivity for radiologists.
Proteinortho: detection of (co-)orthologs in large-scale analysis.
Lechner, Marcus; Findeiss, Sven; Steiner, Lydia; Marz, Manja; Stadler, Peter F; Prohaska, Sonja J
2011-04-28
Orthology analysis is an important part of data analysis in many areas of bioinformatics such as comparative genomics and molecular phylogenetics. The ever-increasing flood of sequence data, and hence the rapidly increasing number of genomes that can be compared simultaneously, calls for efficient software tools as brute-force approaches with quadratic memory requirements become infeasible in practise. The rapid pace at which new data become available, furthermore, makes it desirable to compute genome-wide orthology relations for a given dataset rather than relying on relations listed in databases. The program Proteinortho described here is a stand-alone tool that is geared towards large datasets and makes use of distributed computing techniques when run on multi-core hardware. It implements an extended version of the reciprocal best alignment heuristic. We apply Proteinortho to compute orthologous proteins in the complete set of all 717 eubacterial genomes available at NCBI at the beginning of 2009. We identified thirty proteins present in 99% of all bacterial proteomes. Proteinortho significantly reduces the required amount of memory for orthology analysis compared to existing tools, allowing such computations to be performed on off-the-shelf hardware.
NASA Technical Reports Server (NTRS)
1980-01-01
Twenty-four functional requirements were prepared under six categories and serve to indicate how to integrate dispersed storage generation (DSG) systems with the distribution and other portions of the electric utility system. Results indicate that there are no fundamental technical obstacles to prevent the connection of dispersed storage and generation to the distribution system. However, a communication system of some sophistication is required to integrate the distribution system and the dispersed generation sources for effective control. The large-size span of generators from 10 KW to 30 MW means that a variety of remote monitoring and control may be required. Increased effort is required to develop demonstration equipment to perform the DSG monitoring and control functions and to acquire experience with this equipment in the utility distribution environment.
NASA Technical Reports Server (NTRS)
Shidner, Jeremy D.; Davis, Jody L.; Cianciolo, Alicia D.; Samareh, Jamshid A.; Powell, RIchard W.
2010-01-01
Landing on Mars has been a challenging task. Past NASA missions have shown resilience to increases in spacecraft mass by scaling back requirements such as landing site altitude, landing site location and arrival time. Knowledge of the partials relating requirements to mass is critical for mission designers to understand so that the project can retain margin throughout the process. Looking forward to new missions that will land 1.5 metric tons or greater, the current level of technology is insufficient, and new technologies will need to be developed. Understanding the sensitivity of these new technologies to requirements is the purpose of this paper.
Dynamic behavior of particles in spacecraft
NASA Technical Reports Server (NTRS)
Perrine, B. S.
1981-01-01
The behavior of particles relative to a spacecraft frame of reference was examined. Significant spatial excursions of particles in space can occur relative to the spacecraft frame of reference as a result of drag deceleration of the vehicle. These vehicle excursions tend to be large as time increases. Thus, if the particle is required to remain in a specified volume, constraints may be required. Thus, for example, in levitation experiments it may be extremely difficult to turn off the forces of constraint which keep the particles in a specified region. This means experiments which are sensitive to disturbances may be very difficult to perform if perturbation forces are required to be absent.
SAGA: A project to automate the management of software production systems
NASA Technical Reports Server (NTRS)
Campbell, Roy H.; Beckman-Davies, C. S.; Benzinger, L.; Beshers, G.; Laliberte, D.; Render, H.; Sum, R.; Smith, W.; Terwilliger, R.
1986-01-01
Research into software development is required to reduce its production cost and to improve its quality. Modern software systems, such as the embedded software required for NASA's space station initiative, stretch current software engineering techniques. The requirements to build large, reliable, and maintainable software systems increases with time. Much theoretical and practical research is in progress to improve software engineering techniques. One such technique is to build a software system or environment which directly supports the software engineering process, i.e., the SAGA project, comprising the research necessary to design and build a software development which automates the software engineering process. Progress under SAGA is described.
Advanced technology requirements for large space structures. Part 5: Atlas program requirements
NASA Technical Reports Server (NTRS)
Katz, E.; Lillenas, A. N.; Broddy, J. A.
1977-01-01
The results of a special study which identifies and assigns priorities to technology requirements needed to accomplish a particular scenario of future large area space systems are described. Proposed future systems analyzed for technology requirements included large Electronic Mail, Microwave Radiometer, and Radar Surveillance Satellites. Twenty technology areas were identified as requirements to develop the proposed space systems.
NASA Technical Reports Server (NTRS)
Farley, Rodger E.; Quinn, David A.; Brodeur, Stephen J. (Technical Monitor)
2001-01-01
With the success of the Hubble Space Telescope, it has become apparent that new frontiers of science and discovery are made every time an improvement in imaging resolution is made. For the HST working primarily in the visible and near-visible spectrum, this meant designing, building, and launching a primary mirror approximately three meters in diameter. Conventional thinking tells us that accomplishing a comparable improvement in resolution at longer wavelengths for Earth and Space Science applications requires a corresponding increase in the size of the primary mirror. For wavelengths in the sub-millimeter range, a very large telescope with an effective aperture in excess of one kilometer in diameter would be needed to obtain high quality angular resolution. Realistically a single aperture this large is practically impossible. Fortunately such large apertures can be constructed synthetically. Possibly as few as three 34 meter diameter mirrors flying in precision formation could be used to collect light at these longer wavelengths permitting not only very large virtual aperture science to be carried out, but high-resolution interferometry as well. To ensure the longest possible mission duration, a system of tethered spacecraft will be needed to mitigate the need for a great deal of propellant. A spin-stabilized, tethered formation will likely meet these requirements. Several configurations have been proposed which possibly meet the needs of the Space Science community. This paper discusses two of them, weighing the relative pros and cons of each concept. The ultimate goal being to settle on a configuration which combines the best features of structure, tethers, and formation flying to meet the ambitious requirements necessary to make future large synthetic aperture and interferometric science missions successful.
NASA Technical Reports Server (NTRS)
Farley, Rodger E.; Quinn, David A.
2004-01-01
With the success of the Hubble Space Telescope, it has become apparent that new frontiers of science and discovery are made every time an improvement in imaging resolution is made. For the HST working primarily in the visible and near-visible spectrum, this meant designing, building and launching a primary mirror approximately three meters in diameter. Conventional thinking tells us that accomplishing a comparable improvement in resolution at longer wavelengths for Earth and Space Science applications requires a corresponding increase in the size of the primary mirror. For wavelengths in the sub-millimeter range, a very large telescope with an effective aperture in excess of one kilometer in diameter would be needed to obtain high quality angular resolution. Realistically a single aperture this large is practically impossible. Fortunately such large apertures can be constructed synthetically. Possibly as few as three 3 - 4 meter diameter mirrors flying in precision formation could be used to collect light at these longer wavelengths permitting not only very large virtual aperture science to be carried out, but high-resolution interferometry as well. To ensure the longest possible mission duration, a system of tethered spacecraft will be needed to mitigate the need for a great deal of propellant. A spin-stabilized, tethered formation will likely meet these requirements. Several configurations have been proposed which possibly meet the needs of the Space Science community. This paper discusses two of them, weighing the relative pros and cons of each concept. The ultimate goal being to settle on a configuration which combines the best features of structure, tethers and formation flying to meet the ambitious requirements necessary to make future large synthetic aperture and interferometric science missions successful.
Quinley, J C; Baker, T D
1986-01-01
Historically, the Agency for International Development (AIDS) health budget has been closely tied to overall development spending. A large increase in the international health appropriations in 1984 broke this pattern. Investigation shows that active grass roots organizing and congressional lobbying are the most likely responsible factors in the increase. Maintenance and expansion of this success will require increased recognition of and participation in these activities by individuals and organizations involved in international health. PMID:3717466
Does parental consent for birth control affect underage pregnancy rates? The case of Texas.
Girma, Sourafel; Paton, David
2013-12-01
Previous work based on conjectural responses of minors predicted that the 2003 Texas requirement for parental consent for state-funded birth control to minors would lead to a large increase in underage pregnancies. We use state- and county-level data to test this prediction. The latter allow us to compare the impact of parental consent in counties with and without state-funded family planning clinics. We control for characteristics systematically correlated with the presence of state-funded clinics by combining difference-in-difference estimation with propensity score-weighted regressions. The evidence suggests that the parental consent mandate led to a large decrease in attendance at family planning clinics among teens but did not lead to an increase in underage pregnancies.
Size-dependent enhancement of water relations during post-fire resprouting.
Schafer, Jennifer L; Breslow, Bradley P; Hollingsworth, Stephanie N; Hohmann, Matthew G; Hoffmann, William A
2014-04-01
In resprouting species, fire-induced topkill causes a reduction in height and leaf area without a comparable reduction in the size of the root system, which should lead to an increase in the efficiency of water transport after fire. However, large plants undergo a greater relative reduction in size, compared with small plants, so we hypothesized that this enhancement in hydraulic efficiency would be greatest among large growth forms. In the ecotone between long-leaf pine (Pinus palustris Mill.) savannas and wetlands, we measured stomatal conductance (gs), mid-day leaf water potential (Ψleaf), leaf-specific whole-plant hydraulic conductance (KL.p), leaf area and height of 10 species covering a range of growth forms in burned and unburned sites. As predicted, KL.p was higher in post-fire resprouts than in unburned plants, and the post-fire increase in KL.p was positively related to plant size. Specifically, large-statured species tended to undergo the greatest relative reductions in leaf area and height, and correspondingly experienced the greatest increases in KL.p. The post-fire increase in KL.p was smaller than expected, however, due to a decrease in absolute root hydraulic conductance (i.e., not scaled to leaf area). The higher KL.p in burned sites was manifested as an increase in gs rather than an increase in Ψleaf. Post-fire increases in gs should promote high rates of photosynthesis for recovery of carbohydrate reserves and aboveground biomass, which is particularly important for large-statured species that require more time to recover their pre-fire size.
NASA Technical Reports Server (NTRS)
Navon, I. M.
1984-01-01
A Lagrange multiplier method using techniques developed by Bertsekas (1982) was applied to solving the problem of enforcing simultaneous conservation of the nonlinear integral invariants of the shallow water equations on a limited area domain. This application of nonlinear constrained optimization is of the large dimensional type and the conjugate gradient method was found to be the only computationally viable method for the unconstrained minimization. Several conjugate-gradient codes were tested and compared for increasing accuracy requirements. Robustness and computational efficiency were the principal criteria.
Financial incentives for reducing proliferation risks
DOE Office of Scientific and Technical Information (OSTI.GOV)
Weise, Rachel A.; Hund, Gretchen
This article submitted for publication to the Bulletin of Atomic Scientists explains the possible financial incentives for financial institutions and large integrators to reduce nuclear proliferation risks by including anti-proliferation measures in their due diligence and requiring their suppliers to meet heightened compliance standards. Because manufacturers of dual-use nuclear goods are diverse and numerous outreach is difficult. However, financial institutions and large integrators work with nearly all dual-use manufacturers, making financial institutions and integrators well-positioned to increase awareness of proliferation and trafficking risks throughout the nuclear supply chain
Investigation of orbitofrontal sulcogyral pattern in chronic schizophrenia.
Cropley, Vanessa L; Bartholomeusz, Cali F; Wu, Peter; Wood, Stephen J; Proffitt, Tina; Brewer, Warrick J; Desmond, Patricia M; Velakoulis, Dennis; Pantelis, Christos
2015-11-30
Abnormalities of orbitofrontal cortex (OFC) pattern type distribution have been associated with schizophrenia-spectrum disorders. We investigated OFC pattern type in a large sample of chronic schizophrenia patients and healthy controls. We found an increased frequency of Type II but no difference in Type I or III folding pattern in the schizophrenia group in comparison to controls. Further large studies are required to investigate the diagnostic specificity of altered OFC pattern type and to confirm the distribution of pattern type in the normal population. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
Flaxman, Abraham D; Stewart, Andrea; Joseph, Jonathan C; Alam, Nurul; Alam, Sayed Saidul; Chowdhury, Hafizur; Mooney, Meghan D; Rampatige, Rasika; Remolador, Hazel; Sanvictores, Diozele; Serina, Peter T; Streatfield, Peter Kim; Tallo, Veronica; Murray, Christopher J L; Hernandez, Bernardo; Lopez, Alan D; Riley, Ian Douglas
2018-02-01
There is increasing interest in using verbal autopsy to produce nationally representative population-level estimates of causes of death. However, the burden of processing a large quantity of surveys collected with paper and pencil has been a barrier to scaling up verbal autopsy surveillance. Direct electronic data capture has been used in other large-scale surveys and can be used in verbal autopsy as well, to reduce time and cost of going from collected data to actionable information. We collected verbal autopsy interviews using paper and pencil and using electronic tablets at two sites, and measured the cost and time required to process the surveys for analysis. From these cost and time data, we extrapolated costs associated with conducting large-scale surveillance with verbal autopsy. We found that the median time between data collection and data entry for surveys collected on paper and pencil was approximately 3 months. For surveys collected on electronic tablets, this was less than 2 days. For small-scale surveys, we found that the upfront costs of purchasing electronic tablets was the primary cost and resulted in a higher total cost. For large-scale surveys, the costs associated with data entry exceeded the cost of the tablets, so electronic data capture provides both a quicker and cheaper method of data collection. As countries increase verbal autopsy surveillance, it is important to consider the best way to design sustainable systems for data collection. Electronic data capture has the potential to greatly reduce the time and costs associated with data collection. For long-term, large-scale surveillance required by national vital statistical systems, electronic data capture reduces costs and allows data to be available sooner.
LARGE-SCALE HYDROGEN PRODUCTION FROM NUCLEAR ENERGY USING HIGH TEMPERATURE ELECTROLYSIS
DOE Office of Scientific and Technical Information (OSTI.GOV)
James E. O'Brien
2010-08-01
Hydrogen can be produced from water splitting with relatively high efficiency using high-temperature electrolysis. This technology makes use of solid-oxide cells, running in the electrolysis mode to produce hydrogen from steam, while consuming electricity and high-temperature process heat. When coupled to an advanced high temperature nuclear reactor, the overall thermal-to-hydrogen efficiency for high-temperature electrolysis can be as high as 50%, which is about double the overall efficiency of conventional low-temperature electrolysis. Current large-scale hydrogen production is based almost exclusively on steam reforming of methane, a method that consumes a precious fossil fuel while emitting carbon dioxide to the atmosphere. Demandmore » for hydrogen is increasing rapidly for refining of increasingly low-grade petroleum resources, such as the Athabasca oil sands and for ammonia-based fertilizer production. Large quantities of hydrogen are also required for carbon-efficient conversion of biomass to liquid fuels. With supplemental nuclear hydrogen, almost all of the carbon in the biomass can be converted to liquid fuels in a nearly carbon-neutral fashion. Ultimately, hydrogen may be employed as a direct transportation fuel in a “hydrogen economy.” The large quantity of hydrogen that would be required for this concept should be produced without consuming fossil fuels or emitting greenhouse gases. An overview of the high-temperature electrolysis technology will be presented, including basic theory, modeling, and experimental activities. Modeling activities include both computational fluid dynamics and large-scale systems analysis. We have also demonstrated high-temperature electrolysis in our laboratory at the 15 kW scale, achieving a hydrogen production rate in excess of 5500 L/hr.« less
Upgrade plans for the ATLAS Forward Calorimeter at the HL-LHC
NASA Astrophysics Data System (ADS)
Rutherfoord, John; ATLAS Liquid Argon Calorimeter Group
2012-12-01
Although data-taking at CERN's Large Hadron Collider (LHC) is expected to continue for a number of years, plans are already being developed for operation of the LHC and associated detectors at an increased instantaneous luminosity about 5 times the original design value of 1034 cm-2 s-1. The increased particle flux at this high luminosity (HL) will have an impact on many sub-systems of the ATLAS detector. In particular, in the liquid argon forward calorimeter (FCal), which was designed for operation at LHC luminosities, the associated increase in the ionization load at HL-LHC luminosities creates a number of problems which can degrade its performance. These include space-charge effects in the liquid argon gaps, excessive drop in potential across the gaps due to large HV supply currents through the protection resistors, and an increase in temperature which may cause the liquid argon to boil. One solution, which would require opening both End-Cap cryostats, is the construction and installation of new FCals with narrower liquid argon gaps, lowering the values of the protection resistors, and the addition of cooling loops. A second proposed solution, which does not require opening the cryostat cold volume, is the addition of a small, warm calorimeter in front of each existing FCal, resulting in a reduction of the particle flux to levels at which the existing FCal can operate normally.
Discontinuous Spectral Difference Method for Conservation Laws on Unstructured Grids
NASA Technical Reports Server (NTRS)
Liu, Yen; Vinokur, Marcel
2004-01-01
A new, high-order, conservative, and efficient discontinuous spectral finite difference (SD) method for conservation laws on unstructured grids is developed. The concept of discontinuous and high-order local representations to achieve conservation and high accuracy is utilized in a manner similar to the Discontinuous Galerkin (DG) and the Spectral Volume (SV) methods, but while these methods are based on the integrated forms of the equations, the new method is based on the differential form to attain a simpler formulation and higher efficiency. Conventional unstructured finite-difference and finite-volume methods require data reconstruction based on the least-squares formulation using neighboring point or cell data. Since each unknown employs a different stencil, one must repeat the least-squares inversion for every point or cell at each time step, or to store the inversion coefficients. In a high-order, three-dimensional computation, the former would involve impractically large CPU time, while for the latter the memory requirement becomes prohibitive. In addition, the finite-difference method does not satisfy the integral conservation in general. By contrast, the DG and SV methods employ a local, universal reconstruction of a given order of accuracy in each cell in terms of internally defined conservative unknowns. Since the solution is discontinuous across cell boundaries, a Riemann solver is necessary to evaluate boundary flux terms and maintain conservation. In the DG method, a Galerkin finite-element method is employed to update the nodal unknowns within each cell. This requires the inversion of a mass matrix, and the use of quadratures of twice the order of accuracy of the reconstruction to evaluate the surface integrals and additional volume integrals for nonlinear flux functions. In the SV method, the integral conservation law is used to update volume averages over subcells defined by a geometrically similar partition of each grid cell. As the order of accuracy increases, the partitioning for 3D requires the introduction of a large number of parameters, whose optimization to achieve convergence becomes increasingly more difficult. Also, the number of interior facets required to subdivide non-planar faces, and the additional increase in the number of quadrature points for each facet, increases the computational cost greatly.
Design of a large span-distributed load flying-wing cargo airplane with laminar flow control
NASA Technical Reports Server (NTRS)
Lovell, W. A.; Price, J. E.; Quartero, C. B.; Turriziani, R. V.; Washburn, G. F.
1978-01-01
A design study was conducted to add laminar flow control to a previously design span-distributed load airplane while maintaining constant range and payload. With laminar flow control applied to 100 percent of the wing and vertical tail chords, the empty weight increased by 4.2 percent, the drag decreased by 27.4 percent, the required engine thrust decreased by 14.8 percent, and the fuel consumption decreased by 21.8 percent. When laminar flow control was applied to a lesser extent of the chord (approximately 80 percent), the empty weight increased by 3.4 percent, the drag decreased by 20.0 percent, the required engine thrust decreased by 13.0 percent, and the fuel consumption decreased by 16.2 percent. In both cases the required take-off gross weight of the aircraft was less than the original turbulent aircraft.
Columbus, Daniel A; Lapierre, Hélène; Htoo, John K; de Lange, Cornelis F M
2014-05-01
Nitrogen absorption from the large intestine, largely as ammonia and possibly as amino acids (AAs), is generally thought to be of little nutritional value to nonruminant animals and humans. Ammonia-nitrogen absorbed from the large intestine, however, may be recycled into the small intestine as urea and incorporated into microbial AAs, which may then be used by the host. A cecal infusion study was performed to determine the form in which nitrogen is absorbed from the large intestine and the impact of large intestine nitrogen supply on nitrogen balance in growing pigs. Eighteen cecally cannulated barrows (initial body weight: 22.4 ± 1.2 kg) were used to determine the effect of supplying nitrogen into the large intestine from either casein or urea on whole-body nitrogen retention and urea kinetics. Treatments were cecal infusions of saline (control), casein, or urea with nitrogen infused at a rate of 40% of nitrogen intake. In a subsample of 9 pigs, (15)N(15)N-urea was infused via i.v. during the nitrogen-balance period to determine urea kinetics. All pigs were fed a valine-limiting cornstarch-soybean meal-based diet. More than 80% of infused nitrogen was apparently absorbed. Urea flux and urinary nitrogen excretion increased (P ≤ 0.05) by the same amount for both nitrogen sources, but this increase did not fully account for the increase in nitrogen absorption from the large intestine. Whole-body nitrogen retention improved with nitrogen infusions (129 vs. 114 g/d; P < 0.01) and did not differ (P > 0.05) between nitrogen sources. Absorption of nitrogen from the large intestine appears to be in the form of nonprotein nitrogen, which appears to be returned to the small intestine via urea and used there for microbial AA production and should therefore be considered when determining nitrogen and AA supply and requirements.
Parallel Clustering Algorithm for Large-Scale Biological Data Sets
Wang, Minchao; Zhang, Wu; Ding, Wang; Dai, Dongbo; Zhang, Huiran; Xie, Hao; Chen, Luonan; Guo, Yike; Xie, Jiang
2014-01-01
Backgrounds Recent explosion of biological data brings a great challenge for the traditional clustering algorithms. With increasing scale of data sets, much larger memory and longer runtime are required for the cluster identification problems. The affinity propagation algorithm outperforms many other classical clustering algorithms and is widely applied into the biological researches. However, the time and space complexity become a great bottleneck when handling the large-scale data sets. Moreover, the similarity matrix, whose constructing procedure takes long runtime, is required before running the affinity propagation algorithm, since the algorithm clusters data sets based on the similarities between data pairs. Methods Two types of parallel architectures are proposed in this paper to accelerate the similarity matrix constructing procedure and the affinity propagation algorithm. The memory-shared architecture is used to construct the similarity matrix, and the distributed system is taken for the affinity propagation algorithm, because of its large memory size and great computing capacity. An appropriate way of data partition and reduction is designed in our method, in order to minimize the global communication cost among processes. Result A speedup of 100 is gained with 128 cores. The runtime is reduced from serval hours to a few seconds, which indicates that parallel algorithm is capable of handling large-scale data sets effectively. The parallel affinity propagation also achieves a good performance when clustering large-scale gene data (microarray) and detecting families in large protein superfamilies. PMID:24705246
Rodger, Marc; Langlois, Nicole; Middeldorp, Saskia; Kahn, Susan; Sandset, Per Morten; Brighton, Timothy; Huisman, Menno V; Meyer, Guy; Konstantinides, Stavros; Ageno, Walter; Morange, Pierre; Garcia, David; Kreuziger, Lisa Baumann; Young, Laura; Key, Nigel; Monreal, Manuel; Jiménez, David
2018-03-01
Venous thromboembolism (VTE) represents a major global burden of disease and requires collaborative efforts to conduct large, high-quality investigator-initiated and academically sponsored studies addressing the most relevant clinical questions. Owing to increasing regulatory requirements, the highly competitive nature of peer-reviewed funding and costs associated with conducting large, multinational clinical trials, completing practice-changing research constitutes a growing challenge for clinical investigators. As clinical trialists interested in VTE, we founded INVENT (International Network of Venous Thromboembolism Clinical Research Networks) in an effort to promote and accelerate patient-oriented, investigator-initiated, international collaborative research, to identify, prioritize and answer key clinical research questions for patients with VTE. We report on our activities to formalize the INVENT network and our accomplishments in our first year. Copyright © 2018 Elsevier Ltd. All rights reserved.
Controls, health assessment, and conditional monitoring for large, reusable, liquid rocket engines
NASA Technical Reports Server (NTRS)
Cikanek, H. A., III
1986-01-01
Past and future progress in the performance of control systems for large, liquid rocket engines typified such as current state-of-the-art, the Shuttle Main Engine (SSME), is discussed. Details of the first decade of efforts, which culminates in the F-1 and J-2 Saturn engines control systems, are traced, noting problem modes and improvements which were implemented to realize the SSME. Future control system designs, to accommodate the requirements of operation of engines for a heavy lift launch vehicle, an orbital transfer vehicle and the aerospace plane, are summarized. Generic design upgrades needed include an expanded range of fault detection, maintenance as-needed instead of as-scheduled, reduced human involvement in engine operations, and increased control of internal engine states. Current NASA technology development programs aimed at meeting the future control system requirements are described.
A Primer on Infectious Disease Bacterial Genomics
Petkau, Aaron; Knox, Natalie; Graham, Morag; Van Domselaar, Gary
2016-01-01
SUMMARY The number of large-scale genomics projects is increasing due to the availability of affordable high-throughput sequencing (HTS) technologies. The use of HTS for bacterial infectious disease research is attractive because one whole-genome sequencing (WGS) run can replace multiple assays for bacterial typing, molecular epidemiology investigations, and more in-depth pathogenomic studies. The computational resources and bioinformatics expertise required to accommodate and analyze the large amounts of data pose new challenges for researchers embarking on genomics projects for the first time. Here, we present a comprehensive overview of a bacterial genomics projects from beginning to end, with a particular focus on the planning and computational requirements for HTS data, and provide a general understanding of the analytical concepts to develop a workflow that will meet the objectives and goals of HTS projects. PMID:28590251
Linear Approximation SAR Azimuth Processing Study
NASA Technical Reports Server (NTRS)
Lindquist, R. B.; Masnaghetti, R. K.; Belland, E.; Hance, H. V.; Weis, W. G.
1979-01-01
A segmented linear approximation of the quadratic phase function that is used to focus the synthetic antenna of a SAR was studied. Ideal focusing, using a quadratic varying phase focusing function during the time radar target histories are gathered, requires a large number of complex multiplications. These can be largely eliminated by using linear approximation techniques. The result is a reduced processor size and chip count relative to ideally focussed processing and a correspondingly increased feasibility for spaceworthy implementation. A preliminary design and sizing for a spaceworthy linear approximation SAR azimuth processor meeting requirements similar to those of the SEASAT-A SAR was developed. The study resulted in a design with approximately 1500 IC's, 1.2 cubic feet of volume, and 350 watts of power for a single look, 4000 range cell azimuth processor with 25 meters resolution.
Remote visualization and scale analysis of large turbulence datatsets
NASA Astrophysics Data System (ADS)
Livescu, D.; Pulido, J.; Burns, R.; Canada, C.; Ahrens, J.; Hamann, B.
2015-12-01
Accurate simulations of turbulent flows require solving all the dynamically relevant scales of motions. This technique, called Direct Numerical Simulation, has been successfully applied to a variety of simple flows; however, the large-scale flows encountered in Geophysical Fluid Dynamics (GFD) would require meshes outside the range of the most powerful supercomputers for the foreseeable future. Nevertheless, the current generation of petascale computers has enabled unprecedented simulations of many types of turbulent flows which focus on various GFD aspects, from the idealized configurations extensively studied in the past to more complex flows closer to the practical applications. The pace at which such simulations are performed only continues to increase; however, the simulations themselves are restricted to a small number of groups with access to large computational platforms. Yet the petabytes of turbulence data offer almost limitless information on many different aspects of the flow, from the hierarchy of turbulence moments, spectra and correlations, to structure-functions, geometrical properties, etc. The ability to share such datasets with other groups can significantly reduce the time to analyze the data, help the creative process and increase the pace of discovery. Using the largest DOE supercomputing platforms, we have performed some of the biggest turbulence simulations to date, in various configurations, addressing specific aspects of turbulence production and mixing mechanisms. Until recently, the visualization and analysis of such datasets was restricted by access to large supercomputers. The public Johns Hopkins Turbulence database simplifies the access to multi-Terabyte turbulence datasets and facilitates turbulence analysis through the use of commodity hardware. First, one of our datasets, which is part of the database, will be described and then a framework that adds high-speed visualization and wavelet support for multi-resolution analysis of turbulence will be highlighted. The addition of wavelet support reduces the latency and bandwidth requirements for visualization, allowing for many concurrent users, and enables new types of analyses, including scale decomposition and coherent feature extraction.
Space Fed Subarray Synthesis Using Displaced Feed Location
NASA Astrophysics Data System (ADS)
Mailloux, Robert J.
2002-01-01
Wideband space-fed subarray systems are often proposed for large airborne or spaceborne scanning array applications. These systems allow the introduction of time delay devices at the subarray input terminals while using phase shifters in the array face. This can sometimes reduce the number of time delayed controls by an order of magnitude or more. The implementation of this technology has been slowed because the feed network, usually a Rotman Lens or Butler Matrix, is bulky, heavy and often has significant RF loss. In addition, the large lens aperture is necessarily filled with phase shifters, and so it introduces further loss, weight, and perhaps unacceptable phase shifter control power. These systems are currently viewed with increased interest because combination of low loss, low power MEMS phase shifters in the main aperture and solid state T/R modules in the feed might lead to large scanning arrays with much higher efficiency then previously realizable. Unfortunately, the conventional system design imposes an extremely large dynamic range requirement when used in the transmit mode, and requires very high output power from the T/R modules. This paper presents one possible solution to this problem using a modified feed geometry.
A global probabilistic tsunami hazard assessment from earthquake sources
Davies, Gareth; Griffin, Jonathan; Lovholt, Finn; Glimsdal, Sylfest; Harbitz, Carl; Thio, Hong Kie; Lorito, Stefano; Basili, Roberto; Selva, Jacopo; Geist, Eric L.; Baptista, Maria Ana
2017-01-01
Large tsunamis occur infrequently but have the capacity to cause enormous numbers of casualties, damage to the built environment and critical infrastructure, and economic losses. A sound understanding of tsunami hazard is required to underpin management of these risks, and while tsunami hazard assessments are typically conducted at regional or local scales, globally consistent assessments are required to support international disaster risk reduction efforts, and can serve as a reference for local and regional studies. This study presents a global-scale probabilistic tsunami hazard assessment (PTHA), extending previous global-scale assessments based largely on scenario analysis. Only earthquake sources are considered, as they represent about 80% of the recorded damaging tsunami events. Globally extensive estimates of tsunami run-up height are derived at various exceedance rates, and the associated uncertainties are quantified. Epistemic uncertainties in the exceedance rates of large earthquakes often lead to large uncertainties in tsunami run-up. Deviations between modelled tsunami run-up and event observations are quantified, and found to be larger than suggested in previous studies. Accounting for these deviations in PTHA is important, as it leads to a pronounced increase in predicted tsunami run-up for a given exceedance rate.
A new way to protect privacy in large-scale genome-wide association studies.
Kamm, Liina; Bogdanov, Dan; Laur, Sven; Vilo, Jaak
2013-04-01
Increased availability of various genotyping techniques has initiated a race for finding genetic markers that can be used in diagnostics and personalized medicine. Although many genetic risk factors are known, key causes of common diseases with complex heritage patterns are still unknown. Identification of such complex traits requires a targeted study over a large collection of data. Ideally, such studies bring together data from many biobanks. However, data aggregation on such a large scale raises many privacy issues. We show how to conduct such studies without violating privacy of individual donors and without leaking the data to third parties. The presented solution has provable security guarantees. Supplementary data are available at Bioinformatics online.
Pre- and postnatal health: evidence of increased choline needs.
Caudill, Marie A
2010-08-01
Choline, a micronutrient found in food, serves as the starting material for several important metabolites that play key roles in fetal development, particularly the brain. Although human beings' requirement for choline is unknown, an Adequate Intake level of 425 mg/day was established for women with upward adjustments to 450 and 550 mg/day during pregnancy and lactation, respectively. The importance of choline in human development is supported by observations that a human fetus receives a large supply of choline during gestation; pregnancy causes depletion of hepatic choline pools in rats consuming a normal diet; human neonates are born with blood levels that are three times higher than maternal blood concentrations; and large amounts of choline are present in human milk. The development of the central nervous system is particularly sensitive to choline availability with evidence of effects on neural tube closure and cognition. Existing data show that the majority of pregnant (and presumably lactating) women are not achieving the target intake levels and that certain common genetic variants may increase requirements for choline beyond current recommendations. Because choline is not found in most varieties of prenatal vitamins (or regular multivitamins), increased consumption of choline-rich foods may be needed to meet the high pre- and postnatal demands for choline. 2010 American Dietetic Association. Published by Elsevier Inc. All rights reserved.
Thermographic Imaging of Defects in Anisotropic Composites
NASA Technical Reports Server (NTRS)
Plotnikov, Y. A.; Winfree, W. P.
2000-01-01
Composite materials are of increasing interest to the aerospace industry as a result of their weight versus performance characteristics. One of the disadvantages of composites is the high cost of fabrication and post inspection with conventional ultrasonic scanning systems. The high cost of inspection is driven by the need for scanning systems which can follow large curve surfaces. Additionally, either large water tanks or water squirters are required to couple the ultrasonics into the part. Thermographic techniques offer significant advantages over conventional ultrasonics by not requiring physical coupling between the part and sensor. The thermographic system can easily inspect large curved surface without requiring a surface following scanner. However, implementation of Thermal Nondestructive Evaluations (TNDE) for flaw detection in composite materials and structures requires determining its limit. Advanced algorithms have been developed to enable locating and sizing defects in carbon fiber reinforced plastic (CFRP). Thermal Tomography is a very promising method for visualizing the size and location of defects in materials such as CFRP. However, further investigations are required to determine its capabilities for inspection of thick composites. In present work we have studied influence of the anisotropy on the reconstructed image of a defect generated by an inversion technique. The composite material is considered as homogeneous with macro properties: thermal conductivity K, specific heat c, and density rho. The simulation process involves two sequential steps: solving the three dimensional transient heat diffusion equation for a sample with a defect, then estimating the defect location and size from the surface spatial and temporal thermal distributions (inverse problem), calculated from the simulations.
Using Unplanned Fires to Help Suppressing Future Large Fires in Mediterranean Forests
Regos, Adrián; Aquilué, Núria; Retana, Javier; De Cáceres, Miquel; Brotons, Lluís
2014-01-01
Despite the huge resources invested in fire suppression, the impact of wildfires has considerably increased across the Mediterranean region since the second half of the 20th century. Modulating fire suppression efforts in mild weather conditions is an appealing but hotly-debated strategy to use unplanned fires and associated fuel reduction to create opportunities for suppression of large fires in future adverse weather conditions. Using a spatially-explicit fire–succession model developed for Catalonia (Spain), we assessed this opportunistic policy by using two fire suppression strategies that reproduce how firefighters in extreme weather conditions exploit previous fire scars as firefighting opportunities. We designed scenarios by combining different levels of fire suppression efficiency and climatic severity for a 50-year period (2000–2050). An opportunistic fire suppression policy induced large-scale changes in fire regimes and decreased the area burnt under extreme climate conditions, but only accounted for up to 18–22% of the area to be burnt in reference scenarios. The area suppressed in adverse years tended to increase in scenarios with increasing amounts of area burnt during years dominated by mild weather. Climate change had counterintuitive effects on opportunistic fire suppression strategies. Climate warming increased the incidence of large fires under uncontrolled conditions but also indirectly increased opportunities for enhanced fire suppression. Therefore, to shift fire suppression opportunities from adverse to mild years, we would require a disproportionately large amount of area burnt in mild years. We conclude that the strategic planning of fire suppression resources has the potential to become an important cost-effective fuel-reduction strategy at large spatial scale. We do however suggest that this strategy should probably be accompanied by other fuel-reduction treatments applied at broad scales if large-scale changes in fire regimes are to be achieved, especially in the wider context of climate change. PMID:24727853
Using unplanned fires to help suppressing future large fires in Mediterranean forests.
Regos, Adrián; Aquilué, Núria; Retana, Javier; De Cáceres, Miquel; Brotons, Lluís
2014-01-01
Despite the huge resources invested in fire suppression, the impact of wildfires has considerably increased across the Mediterranean region since the second half of the 20th century. Modulating fire suppression efforts in mild weather conditions is an appealing but hotly-debated strategy to use unplanned fires and associated fuel reduction to create opportunities for suppression of large fires in future adverse weather conditions. Using a spatially-explicit fire-succession model developed for Catalonia (Spain), we assessed this opportunistic policy by using two fire suppression strategies that reproduce how firefighters in extreme weather conditions exploit previous fire scars as firefighting opportunities. We designed scenarios by combining different levels of fire suppression efficiency and climatic severity for a 50-year period (2000-2050). An opportunistic fire suppression policy induced large-scale changes in fire regimes and decreased the area burnt under extreme climate conditions, but only accounted for up to 18-22% of the area to be burnt in reference scenarios. The area suppressed in adverse years tended to increase in scenarios with increasing amounts of area burnt during years dominated by mild weather. Climate change had counterintuitive effects on opportunistic fire suppression strategies. Climate warming increased the incidence of large fires under uncontrolled conditions but also indirectly increased opportunities for enhanced fire suppression. Therefore, to shift fire suppression opportunities from adverse to mild years, we would require a disproportionately large amount of area burnt in mild years. We conclude that the strategic planning of fire suppression resources has the potential to become an important cost-effective fuel-reduction strategy at large spatial scale. We do however suggest that this strategy should probably be accompanied by other fuel-reduction treatments applied at broad scales if large-scale changes in fire regimes are to be achieved, especially in the wider context of climate change.
Low-Gain Circularly Polarized Antenna with Torus-Shaped Pattern
NASA Technical Reports Server (NTRS)
Amaro, Luis R.; Kruid, Ronald C.; Vacchione, Joseph D.; Prata, Aluizio
2012-01-01
The Juno mission to Jupiter requires an antenna with a torus-shaped antenna pattern with approximately 6 dBic gain and circular polarization over the Deep Space Network (DSN) 7-GHz transmit frequency and the 8-GHz receive frequency. Given the large distances that accumulate en-route to Jupiter and the limited power afforded by the solar-powered vehicle, this toroidal low-gain antenna requires as much gain as possible while maintaining a beam width that could facilitate a +/-10deg edge of coverage. The natural antenna that produces a toroidal antenna pattern is the dipole, but the limited approx. = 2.2 dB peak gain would be insufficient. Here a shaped variation of the standard bicone antenna is proposed that could achieve the required gains and bandwidths while maintaining a size that was not excessive. The final geometry that was settled on consisted of a corrugated, shaped bicone, which is fed by a WR112 waveguide-to-coaxial- waveguide transition. This toroidal low-gain antenna (TLGA) geometry produced the requisite gain, moderate sidelobes, and the torus-shaped antenna pattern while maintaining a very good match over the entire required frequency range. Its "horn" geometry is also low-loss and capable of handling higher powers with large margins against multipactor breakdown. The final requirement for the antenna was to link with the DSN with circular polarization. A four-layer meander-line array polarizer was implemented; an approach that was fairly well suited to the TLGA geometry. The principal development of this work was to adapt the standard linear bicone such that its aperture could be increased in order to increase the available gain of the antenna. As one increases the aperture of a standard bicone, the phase variation across the aperture begins to increase, so the larger the aperture becomes, the greater the phase variation. In order to maximize the gain from any aperture antenna, the phase should be kept as uniform as possible. Thus, as the standard bicone fs aperture increases, the gain increase becomes less until one reaches a point of diminishing returns. In order to overcome this problem, a shaped aperture is used. Rather than the standard linear bicone, a parabolic bicone was found to reduce the amount of phase variation as the aperture increases. In fact, the phase variation is half of the standard linear bicone, which leads to higher gain with smaller aperture sizes. The antenna pattern radiated from this parabolic-shaped bicone antenna has fairly high side lobes. The Juno project requested that these sidelobes be minimized. This was accomplished by adding corrugations to the parabolic shape. This corrugated-shaped bicone antenna had reasonably low sidelobes, and the appropriate gain and beamwidth to meet project requirements.
Increasing older persons' employment in Finland: in search of a new strategy.
Sihto, M
1999-01-01
Using Finland as a case study, it is argued that early retirement will probably no longer be used on a large scale to reduce older-worker labor-force participation and unemployment among older workers. Instead, new strategies are needed to enhance the ability of older workers to remain productive and in the labor force, and to facilitate the reintegration of unemployed older persons back into working life. Both tasks require massive pioneering efforts. Reducing unemployment rates among older workers, particularly, requires completely new kinds of labor-market measures.
Localization Algorithm Based on a Spring Model (LASM) for Large Scale Wireless Sensor Networks.
Chen, Wanming; Mei, Tao; Meng, Max Q-H; Liang, Huawei; Liu, Yumei; Li, Yangming; Li, Shuai
2008-03-15
A navigation method for a lunar rover based on large scale wireless sensornetworks is proposed. To obtain high navigation accuracy and large exploration area, highnode localization accuracy and large network scale are required. However, thecomputational and communication complexity and time consumption are greatly increasedwith the increase of the network scales. A localization algorithm based on a spring model(LASM) method is proposed to reduce the computational complexity, while maintainingthe localization accuracy in large scale sensor networks. The algorithm simulates thedynamics of physical spring system to estimate the positions of nodes. The sensor nodesare set as particles with masses and connected with neighbor nodes by virtual springs. Thevirtual springs will force the particles move to the original positions, the node positionscorrespondingly, from the randomly set positions. Therefore, a blind node position can bedetermined from the LASM algorithm by calculating the related forces with the neighbornodes. The computational and communication complexity are O(1) for each node, since thenumber of the neighbor nodes does not increase proportionally with the network scale size.Three patches are proposed to avoid local optimization, kick out bad nodes and deal withnode variation. Simulation results show that the computational and communicationcomplexity are almost constant despite of the increase of the network scale size. The time consumption has also been proven to remain almost constant since the calculation steps arealmost unrelated with the network scale size.
Water and Land Limitations to Future Agricultural Production in the Middle East
NASA Astrophysics Data System (ADS)
Koch, J. A. M.; Wimmer, F.; Schaldach, R.
2015-12-01
Countries in the Middle East use a large fraction of their scarce water resources to produce cash crops, such as fruit and vegetables, for international markets. At the same time, these countries import large amounts of staple crops, such as cereals, required to meet the nutritional demand of their populations. This makes food security in the Middle East heavily dependent on world market prices for staple crops. Under these preconditions, increasing food demand due to population growth, urban expansion on fertile farmlands, and detrimental effects of a changing climate on the production of agricultural commodities present major challenges to countries in the Middle East that try to improve food security by increasing their self-sufficiency rate of staple crops.We applied the spatio-temporal land-use change model LandSHIFT.JR to simulate how an expansion of urban areas may affect the production of agricultural commodities in Jordan. We furthermore evaluated how climate change and changes in socio-economic conditions may influence crop production. The focus of our analysis was on potential future irrigated and rainfed production (crop yield and area demand) of fruit, vegetables, and cereals. Our simulation results show that the expansion of urban areas and the resulting displacement of agricultural areas does result in a slight decrease in crop yields. This leads to almost no additional irrigation water requirements due to the relocation of agricultural areas, i.e. there is the same amount of "crop per drop". However, taking into account projected changes in socio-economic conditions and climate conditions, a large volume of water would be required for cereal production in order to safeguard current self-sufficiency rates for staple crops. Irrigation water requirements are expected to double until 2025 and to triple until 2050. Irrigated crop yields are projected to decrease by about 25%, whereas there is no decrease in rainfed crop yields to be expected.
Tanger, Paul; Klassen, Stephen; Mojica, Julius P.; Lovell, John T.; Moyers, Brook T.; Baraoidan, Marietta; Naredo, Maria Elizabeth B.; McNally, Kenneth L.; Poland, Jesse; Bush, Daniel R.; Leung, Hei; Leach, Jan E.; McKay, John K.
2017-01-01
To ensure food security in the face of population growth, decreasing water and land for agriculture, and increasing climate variability, crop yields must increase faster than the current rates. Increased yields will require implementing novel approaches in genetic discovery and breeding. Here we demonstrate the potential of field-based high throughput phenotyping (HTP) on a large recombinant population of rice to identify genetic variation underlying important traits. We find that detecting quantitative trait loci (QTL) with HTP phenotyping is as accurate and effective as traditional labor-intensive measures of flowering time, height, biomass, grain yield, and harvest index. Genetic mapping in this population, derived from a cross of an modern cultivar (IR64) with a landrace (Aswina), identified four alleles with negative effect on grain yield that are fixed in IR64, demonstrating the potential for HTP of large populations as a strategy for the second green revolution. PMID:28220807
Transforming the NAS: The Next Generation Air Traffic Control System
NASA Technical Reports Server (NTRS)
Erzberger, Heinz
2004-01-01
The next-generation air traffic control system must be designed to safely and efficiently accommodate the large growth of traffic expected in the near future. It should be sufficiently scalable to contend with the factor of 2 or more increase in demand expected by the year 2020. Analysis has shown that the current method of controlling air traffic cannot be scaled up to provide such levels of capacity. Therefore, to achieve a large increase in capacity while also giving pilots increased freedom to optimize their flight trajectories requires a fundamental change in the way air traffic is controlled. The key to achieving a factor of 2 or more increase in airspace capacity is to automate separation monitoring and control and to use an air-ground data link to send trajectories and clearances directly between ground-based and airborne systems. In addition to increasing capacity and offering greater flexibility in the selection of trajectories, this approach also has the potential to increase safety by reducing controller and pilot errors that occur in routine monitoring and voice communication tasks.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wood, E.; Burton, E.; Duran, A.
Understanding the real-world power demand of modern automobiles is of critical importance to engineers using modeling and simulation to inform the intelligent design of increasingly efficient powertrains. Increased use of global positioning system (GPS) devices has made large scale data collection of vehicle speed (and associated power demand) a reality. While the availability of real-world GPS data has improved the industry's understanding of in-use vehicle power demand, relatively little attention has been paid to the incremental power requirements imposed by road grade. This analysis quantifies the incremental efficiency impacts of real-world road grade by appending high fidelity elevation profiles tomore » GPS speed traces and performing a large simulation study. Employing a large real-world dataset from the National Renewable Energy Laboratory's Transportation Secure Data Center, vehicle powertrain simulations are performed with and without road grade under five vehicle models. Aggregate results of this study suggest that road grade could be responsible for 1% to 3% of fuel use in light-duty automobiles.« less
Glahn, David; Nuccitelli, Richard
2003-04-01
Voltage-clamped mature, jelly-intact Xenopus eggs were used to carefully examine the ionic currents crossing the plasma membrane before, during, and after fertilization. The bulk of the fertilization current was transient, of large amplitude, and reversed at the predicted Cl- reversal potential. However, the large amplitude fertilization current was preceded by a small, step-like increase in holding current. This small increase in holding current is referred to in this paper as Ion to acknowledge its qualitative similarity to the Ion current previously described in the sea urchin. It was observed in both fertilized and artificially activated eggs, and was found to be unaffected by 10 mm tetra-ethyl ammonium (TEA), a concentration found to block K+ currents in Rana pipiens. Current-voltage relationships are presented for the large fertilization potential, and show that the fertilization currents have a marked outward rectification and are voltage sensitive. These properties are in contrast to the total lack of rectification and slight voltage sensitivity seen before or after the fertilization currents. The time required for sperm to fertilize the egg was found to be voltage dependent with a relatively more depolarized voltage requiring a longer time for fertilization to occur. The percentage of eggs blocked with varying potential levels was determined and this information was fitted to a modified Boltzmann equation having a midpoint of -9 mV.
Green and blue water demand from large-scale land acquisitions in Africa
Johansson, Emma Li; Fader, Marianela; Seaquist, Jonathan W.; Nicholas, Kimberly A.
2016-01-01
In the last decade, more than 22 million ha of land have been contracted to large-scale land acquisitions in Africa, leading to increased pressures, competition, and conflicts over freshwater resources. Currently, 3% of contracted land is in production, for which we model site-specific water demands to indicate where freshwater appropriation might pose high socioenvironmental challenges. We use the dynamic global vegetation model Lund–Potsdam–Jena managed Land to simulate green (precipitation stored in soils and consumed by plants through evapotranspiration) and blue (extracted from rivers, lakes, aquifers, and dams) water demand and crop yields for seven irrigation scenarios, and compare these data with two baseline scenarios of staple crops representing previous water demand. We find that most land acquisitions are planted with crops that demand large volumes of water (>9,000 m3⋅ha−1) like sugarcane, jatropha, and eucalyptus, and that staple crops have lower water requirements (<7,000 m3⋅ha−1). Blue water demand varies with irrigation system, crop choice, and climate. Even if the most efficient irrigation systems were implemented, 18% of the land acquisitions, totaling 91,000 ha, would still require more than 50% of water from blue water sources. These hotspots indicate areas at risk for transgressing regional constraints for freshwater use as a result of overconsumption of blue water, where socioenvironmental systems might face increased conflicts and tensions over water resources. PMID:27671634
Bionimbus: a cloud for managing, analyzing and sharing large genomics datasets.
Heath, Allison P; Greenway, Matthew; Powell, Raymond; Spring, Jonathan; Suarez, Rafael; Hanley, David; Bandlamudi, Chai; McNerney, Megan E; White, Kevin P; Grossman, Robert L
2014-01-01
As large genomics and phenotypic datasets are becoming more common, it is increasingly difficult for most researchers to access, manage, and analyze them. One possible approach is to provide the research community with several petabyte-scale cloud-based computing platforms containing these data, along with tools and resources to analyze it. Bionimbus is an open source cloud-computing platform that is based primarily upon OpenStack, which manages on-demand virtual machines that provide the required computational resources, and GlusterFS, which is a high-performance clustered file system. Bionimbus also includes Tukey, which is a portal, and associated middleware that provides a single entry point and a single sign on for the various Bionimbus resources; and Yates, which automates the installation, configuration, and maintenance of the software infrastructure required. Bionimbus is used by a variety of projects to process genomics and phenotypic data. For example, it is used by an acute myeloid leukemia resequencing project at the University of Chicago. The project requires several computational pipelines, including pipelines for quality control, alignment, variant calling, and annotation. For each sample, the alignment step requires eight CPUs for about 12 h. BAM file sizes ranged from 5 GB to 10 GB for each sample. Most members of the research community have difficulty downloading large genomics datasets and obtaining sufficient storage and computer resources to manage and analyze the data. Cloud computing platforms, such as Bionimbus, with data commons that contain large genomics datasets, are one choice for broadening access to research data in genomics. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://group.bmj.com/group/rights-licensing/permissions.
ERIC Educational Resources Information Center
Borsari, Brian; Hustad, John T. P.; Mastroleo, Nadine R.; Tevyaw, Tracy O'Leary; Barnett, Nancy P.; Kahler, Christopher W.; Short, Erica Eaton; Monti, Peter M.
2012-01-01
Objective: Over the past 2 decades, colleges and universities have seen a large increase in the number of students referred to the administration for alcohol policies violations. However, a substantial portion of mandated students may not require extensive treatment. Stepped care may maximize treatment efficiency and greatly reduce the demands on…
Curious or spurious correlations within a national-scale forest inventory?
Christopher W. Woodall; James A. Westfall
2012-01-01
Foresters are increasingly required to assess trends not only in traditional forest attributes (e.g., growing-stock volumes), but also across suites of forest health indicators and site/climate variables. Given the tenuous relationship between correlation and causality within extremely large datasets, the goal of this study was to use a nationwide annual forest...
Andrea Havron; Chris Goldfinger; Sarah Henkel; Bruce G. Marcot; Chris Romsos; Lisa Gilbane
2017-01-01
Resource managers increasingly use habitat suitability map products to inform risk management and policy decisions. Modeling habitat suitability of data-poor species over large areas requires careful attention to assumptions and limitations. Resulting habitat suitability maps can harbor uncertainties from data collection and modeling processes; yet these limitations...
ERIC Educational Resources Information Center
Lawrence, Anna; Turnhout, Esther
2010-01-01
The demand for biodiversity data is increasing. Governments require standardised, objective data to underpin planning and conservation decisions. These data are produced by large numbers of (volunteer) natural historians and non-governmental organisations. This article analyses the interface between the state and the volunteer naturalists to…
Using a Learning Management System to Enhance an Extensive Reading Program
ERIC Educational Resources Information Center
Koby, Cory J.
2017-01-01
The Extensive Reading (ER) approach to second language acquisition is increasingly one of the methods of choice amongst English as a Foreign Language (EFL) educators around the world. This method requires learners to read a large volume of easily comprehensible text, and teachers to track and manage their students' progress in some manner. There…
Building 4-H Program Capacity and Sustainability through Collaborative Fee-Based Programs
ERIC Educational Resources Information Center
Pellien, Tamara
2016-01-01
Shrinking budgets and increased demands for services and programs are the norm for today's Extension professional. The tasks of procuring grants, developing fund raisers, and pursuing donors require a large investment of time and can lead to mission drift in the pursuit of funding. Implementing a collaborative fee-based program initiative can fund…
Bringing the "Two Cultures" Together through "A World of Light and Color"
ERIC Educational Resources Information Center
Mian, Shabbir M.; Marx, Jeffrey D.; Pagonis, Vasilis
2008-01-01
In the United States, the undergraduate general education curriculum by and large requires students take courses from the arts and humanities as well as the sciences in order to produce well-rounded or liberally educated individuals. This educational philosophy is in line with C. P. Snow's recommendation for increased communication between the…
ERIC Educational Resources Information Center
Adams, Troy B.; Colner, Willa
2008-01-01
Few college students meet fruit and vegetable intake recommended requirements, and most receive no information from their institutions about this issue. The avoidable disease burden among students is large, the necessary information infrastructure exists, and "Healthy People 2010" objectives indicate efforts should be taken to increase intake.…
Modeling the cumulative watershed effects of forest management strategies
R. R. Ziemer; J. Lewis; R. M. Rice; T. E. Lisle
1991-01-01
Abstract - There is increasing concern over the possibility of adverse cumulative watershed effects from intensive forest management. It is impractical to address many aspects of the problem experimentally because to do so would require studying large watersheds for 100 yr or more. One such aspect is the long-term effect of forest management strategies on erosion and...
Karim, Md Rezaul; Michel, Audrey; Zappa, Achille; Baranov, Pavel; Sahay, Ratnesh; Rebholz-Schuhmann, Dietrich
2017-04-16
Data workflow systems (DWFSs) enable bioinformatics researchers to combine components for data access and data analytics, and to share the final data analytics approach with their collaborators. Increasingly, such systems have to cope with large-scale data, such as full genomes (about 200 GB each), public fact repositories (about 100 TB of data) and 3D imaging data at even larger scales. As moving the data becomes cumbersome, the DWFS needs to embed its processes into a cloud infrastructure, where the data are already hosted. As the standardized public data play an increasingly important role, the DWFS needs to comply with Semantic Web technologies. This advancement to DWFS would reduce overhead costs and accelerate the progress in bioinformatics research based on large-scale data and public resources, as researchers would require less specialized IT knowledge for the implementation. Furthermore, the high data growth rates in bioinformatics research drive the demand for parallel and distributed computing, which then imposes a need for scalability and high-throughput capabilities onto the DWFS. As a result, requirements for data sharing and access to public knowledge bases suggest that compliance of the DWFS with Semantic Web standards is necessary. In this article, we will analyze the existing DWFS with regard to their capabilities toward public open data use as well as large-scale computational and human interface requirements. We untangle the parameters for selecting a preferable solution for bioinformatics research with particular consideration to using cloud services and Semantic Web technologies. Our analysis leads to research guidelines and recommendations toward the development of future DWFS for the bioinformatics research community. © The Author 2017. Published by Oxford University Press.
Ayus, J C; Bellido, T; Negri, A L
2017-05-01
The Fracture Risk Assessment Tool (FRAX®) was developed by the WHO Collaborating Centre for metabolic bone diseases to evaluate fracture risk of patients. It is based on patient models that integrate the risk associated with clinical variables and bone mineral density (BMD) at the femoral neck. The clinical risk factors included in FRAX were chosen to include only well-established and independent variables related to skeletal fracture risk. The FRAX tool has acquired worldwide acceptance despite having several limitations. FRAX models have not included biochemical derangements in estimation of fracture risk due to the lack of validation in large prospective studies. Recently, there has been an increasing number of studies showing a relationship between hyponatremia and the occurrence of fractures. Hyponatremia is the most frequent electrolyte abnormality measured in the clinic, and serum sodium concentration is a very reproducible, affordable, and readily obtainable measurement. Thus, we think that hyponatremia should be further studied as a biochemical risk factor for skeletal fractures prediction, particularly those at the hip which carries the greatest morbidity and mortality. To achieve this will require the collection of large patient cohorts from diverse geographical locations that include a measure of serum sodium in addition to the other FRAX variables in large numbers, in both sexes, over a wide age range and with wide geographical representation. It would also require the inclusion of data on duration and severity of hyponatremia. Information will be required both on the risk of fracture associated with the occurrence and length of exposure to hyponatremia and to the relationship with the other risk variables included in FRAX and also the independent effect on the occurrence of death which is increased by hyponatremia.
Legacy Phosphorus Effect and Need to Re-calibrate Soil Test P Methods for Organic Crop Production.
NASA Astrophysics Data System (ADS)
Dao, Thanh H.; Schomberg, Harry H.; Cavigelli, Michel A.
2015-04-01
Phosphorus (P) is a required nutrient for the normal development and growth of plants and supplemental P is needed in most cultivated soils. Large inputs of cover crop residues and nutrient-rich animal manure are added to supply needed nutrients to promote the optimal production of organic grain crops and forages. The effects of crop rotations and tillage management of the near-surface zone on labile phosphorus (P) forms were studied in soil under conventional and organic crop management systems in the mid-Atlantic region of the U.S. after 18 years due to the increased interest in these alternative systems. Soil nutrient surpluses likely caused by low grain yields resulted in large pools of exchangeable phosphate-P and equally large pools of enzyme-labile organic P (Po) in soils under organic management. In addition, the difference in the P loading rates between the conventional and organic treatments as guided by routine soil test recommendations suggested that overestimating plant P requirements contributed to soil P surpluses because routine soil testing procedures did not account for the presence and size of the soil enzyme-labile Po pool. The effect of large P additions is long-lasting as they continued to contribute to elevated soil total bioactive P concentrations 12 or more years later. Consequently, accurate estimates of crop P requirements, P turnover in soil, and real-time plant and soil sensing systems are critical considerations to optimally manage manure-derived nutrients in organic crop production.
Luman, Marjolein; Sergeant, Joseph A; Knol, Dirk L; Oosterlaan, Jaap
2010-08-15
When making decisions, children with oppositional defiant disorder (ODD) are thought to focus on reward and ignore penalty. This is suggested to be associated with a state of low psychophysiological arousal. This study investigates decision making in 18 children with oppositional defiant disorder and 24 typically developing control subjects. Children were required to choose between three alternatives that carried either frequent small rewards and occasional small penalties (advantageous), frequent large rewards and increasing penalties (seductive), or frequent small rewards and increasing penalties (disadvantageous). Penalties in the seductive and disadvantageous alternatives increased either in frequency or magnitude in two conditions. Heart rate (HR) and skin conductance responses to reinforcement were obtained. In the magnitude condition, children with ODD showed an increased preference for the seductive alternative (carrying large rewards); this was not observed in the frequency condition. Children with ODD, compared with typically developing children, displayed greater HR reactivity to reward (more HR deceleration) and smaller HR reactivity to penalty. Correlation analyses showed that decreased HR responses to penalty were related to an increased preference for large rewards. No group differences were observed in skin conductance responses to reward or penalty. The findings suggest that an increased preference for large rewards in children with ODD is related to a reduced cardiac reactivity to aversive stimuli. This confirms notions of impaired decision making and altered reinforcement sensitivity in children with ODD and adds to the literature linking altered autonomic control to antisocial behavior. Copyright 2010 Society of Biological Psychiatry. Published by Elsevier Inc. All rights reserved.
NASA Technical Reports Server (NTRS)
1975-01-01
A set of energy conservation actions that cut across all sectors of the economy were analyzed so that all actions under consideration be analyzed systematically and as a whole. The actions considered were as follows: (1) roll back the price of newly discovered oil, (2) freeze gasoline production for 3 years at 1972 levels, (3) mandate automobile mileage improvements, (4) require industry to improve energy efficiency, (5) require manufacture of household appliances with greater efficiency, (6) force conversion of many power plants from gas and oil to coal. The results showed that considerable gas and oil would be saved by forcing switches to coal. However, the large scale switch to coal was shown to require greatly increased outputs from many other industries that in turn require more energy. It was estimated that nearly 2.5 quads of additional coal were needed to produce these additional requirements. Also, the indirect requirements would create more jobs.
NASA Astrophysics Data System (ADS)
Greiner, Nathan J.
Modern turbine engines require high turbine inlet temperatures and pressures to maximize thermal efficiency. Increasing the turbine inlet temperature drives higher heat loads on the turbine surfaces. In addition, increasing pressure ratio increases the turbine coolant temperature such that the ability to remove heat decreases. As a result, highly effective external film cooling is required to reduce the heat transfer to turbine surfaces. Testing of film cooling on engine hardware at engine temperatures and pressures can be exceedingly difficult and expensive. Thus, modern studies of film cooling are often performed at near ambient conditions. However, these studies are missing an important aspect in their characterization of film cooling effectiveness. Namely, they do not model effect of thermal property variations that occur within the boundary and film cooling layers at engine conditions. Also, turbine surfaces can experience significant radiative heat transfer that is not trivial to estimate analytically. The present research first computationally examines the effect of large temperature variations on a turbulent boundary layer. Subsequently, a method to model the effect of large temperature variations within a turbulent boundary layer in an environment coupled with significant radiative heat transfer is proposed and experimentally validated. Next, a method to scale turbine cooling from ambient to engine conditions via non-dimensional matching is developed computationally and the experimentally validated at combustion temperatures. Increasing engine efficiency and thrust to weight ratio demands have driven increased combustor fuel-air ratios. Increased fuel-air ratios increase the possibility of unburned fuel species entering the turbine. Alternatively, advanced ultra-compact combustor designs have been proposed to decrease combustor length, increase thrust, or generate power for directed energy weapons. However, the ultra-compact combustor design requires a film cooled vane within the combustor. In both these environments, the unburned fuel in the core flow encounters the oxidizer rich film cooling stream, combusts, and can locally heat the turbine surface rather than the intended cooling of the surface. Accordingly, a method to quantify film cooling performance in a fuel rich environment is prescribed. Finally, a method to film cool in a fuel rich environment is experimentally demonstrated.
Excess electron localization in solvated DNA bases.
Smyth, Maeve; Kohanoff, Jorge
2011-06-10
We present a first-principles molecular dynamics study of an excess electron in condensed phase models of solvated DNA bases. Calculations on increasingly large microsolvated clusters taken from liquid phase simulations show that adiabatic electron affinities increase systematically upon solvation, as for optimized gas-phase geometries. Dynamical simulations after vertical attachment indicate that the excess electron, which is initially found delocalized, localizes around the nucleobases within a 15 fs time scale. This transition requires small rearrangements in the geometry of the bases.
Excess Electron Localization in Solvated DNA Bases
DOE Office of Scientific and Technical Information (OSTI.GOV)
Smyth, Maeve; Kohanoff, Jorge
2011-06-10
We present a first-principles molecular dynamics study of an excess electron in condensed phase models of solvated DNA bases. Calculations on increasingly large microsolvated clusters taken from liquid phase simulations show that adiabatic electron affinities increase systematically upon solvation, as for optimized gas-phase geometries. Dynamical simulations after vertical attachment indicate that the excess electron, which is initially found delocalized, localizes around the nucleobases within a 15 fs time scale. This transition requires small rearrangements in the geometry of the bases.
Disaster and Contingency Planning for Scientific Shared Resource Cores.
Mische, Sheenah; Wilkerson, Amy
2016-04-01
Progress in biomedical research is largely driven by improvements, innovations, and breakthroughs in technology, accelerating the research process, and an increasingly complex collaboration of both clinical and basic science. This increasing sophistication has driven the need for centralized shared resource cores ("cores") to serve the scientific community. From a biomedical research enterprise perspective, centralized resource cores are essential to increased scientific, operational, and cost effectiveness; however, the concentration of instrumentation and resources in the cores may render them highly vulnerable to damage from severe weather and other disasters. As such, protection of these assets and the ability to recover from a disaster is increasingly critical to the mission and success of the institution. Therefore, cores should develop and implement both disaster and business continuity plans and be an integral part of the institution's overall plans. Here we provide an overview of key elements required for core disaster and business continuity plans, guidance, and tools for developing these plans, and real-life lessons learned at a large research institution in the aftermath of Superstorm Sandy.
Crystal Structure Variations of Sn Nanoparticles upon Heating
NASA Astrophysics Data System (ADS)
Mittal, Jagjiwan; Lin, Kwang-Lung
2018-04-01
Structural changes in Sn nanoparticles during heating below the melting point have been investigated using differential scanning calorimetry (DSC), x-ray diffraction (XRD) analysis, electron diffraction (ED), and high-resolution transmission electron microscopy (HRTEM). DSC revealed that the heat required to melt the nanoparticles (28.43 J/g) was about half compared with Sn metal (52.80 J/g), which was attributed to the large surface energy contribution for the nanoparticles. ED and XRD analyses of the Sn nanoparticles revealed increased intensity for crystal planes having large interplaner distances compared with regular crystal planes with increasing heat treatment temperature (HTT). HRTEM revealed an increase in interlayer spacing at the surface and near joints between nanoparticles with the HTT, leading to an amorphous structure of nanoparticles at the surface at 220°C. These results highlight the changes that occur in the morphology and crystal structure of Sn nanoparticles at the surface and in the interior with increase of the heat treatment temperature.
A fast time-difference inverse solver for 3D EIT with application to lung imaging.
Javaherian, Ashkan; Soleimani, Manuchehr; Moeller, Knut
2016-08-01
A class of sparse optimization techniques that require solely matrix-vector products, rather than an explicit access to the forward matrix and its transpose, has been paid much attention in the recent decade for dealing with large-scale inverse problems. This study tailors application of the so-called Gradient Projection for Sparse Reconstruction (GPSR) to large-scale time-difference three-dimensional electrical impedance tomography (3D EIT). 3D EIT typically suffers from the need for a large number of voxels to cover the whole domain, so its application to real-time imaging, for example monitoring of lung function, remains scarce since the large number of degrees of freedom of the problem extremely increases storage space and reconstruction time. This study shows the great potential of the GPSR for large-size time-difference 3D EIT. Further studies are needed to improve its accuracy for imaging small-size anomalies.
Effectiveness of Treatment Modalities on Kidney Stone Recurrence.
Zisman, Anna L
2017-10-06
Nephrolithiasis is highly prevalent across all demographic groups in the Western world and beyond, and its incidence rates are rising. In addition to the morbidity of the acute event, stone disease often becomes a lifelong problem that requires preventative therapy to diminish ongoing morbidity. Across the majority of stone types, increased fluid intake and targeted dietary modifications are mainstays of therapy. Specific dietary interventions associated with reduced calcium stone risk include adequate dietary calcium intake and restriction of sodium, protein, and oxalate intake, among others. Pharmaceutical therapy may be required if lifestyle changes are insufficient to minimize risk of stone recurrence, and must be targeted to the specific metabolic abnormalities portending risk for a given patient. Therapeutic options for idiopathic calcium stone disease include thiazides, citrate salts, and uric acid-lowering agents. Alkali salts are also the treatment of choice for uric acid stone disease. Management of struvite stone disease is largely surgical, but acetohydroxamic acid is a proven second line therapy. Cystinuria requires lifestyle modifications and may call for thiol-binding agents. Significant heterogeneity of the clinical population with stone disease has previously limited opportunities for large randomized controlled trials. However, as clinical phenotypes and genotypes are increasingly clarified, there are mounting opportunities for targeted randomized controlled trials in stone prevention. In the meantime, the currently available evidence for both lifestyle and pharmacologic interventions is reviewed herein. Copyright © 2017 by the American Society of Nephrology.
Blaas, Harry; Kroeze, Carolien
2014-10-15
Biodiesel is increasingly considered as an alternative for fossil diesel. Biodiesel can be produced from rapeseed, palm, sunflower, soybean and algae. In this study, the consequences of large-scale production of biodiesel from micro-algae for eutrophication in four large European seas are analysed. To this end, scenarios for the year 2050 are analysed, assuming that in the 27 countries of the European Union fossil diesel will be replaced by biodiesel from algae. Estimates are made for the required fertiliser inputs to algae parks, and how this may increase concentrations of nitrogen and phosphorus in coastal waters, potentially leading to eutrophication. The Global NEWS (Nutrient Export from WaterSheds) model has been used to estimate the transport of nitrogen and phosphorus to the European coastal waters. The results indicate that the amount of nitrogen and phosphorus in the coastal waters may increase considerably in the future as a result of large-scale production of algae for the production of biodiesel, even in scenarios assuming effective waste water treatment and recycling of waste water in algae production. To ensure sustainable production of biodiesel from micro-algae, it is important to develop cultivation systems with low nutrient losses to the environment. Copyright © 2014 Elsevier B.V. All rights reserved.
A practical large scale/high speed data distribution system using 8 mm libraries
NASA Technical Reports Server (NTRS)
Howard, Kevin
1993-01-01
Eight mm tape libraries are known primarily for their small size, large storage capacity, and low cost. However, many applications require an additional attribute which, heretofore, has been lacking -- high transfer rate. Transfer rate is particularly important in a large scale data distribution environment -- an environment in which 8 mm tape should play a very important role. Data distribution is a natural application for 8 mm for several reasons: most large laboratories have access to 8 mm tape drives, 8 mm tapes are upwardly compatible, 8 mm media are very inexpensive, 8 mm media are light weight (important for shipping purposes), and 8 mm media densely pack data (5 gigabytes now and 15 gigabytes on the horizon). If the transfer rate issue were resolved, 8 mm could offer a good solution to the data distribution problem. To that end Exabyte has analyzed four ways to increase its transfer rate: native drive transfer rate increases, data compression at the drive level, tape striping, and homogeneous drive utilization. Exabyte is actively pursuing native drive transfer rate increases and drive level data compression. However, for non-transmitted bulk data applications (which include data distribution) the other two methods (tape striping and homogeneous drive utilization) hold promise.
Large-scale quantum photonic circuits in silicon
NASA Astrophysics Data System (ADS)
Harris, Nicholas C.; Bunandar, Darius; Pant, Mihir; Steinbrecher, Greg R.; Mower, Jacob; Prabhu, Mihika; Baehr-Jones, Tom; Hochberg, Michael; Englund, Dirk
2016-08-01
Quantum information science offers inherently more powerful methods for communication, computation, and precision measurement that take advantage of quantum superposition and entanglement. In recent years, theoretical and experimental advances in quantum computing and simulation with photons have spurred great interest in developing large photonic entangled states that challenge today's classical computers. As experiments have increased in complexity, there has been an increasing need to transition bulk optics experiments to integrated photonics platforms to control more spatial modes with higher fidelity and phase stability. The silicon-on-insulator (SOI) nanophotonics platform offers new possibilities for quantum optics, including the integration of bright, nonclassical light sources, based on the large third-order nonlinearity (χ(3)) of silicon, alongside quantum state manipulation circuits with thousands of optical elements, all on a single phase-stable chip. How large do these photonic systems need to be? Recent theoretical work on Boson Sampling suggests that even the problem of sampling from e30 identical photons, having passed through an interferometer of hundreds of modes, becomes challenging for classical computers. While experiments of this size are still challenging, the SOI platform has the required component density to enable low-loss and programmable interferometers for manipulating hundreds of spatial modes. Here, we discuss the SOI nanophotonics platform for quantum photonic circuits with hundreds-to-thousands of optical elements and the associated challenges. We compare SOI to competing technologies in terms of requirements for quantum optical systems. We review recent results on large-scale quantum state evolution circuits and strategies for realizing high-fidelity heralded gates with imperfect, practical systems. Next, we review recent results on silicon photonics-based photon-pair sources and device architectures, and we discuss a path towards large-scale source integration. Finally, we review monolithic integration strategies for single-photon detectors and their essential role in on-chip feed forward operations.
NASA Astrophysics Data System (ADS)
Manfredi, Sabato
2016-06-01
Large-scale dynamic systems are becoming highly pervasive in their occurrence with applications ranging from system biology, environment monitoring, sensor networks, and power systems. They are characterised by high dimensionality, complexity, and uncertainty in the node dynamic/interactions that require more and more computational demanding methods for their analysis and control design, as well as the network size and node system/interaction complexity increase. Therefore, it is a challenging problem to find scalable computational method for distributed control design of large-scale networks. In this paper, we investigate the robust distributed stabilisation problem of large-scale nonlinear multi-agent systems (briefly MASs) composed of non-identical (heterogeneous) linear dynamical systems coupled by uncertain nonlinear time-varying interconnections. By employing Lyapunov stability theory and linear matrix inequality (LMI) technique, new conditions are given for the distributed control design of large-scale MASs that can be easily solved by the toolbox of MATLAB. The stabilisability of each node dynamic is a sufficient assumption to design a global stabilising distributed control. The proposed approach improves some of the existing LMI-based results on MAS by both overcoming their computational limits and extending the applicative scenario to large-scale nonlinear heterogeneous MASs. Additionally, the proposed LMI conditions are further reduced in terms of computational requirement in the case of weakly heterogeneous MASs, which is a common scenario in real application where the network nodes and links are affected by parameter uncertainties. One of the main advantages of the proposed approach is to allow to move from a centralised towards a distributed computing architecture so that the expensive computation workload spent to solve LMIs may be shared among processors located at the networked nodes, thus increasing the scalability of the approach than the network size. Finally, a numerical example shows the applicability of the proposed method and its advantage in terms of computational complexity when compared with the existing approaches.
Intensive Care Management of Patients with Cirrhosis.
Olson, Jody C
2018-06-01
Cirrhosis is a major worldwide health problem which results in a high level of morbidity and mortality. Patients with cirrhosis who require intensive care support have high mortality rates of near 50%. The goal of this review is to address the management of common complications of cirrhosis in the ICU. Recent epidemiological studies have shown an increase in hospitalizations due to advanced liver disease with an associated increase in intensive care utilization. Given an increasing burden on the healthcare system, it is imperative that we strive to improve our management cirrhotic patients in the intensive care unit. Large studies evaluating the management of patients in the intensive care setting are lacking. To date, most recommendations are based on extrapolation of data from studies in cirrhosis outside of the ICU or by applying general critical care principles which may or may not be appropriate for the critically ill cirrhotic patient. Future research is required to answer important management questions.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pinkston, David
Four slides present a historical perspective on the evolution of nuclear safety, a description of systemic misalignment (available resources do not match expectations, demographic cliff developing, promulgation of increased expectations and new requirements proceeds unabated), and needs facing nuclear safety (financial stability, operational stability, and succession planning). The following conclusions are stated under the heading "Nuclear Safety - 'The System'": the current universe of requirements is too large for the resource pool available; the current universe of requirements has too many different sources of interpretation; there are so many indicators that it’s hard to know what is leading (or important);more » and the net result can come to defy integrated comprehension at the worker level.« less
Orbiting Deep Space Relay Station (ODSRS). Volume 1: Requirement determination
NASA Technical Reports Server (NTRS)
Hunter, J. A.
1979-01-01
The deep space communications requirements of the post-1985 time frame are described and the orbiting deep space relay station (ODSRS) is presented as an option for meeting these requirements. Under current conditions, the ODSRS is not yet cost competitive with Earth based stations to increase DSN telemetry performance, but has significant advantages over a ground station, and these are sufficient to maintain it as a future option. These advantages include: the ability to track a spacecraft 24 hours per day with ground stations located only in the USA; the ability to operate at higher frequencies that would be attenuated by Earth's atmosphere; and the potential for building very large structures without the constraints of Earth's gravity.
Visualization and Analysis for Near-Real-Time Decision Making in Distributed Workflows
Pugmire, David; Kress, James; Choi, Jong; ...
2016-08-04
Data driven science is becoming increasingly more common, complex, and is placing tremendous stresses on visualization and analysis frameworks. Data sources producing 10GB per second (and more) are becoming increasingly commonplace in both simulation, sensor and experimental sciences. These data sources, which are often distributed around the world, must be analyzed by teams of scientists that are also distributed. Enabling scientists to view, query and interact with such large volumes of data in near-real-time requires a rich fusion of visualization and analysis techniques, middleware and workflow systems. Here, this paper discusses initial research into visualization and analysis of distributed datamore » workflows that enables scientists to make near-real-time decisions of large volumes of time varying data.« less
Theoretical study of the effect of ground proximity on the induced efficiency of helicopter rotors
NASA Technical Reports Server (NTRS)
Heyson, H. H.
1977-01-01
A study of rotors in forward flight within ground effect showed that the ground-induced interference is an upwash and a decrease in forward velocity. The interference velocities are large, oppose the normal flow through the rotor, and have large effects on the induced efficiency. Hovering with small ground clearances may result in significant blade stall. As speed is increased from hover in ground effect, power initially increases rather than decreases. At very low heights above the ground, the power requirements become nonlinear with speed as a result of the streamwise interference. The streamwise interference becomes greater as the wake approaches the ground and eventually distorts the wake to form the ground vortex which contributes to certain observed directional stability problems.
Scalable loading of a two-dimensional trapped-ion array
Bruzewicz, Colin D.; McConnell, Robert; Chiaverini, John; Sage, Jeremy M.
2016-01-01
Two-dimensional arrays of trapped-ion qubits are attractive platforms for scalable quantum information processing. Sufficiently rapid reloading capable of sustaining a large array, however, remains a significant challenge. Here with the use of a continuous flux of pre-cooled neutral atoms from a remotely located source, we achieve fast loading of a single ion per site while maintaining long trap lifetimes and without disturbing the coherence of an ion quantum bit in an adjacent site. This demonstration satisfies all major criteria necessary for loading and reloading extensive two-dimensional arrays, as will be required for large-scale quantum information processing. Moreover, the already high loading rate can be increased by loading ions in parallel with only a concomitant increase in photo-ionization laser power and no need for additional atomic flux. PMID:27677357
HITECH spurs EHR vendor competition and innovation, resulting in increased adoption.
Joseph, Seth; Sow, Max; Furukawa, Michael F; Posnack, Steven; Chaffee, Mary Ann
2014-09-01
The Health Information Technology for Economic and Clinical Health (HITECH) Act was enacted to increase electronic health record (EHR) adoption by providers and hospitals. Experts expressed skepticism about whether the program would indeed hasten adoption and could be implemented in time for the initial reporting period. Could EHR vendors meet the certification requirements, and could the industry innovate to meet small-practice needs? This study, in addition to documenting increased provider adoption, provides the first evidence of increased competitiveness and innovation in the EHR industry spurred by HITECH. For example, the number of EHR vendors certified for e-prescribing with Surescripts increased from 96 to 229 over the program's first 3 years. We also find that prescribers in small practices increasingly adopted lower-cost, Web-based e-prescribing and EHR applications at significantly higher rates (15%-35%) than did large practices (3%-4%), which generally have more human and capital resources to make significant investments. These findings suggest that EHR vendors were highly responsive to HITECH requirements and have been adapting their strategies to meet nuanced market needs, providing reason to be optimistic about the Programs' future.
Large field-of-view tiled grating structures for X-ray phase-contrast imaging
NASA Astrophysics Data System (ADS)
Schröter, Tobias J.; Koch, Frieder J.; Meyer, Pascal; Kunka, Danays; Meiser, Jan; Willer, Konstantin; Gromann, Lukas; Marco, Fabio D.; Herzen, Julia; Noel, Peter; Yaroshenko, Andre; Hofmann, Andreas; Pfeiffer, Franz; Mohr, Jürgen
2017-01-01
X-ray grating-based interferometry promises unique new diagnostic possibilities in medical imaging and materials analysis. To transfer this method from scientific laboratories or small-animal applications to clinical radiography applications, compact setups with a large field of view (FoV) are required. Currently the FoV is limited by the grating area, which is restricted due to the complex manufacturing process. One possibility to increase the FoV is tiling individual grating tiles to create one large area grating mounted on a carrier substrate. We investigate theoretically the accuracy needed for a tiling process in all degrees of freedom by applying a simulation approach. We show how the resulting precision requirements can be met using a custom-built frame for exact positioning. Precise alignment is achieved by comparing the fringe patterns of two neighboring grating tiles in a grating interferometer. With this method, the FoV can be extended to practically any desired length in one dimension. First results of a phase-contrast scanning setup with a full FoV of 384 mm × 24 mm show the suitability of this method.
Aad, G.
2015-07-01
Results of a search for new phenomena in final states with an energetic jet and large missing transverse momentum are reported. The search uses 20.3 fb -1 of \\(\\sqrt{s{\\mathrm{}}} = 8\\) TeV data collected in 2012 with the ATLAS detector at the LHC. Events are required to have at least one jet with p T>120 GeV and no leptons. Nine signal regions are considered with increasing missing transverse momentum requirements between E miss T>150 GeV and E miss T>700 GeV. Good agreement is observed between the number of events in data and Standard Model expectations. Results are translated into exclusionmore » limits on models with either large extra spatial dimensions, pair production of weakly interacting dark matter candidates, or production of very light gravitinos in a gauge-mediated supersymmetric model. In addition, limits on the production of an invisibly decaying Higgs-like boson leading to similar topologies in the final state are presented.« less
Hydra: a scalable proteomic search engine which utilizes the Hadoop distributed computing framework
2012-01-01
Background For shotgun mass spectrometry based proteomics the most computationally expensive step is in matching the spectra against an increasingly large database of sequences and their post-translational modifications with known masses. Each mass spectrometer can generate data at an astonishingly high rate, and the scope of what is searched for is continually increasing. Therefore solutions for improving our ability to perform these searches are needed. Results We present a sequence database search engine that is specifically designed to run efficiently on the Hadoop MapReduce distributed computing framework. The search engine implements the K-score algorithm, generating comparable output for the same input files as the original implementation. The scalability of the system is shown, and the architecture required for the development of such distributed processing is discussed. Conclusion The software is scalable in its ability to handle a large peptide database, numerous modifications and large numbers of spectra. Performance scales with the number of processors in the cluster, allowing throughput to expand with the available resources. PMID:23216909
NASA Astrophysics Data System (ADS)
Demir, I.; Krajewski, W. F.
2013-12-01
As geoscientists are confronted with increasingly massive datasets from environmental observations to simulations, one of the biggest challenges is having the right tools to gain scientific insight from the data and communicate the understanding to stakeholders. Recent developments in web technologies make it easy to manage, visualize and share large data sets with general public. Novel visualization techniques and dynamic user interfaces allow users to interact with data, and modify the parameters to create custom views of the data to gain insight from simulations and environmental observations. This requires developing new data models and intelligent knowledge discovery techniques to explore and extract information from complex computational simulations or large data repositories. Scientific visualization will be an increasingly important component to build comprehensive environmental information platforms. This presentation provides an overview of the trends and challenges in the field of scientific visualization, and demonstrates information visualization and communication tools developed within the light of these challenges.
Hydra: a scalable proteomic search engine which utilizes the Hadoop distributed computing framework.
Lewis, Steven; Csordas, Attila; Killcoyne, Sarah; Hermjakob, Henning; Hoopmann, Michael R; Moritz, Robert L; Deutsch, Eric W; Boyle, John
2012-12-05
For shotgun mass spectrometry based proteomics the most computationally expensive step is in matching the spectra against an increasingly large database of sequences and their post-translational modifications with known masses. Each mass spectrometer can generate data at an astonishingly high rate, and the scope of what is searched for is continually increasing. Therefore solutions for improving our ability to perform these searches are needed. We present a sequence database search engine that is specifically designed to run efficiently on the Hadoop MapReduce distributed computing framework. The search engine implements the K-score algorithm, generating comparable output for the same input files as the original implementation. The scalability of the system is shown, and the architecture required for the development of such distributed processing is discussed. The software is scalable in its ability to handle a large peptide database, numerous modifications and large numbers of spectra. Performance scales with the number of processors in the cluster, allowing throughput to expand with the available resources.
NASA Technical Reports Server (NTRS)
Wood, Nathan A.
2005-01-01
Planetary Surface Robot Work Crews (RWC) represent a new class of construction robots for future deployment in planetary exploration. Rovers currently being used for the RWC platform lack the load carrying capabilities required in regular work. Two new rovers, dubbed CrewBots, being designed in JPL's Planetary Robotics Lab specifically for RWC applications greatly increase the load carrying capabilities of the platform. A major component of the rover design was the design of the rocker type suspension, which increases rover mobility. The design of the suspension for the Crewbots departed from the design of recent rovers. While many previous rovers have used internal bevel gear differentials, the increased load requirements of the Crewbots calls for a more robust system. The solution presented is the use of an external modified three-bar, slider-linkage, rocker-style suspension that increases the moment arm of the differential. The final product is a suspension system capable of supporting the extreme loading cases the RWC platform presents, without consuming a large portion of the Crewbots' internal space.
NASA Technical Reports Server (NTRS)
Curreri, Peter A.; Detweiler, Michael
2010-01-01
Creating large space habitats by launching all materials from Earth is prohibitively expensive. Using space resources and space based labor to build space solar power satellites can yield extraordinary profits after a few decades. The economic viability of this program depends on the use of space resources and space labor. To maximize the return on the investment, the early use of high density bolo habitats is required. Other shapes do not allow for the small initial scale required for a quick population increase in space. This study found that 5 Man Year, or 384 person bolo high density habitats will be the most economically feasible for a program started at year 2010 and will cause a profit by year 24 of the program, put over 45,000 people into space, and create a large system of space infrastructure for the further exploration and development of space.
Optimizing distance-based methods for large data sets
NASA Astrophysics Data System (ADS)
Scholl, Tobias; Brenner, Thomas
2015-10-01
Distance-based methods for measuring spatial concentration of industries have received an increasing popularity in the spatial econometrics community. However, a limiting factor for using these methods is their computational complexity since both their memory requirements and running times are in {{O}}(n^2). In this paper, we present an algorithm with constant memory requirements and shorter running time, enabling distance-based methods to deal with large data sets. We discuss three recent distance-based methods in spatial econometrics: the D&O-Index by Duranton and Overman (Rev Econ Stud 72(4):1077-1106, 2005), the M-function by Marcon and Puech (J Econ Geogr 10(5):745-762, 2010) and the Cluster-Index by Scholl and Brenner (Reg Stud (ahead-of-print):1-15, 2014). Finally, we present an alternative calculation for the latter index that allows the use of data sets with millions of firms.
Advanced space system concepts and their orbital support needs (1980 - 2000). Volume 2: Final report
NASA Technical Reports Server (NTRS)
Bekey, I.; Mayer, H. L.; Wolfe, M. G.
1976-01-01
The results are presented of a study which identifies over 100 new and highly capable space systems for the 1980-2000 time period: civilian systems which could bring benefits to large numbers of average citizens in everyday life, much enhance the kinds and levels of public services, increase the economic motivation for industrial investment in space, expand scientific horizons; and, in the military area, systems which could materially alter current concepts of tactical and strategic engagements. The requirements for space transportation, orbital support, and technology for these systems are derived, and those requirements likely to be shared between NASA and the DoD in the time period identified. The high leverage technologies for the time period are identified as very large microwave antennas and optics, high energy power subsystems, high precision and high power lasers, microelectronic circuit complexes and data processors, mosaic solid state sensing devices, and long-life cryogenic refrigerators.
NASA Technical Reports Server (NTRS)
Riester, Peter; Ellis, Colleen; Wagner, Michael; Orren, Scott; Smith, Byron; Skelly, Michael; Zgraggen, Craig; Webber, Matt
1992-01-01
The world is rapidly changing from one with two military superpowers, with which most countries were aligned, to one with many smaller military powers. In this environment, the United States cannot depend on the availability of operating bases from which to respond to crises requiring military intervention. Several studies (e.g. the SAB Global Reach, Global Power Study) have indicated an increased need to be able to rapidly transport large numbers of troops and equipment from the continental United States to potential trouble spots throughout the world. To this end, a request for proposals (RFP) for the concept design of a large aircraft capable of 'projecting' a significant military force without reliance on surface transportation was developed. These design requirements are: minimum payload of 400,000 pounds at 2.5 g maneuver load factor; minimum unfueled range of 6,000 nautical miles; and aircraft must operate from existing domestic air bases and use existing airbases or sites of opportunity at the destination.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aad, G.
Results of a search for new phenomena in final states with an energetic jet and large missing transverse momentum are reported. The search uses 20.3 fb -1 of \\(\\sqrt{s{\\mathrm{}}} = 8\\) TeV data collected in 2012 with the ATLAS detector at the LHC. Events are required to have at least one jet with p T>120 GeV and no leptons. Nine signal regions are considered with increasing missing transverse momentum requirements between E miss T>150 GeV and E miss T>700 GeV. Good agreement is observed between the number of events in data and Standard Model expectations. Results are translated into exclusionmore » limits on models with either large extra spatial dimensions, pair production of weakly interacting dark matter candidates, or production of very light gravitinos in a gauge-mediated supersymmetric model. In addition, limits on the production of an invisibly decaying Higgs-like boson leading to similar topologies in the final state are presented.« less
Mapping CMMI Level 2 to Scrum Practices: An Experience Report
NASA Astrophysics Data System (ADS)
Diaz, Jessica; Garbajosa, Juan; Calvo-Manzano, Jose A.
CMMI has been adopted advantageously in large companies for improvements in software quality, budget fulfilling, and customer satisfaction. However SPI strategies based on CMMI-DEV require heavy software development processes and large investments in terms of cost and time that medium/small companies do not deal with. The so-called light software development processes, such as Agile Software Development (ASD), deal with these challenges. ASD welcomes changing requirements and stresses the importance of adaptive planning, simplicity and continuous delivery of valuable software by short time-framed iterations. ASD is becoming convenient in a more and more global, and changing software market. It would be greatly useful to be able to introduce agile methods such as Scrum in compliance with CMMI process model. This paper intends to increase the understanding of the relationship between ASD and CMMI-DEV reporting empirical results that confirm theoretical comparisons between ASD practices and CMMI level2.
Goldberg, Jesse H.
2011-01-01
Young songbirds produce vocal “babbling,” and the variability of their songs is thought to underlie a process of trial-and-error vocal learning. It is known that this exploratory variability requires the “cortical” component of a basal ganglia (BG) thalamocortical loop, but less understood is the role of the BG and thalamic components in this behavior. We found that large bilateral lesions to the songbird BG homolog Area X had little or no effect on song variability during vocal babbling. In contrast, lesions to the BG-recipient thalamic nucleus DLM (medial portion of the dorsolateral thalamus) largely abolished normal vocal babbling in young birds and caused a dramatic increase in song stereotypy. These findings support the idea that the motor thalamus plays a key role in the expression of exploratory juvenile behaviors during learning. PMID:21430276
Slow light enhanced gas sensing in photonic crystals
NASA Astrophysics Data System (ADS)
Kraeh, Christian; Martinez-Hurtado, J. L.; Popescu, Alexandru; Hedler, Harry; Finley, Jonathan J.
2018-02-01
Infrared spectroscopy allows for highly selective and highly sensitive detection of gas species and concentrations. Conventional gas spectrometers are generally large and unsuitable for on-chip applications. Long absorption path lengths are usually required and impose a challenge for miniaturization. In this work, a gas spectrometer is developed consisting of a microtube photonic crystal structure. This structure of millimetric form factors minimizes the required absorption path length due to slow light effects. The microtube photonic crystal allows for strong transmission in the mid-infrared and, due to its large void space fraction, a strong interaction between light and gas molecules. As a result, enhanced absorption of light increases the gas sensitivity of the device. Slow light enhanced gas absorption by a factor of 5.8 in is experimentally demonstrated at 5400 nm. We anticipate small form factor gas sensors on silicon to be a starting point for on-chip gas sensing architectures.
Wu, Xiaoping; Akgün, Can; Vaughan, J Thomas; Andersen, Peter; Strupp, John; Uğurbil, Kâmil; Van de Moortele, Pierre-François
2010-07-01
Parallel excitation holds strong promises to mitigate the impact of large transmit B1 (B+1) distortion at very high magnetic field. Accelerated RF pulses, however, inherently tend to require larger values in RF peak power which may result in substantial increase in Specific Absorption Rate (SAR) in tissues, which is a constant concern for patient safety at very high field. In this study, we demonstrate adapted rate RF pulse design allowing for SAR reduction while preserving excitation target accuracy. Compared with other proposed implementations of adapted rate RF pulses, our approach is compatible with any k-space trajectories, does not require an analytical expression of the gradient waveform and can be used for large flip angle excitation. We demonstrate our method with numerical simulations based on electromagnetic modeling and we include an experimental verification of transmit pattern accuracy on an 8 transmit channel 9.4 T system.
Production of Bacteriophages by Listeria Cells Entrapped in Organic Polymers.
Roy, Brigitte; Philippe, Cécile; Loessner, Martin J; Goulet, Jacques; Moineau, Sylvain
2018-06-13
Applications for bacteriophages as antimicrobial agents are increasing. The industrial use of these bacterial viruses requires the production of large amounts of suitable strictly lytic phages, particularly for food and agricultural applications. This work describes a new approach for phage production. Phages H387 ( Siphoviridae ) and A511 ( Myoviridae ) were propagated separately using Listeria ivanovii host cells immobilised in alginate beads. The same batch of alginate beads could be used for four successive and efficient phage productions. This technique enables the production of large volumes of high-titer phage lysates in continuous or semi-continuous (fed-batch) cultures.
Use of Patient Registries and Administrative Datasets for the Study of Pediatric Cancer
Rice, Henry E.; Englum, Brian R.; Gulack, Brian C.; Adibe, Obinna O.; Tracy, Elizabeth T.; Kreissman, Susan G.; Routh, Jonathan C.
2015-01-01
Analysis of data from large administrative databases and patient registries is increasingly being used to study childhood cancer care, although the value of these data sources remains unclear to many clinicians. Interpretation of large databases requires a thorough understanding of how the dataset was designed, how data were collected, and how to assess data quality. This review will detail the role of administrative databases and registry databases for the study of childhood cancer, tools to maximize information from these datasets, and recommendations to improve the use of these databases for the study of pediatric oncology. PMID:25807938
Mining large heterogeneous data sets in drug discovery.
Wild, David J
2009-10-01
Increasingly, effective drug discovery involves the searching and data mining of large volumes of information from many sources covering the domains of chemistry, biology and pharmacology amongst others. This has led to a proliferation of databases and data sources relevant to drug discovery. This paper provides a review of the publicly-available large-scale databases relevant to drug discovery, describes the kinds of data mining approaches that can be applied to them and discusses recent work in integrative data mining that looks for associations that pan multiple sources, including the use of Semantic Web techniques. The future of mining large data sets for drug discovery requires intelligent, semantic aggregation of information from all of the data sources described in this review, along with the application of advanced methods such as intelligent agents and inference engines in client applications.
Menzies, Kevin
2014-08-13
The growth in simulation capability over the past 20 years has led to remarkable changes in the design process for gas turbines. The availability of relatively cheap computational power coupled to improvements in numerical methods and physical modelling in simulation codes have enabled the development of aircraft propulsion systems that are more powerful and yet more efficient than ever before. However, the design challenges are correspondingly greater, especially to reduce environmental impact. The simulation requirements to achieve a reduced environmental impact are described along with the implications of continued growth in available computational power. It is concluded that achieving the environmental goals will demand large-scale multi-disciplinary simulations requiring significantly increased computational power, to enable optimization of the airframe and propulsion system over the entire operational envelope. However even with massive parallelization, the limits imposed by communications latency will constrain the time required to achieve a solution, and therefore the position of such large-scale calculations in the industrial design process. © 2014 The Author(s) Published by the Royal Society. All rights reserved.
Material Choice for spindle of machine tools
NASA Astrophysics Data System (ADS)
Gouasmi, S.; Merzoug, B.; Abba, G.; Kherredine, L.
2012-02-01
The requirements of contemporary industry and the flashing development of modern sciences impose restrictions on the majority of the elements of machines; the resulting financial constraints can be satisfied by a better output of the production equipment. As for those concerning the design, the resistance and the correct operation of the product, these require the development of increasingly precise parts, therefore the use of increasingly powerful tools [5]. The precision of machining and the output of the machine tools are generally determined by the precision of rotation of the spindle, indeed, more this one is large more the dimensions to obtain are in the zone of tolerance and the defects of shape are minimized. During the development of the machine tool, the spindle which by definition is a rotating shaft receiving and transmitting to the work piece or the cutting tool the rotational movement, must be designed according to certain optimal parameters to be able to ensure the precision required. This study will be devoted to the choice of the material of the spindle fulfilling the imposed requirements of precision.
NASA Astrophysics Data System (ADS)
Xiong, Ling; Luo, Xiao; Hu, Hai-xiang; Zhang, Zhi-yu; Zhang, Feng; Zheng, Li-gong; Zhang, Xue-jun
2017-08-01
A feasible way to improve the manufacturing efficiency of large reaction-bonded silicon carbide optics is to increase the processing accuracy in the ground stage before polishing, which requires high accuracy metrology. A swing arm profilometer (SAP) has been used to measure large optics during the ground stage. A method has been developed for improving the measurement accuracy of SAP using a capacitive probe and implementing calibrations. The experimental result compared with the interferometer test shows the accuracy of 0.068 μm in root-mean-square (RMS) and maps in 37 low-order Zernike terms show accuracy of 0.048 μm RMS, which shows a powerful capability to provide a major input in high-precision grinding.
Yoo, Sun K; Kim, Dong Keun; Kim, Jung C; Park, Youn Jung; Chang, Byung Chul
2008-01-01
With the increase in demand for high quality medical services, the need for an innovative hospital information system has become essential. An improved system has been implemented in all hospital units of the Yonsei University Health System. Interoperability between multi-units required appropriate hardware infrastructure and software architecture. This large-scale hospital information system encompassed PACS (Picture Archiving and Communications Systems), EMR (Electronic Medical Records) and ERP (Enterprise Resource Planning). It involved two tertiary hospitals and 50 community hospitals. The monthly data production rate by the integrated hospital information system is about 1.8 TByte and the total quantity of data produced so far is about 60 TByte. Large scale information exchange and sharing will be particularly useful for telemedicine applications.
Development of large, horizontal-axis wind turbines
NASA Technical Reports Server (NTRS)
Baldwin, D. H.; Kennard, J.
1985-01-01
A program to develop large, horizontal-axis wind turbines is discussed. The program is directed toward developing the technology for safe, reliable, environmentally acceptable large wind turbines that can generate a significant amount of electricity at costs competitive with those of conventional electricity-generating systems. In addition, these large wind turbines must be fully compatible with electric utility operations and interface requirements. Several ongoing projects in large-wind-turbine development are directed toward meeting the technology requirements for utility applications. The machines based on first-generation technology (Mod-OA and Mod-1) successfully completed their planned periods of experimental operation in June, 1982. The second-generation machines (Mod-2) are in operation at selected utility sites. A third-generation machine (Mod-5) is under contract. Erection and initial operation of the Mod-5 in Hawaii should take place in 1986. Each successive generation of technology increased reliability and energy capture while reducing the cost of electricity. These advances are being made by gaining a better understanding of the system-design drivers, improving the analytical design tools, verifying design methods with operating field data, and incorporating new technology and innovative designs. Information is given on the results from the first- and second-generation machines (Mod-OA, - 1, and -2), the status of the Department of Interior, and the status of the third-generation wind turbine (Mod-5).
Melin, B; Savourey, G
2001-06-30
During ultra-endurance exercise, both increase in body temperature and dehydration due to sweat losses, lead to a decrease in central blood volume. The heart rate drift allows maintaining appropriate cardiac output, in order to satisfy both muscle perfusion and heat transfer requirements by increasing skin blood flow. The resulting dehydration can impair thermal regulation and increase the risks of serious accidents as heat stroke. Endurance events, lasting more than 8 hours, result in large sweat sodium chloride losses. Thus, ingestion of large amounts of water with poor salt intake can induce symptomatic hyponatremia (plasma sodium < 130 mEq/L) which is also a serious accident. Heat environment increases the thermal constraint and when the air humidity is high, evaporation of sweat is compromise. Thus, thermal stress becomes uncompensable which increases the risk of cardiovascular collapse. Cold exposure induces physiological responses to maintain internal temperature by both limiting thermal losses and increasing metabolic heat production. Cold can induce accidental hypothermia and local frost-bites; moreover, it increases the risk of arrhythmia during exercise. Some guidelines (cardiovascular fitness, water and electrolyte intakes, protective clothing) are given for each extreme condition.
Coeli M. Hoover; Mark J. Ducey; R. Andy Colter; Mariko Yamasaki
2018-01-01
There is growing interest in estimating and mapping biomass and carbon content of forests across large landscapes. LiDAR-based inventory methods are increasingly common and have been successfully implemented in multiple forest types. Asner et al. (2011) developed a simple universal forest carbon estimation method for tropical forests that reduces the amount of required...
2000-06-01
As the number of sensors, platforms, exploitation sites, and command and control nodes continues to grow in response to Joint Vision 2010 information ... dominance requirements, Commanders and analysts will have an ever increasing need to collect and process vast amounts of data over wide areas using a large number of disparate sensors and information gathering sources.
ERIC Educational Resources Information Center
Wolf, Alison
The structure of education for 16- to 18-year-olds in Great Britain discourages them from making mathematics, science, and engineering serious options for future study. The emerging structure of the labor market, in which a large proportion of high-status jobs do not require higher mathematics, increases the numbers who decide not to commit…
ERIC Educational Resources Information Center
Handel, Michael J.
2004-01-01
This paper reviews evidence from industry-specific and case studies that shed light on the extent to which computers and automation eliminate jobs, raise job skill requirements, and, consequently, contribute to increased wage inequality between less- and more skilled workers. This paper complements a previous review of large-scale econometric…
A comparison of three methods for measuring local urban tree canopy cover
Kristen L. King; Dexter H. Locke
2013-01-01
Measurements of urban tree canopy cover are crucial for managing urban forests and required for the quantification of the benefits provided by trees. These types of data are increasingly used to secure funding and justify large-scale planting programs in urban areas. Comparisons of tree canopy measurement methods have been conducted before, but a rapidly evolving set...
ERIC Educational Resources Information Center
Nelson, Douglas Allen, Jr.
2017-01-01
Adoption of simulation in healthcare education has increased tremendously over the past two decades. However, the resources necessary to perform simulation are immense. Simulators are large capital investments and require specialized training for both instructors and simulation support staff to develop curriculum using the simulator and to use the…
The Education and Care Divide: The Role of the Early Childhood Workforce in 15 European Countries
ERIC Educational Resources Information Center
Van Laere, Katrien; Peeters, Jan; Vandenbroeck, Michel
2012-01-01
International reports on early childhood education and care tend to attach increasing importance to workforce profiles. Yet a study of 15 European countries reveals that large numbers of (assistant) staff remain invisible in most international reports. As part of the CoRe project (Competence Requirements in Early Childhood Education and Care) we…
Can the United States afford a lunar base
NASA Technical Reports Server (NTRS)
Keaton, Paul W.
1988-01-01
Establishing a lunar base will require steady funding for a decade or two. The question addressed is whether such a large space project is affordable at this time. The relevant facts and methodology are presented so that the reader may formulate independent answers. It is shown that a permanent lunar base can be financed without increasing NASA's historical budgetary trend.
ERIC Educational Resources Information Center
Bowling, William D.
2013-01-01
Post-secondary education is quickly becoming a requirement for many growing careers. Because of this, an increased focused on post-secondary enrollment and attainment has been seen in the education community, particularly in the K-12 systems. To that end a large number of programs and organizations have begun to provide assistance to these…
Novel Concept for Flexible and Resilient Large Power Transformers
DOE Office of Scientific and Technical Information (OSTI.GOV)
Upadhyay, Parag; Englebretson, Steven; Ramanan, V. R. R.
This feasibility study investigates a flexible and adaptable LPT design solution which can facilitate long-term replacement in the event of both catastrophic failures as well as scheduled replacements, thereby increasing grid resilience. The scope of this project has been defined based on an initial system study and identification of the transformer requirements from an overall system load flow perspective.
Brian Roy Lockhart; Jamie E. Kellum
2006-01-01
With the possible re-discovery of the ivory-billed woodpecker (Campephilus principalis), interest has increased in the habitat requirements for the species and the current state of these habitats (Fitzpatrick et al. 2005). Tanner (1942) indicated that the ivory-billed woodpecker needs large, decadent trees for foraging. Trees in such decline provide...
Technical Training Requirements of Middle Management in the Greek Textile and Clothing Industries.
ERIC Educational Resources Information Center
Fotinopoulou, K.; Manolopoulos, N.
A case study of 16 companies in the Greek textile and clothing industry elicited the training needs of the industry's middle managers. The study concentrated on large and medium-sized work units, using a lengthy questionnaire. The study found that middle managers increasingly need to solve problems and ensure the reliability of new equipment and…
ERIC Educational Resources Information Center
Du, Yangbo
2007-01-01
Current U.S. energy policy supports increasing the use of bio-ethanol as a gasoline substitute, which Brazil first produced on a large scale in response to the 1970s energy crises. Brazil's National Alcohol Program stood out among its contemporaries regarding its success at displacing a third of Brazil's gasoline requirements, primarily due to…
ERIC Educational Resources Information Center
Spence, Jane W.
2015-01-01
This study explored elementary and secondary versions of a new principal supervisor role in order to determine whether there might be marked differences in the functions and responsibilities at each level. The new iteration of this role, which is becoming increasingly popular in large urban school districts, requires those occupying it not only to…
Jeffery B. Cannon; Kevin J. Barrett; Benjamin M. Gannon; Robert N. Addington; Mike A. Battaglia; Paula J. Fornwalt; Gregory H. Aplet; Antony S. Cheng; Jeffrey L. Underhill; Jennifer S. Briggs; Peter M. Brown
2018-01-01
In response to large, severe wildfires in historically fire-adapted forests in the western US, policy initiatives, such as the USDA Forest Serviceâs Collaborative Forest Landscape Restoration Program (CFLRP), seek to increase the pace and scale of ecological restoration. One required component of this program is collaborative adaptive management, in which monitoring...
Proteinortho: Detection of (Co-)orthologs in large-scale analysis
2011-01-01
Background Orthology analysis is an important part of data analysis in many areas of bioinformatics such as comparative genomics and molecular phylogenetics. The ever-increasing flood of sequence data, and hence the rapidly increasing number of genomes that can be compared simultaneously, calls for efficient software tools as brute-force approaches with quadratic memory requirements become infeasible in practise. The rapid pace at which new data become available, furthermore, makes it desirable to compute genome-wide orthology relations for a given dataset rather than relying on relations listed in databases. Results The program Proteinortho described here is a stand-alone tool that is geared towards large datasets and makes use of distributed computing techniques when run on multi-core hardware. It implements an extended version of the reciprocal best alignment heuristic. We apply Proteinortho to compute orthologous proteins in the complete set of all 717 eubacterial genomes available at NCBI at the beginning of 2009. We identified thirty proteins present in 99% of all bacterial proteomes. Conclusions Proteinortho significantly reduces the required amount of memory for orthology analysis compared to existing tools, allowing such computations to be performed on off-the-shelf hardware. PMID:21526987
Economic analysis of municipal wastewater utilization for thermoelectric power production
DOE Office of Scientific and Technical Information (OSTI.GOV)
Safari, I.; Walker, M.; Abbasian, J.
2011-01-01
The thermoelectric power industry in the U.S. uses a large amount of freshwater. The large water demand is increasingly a problem, especially for new power plant development, as availability of freshwater for new uses diminishes in the United States. Reusing non-traditional water sources, such as treated municipal wastewater, provides one option to mitigate freshwater usage in the thermoelectric power industry. The amount of freshwater withdrawal that can be displaced with non-traditional water sources at a particular location requires evaluation of the water management and treatment requirements, considering the quality and abundance of the non-traditional water sources. This paper presents themore » development of an integrated costing model to assess the impact of degraded water treatment, as well as the implications of increased tube scaling in the main condenser. The model developed herein is used to perform case studies of various treatment, condenser cleaning and condenser configurations to provide insight into the ramifications of degraded water use in the cooling loops of thermoelectric power plants. Further, this paper lays the groundwork for the integration of relationships between degraded water quality, scaling characteristics and volatile emission within a recirculating cooling loop model.« less
Interacting factors driving a major loss of large trees with cavities in a forest ecosystem.
Lindenmayer, David B; Blanchard, Wade; McBurney, Lachlan; Blair, David; Banks, Sam; Likens, Gene E; Franklin, Jerry F; Laurance, William F; Stein, John A R; Gibbons, Philip
2012-01-01
Large trees with cavities provide critical ecological functions in forests worldwide, including vital nesting and denning resources for many species. However, many ecosystems are experiencing increasingly rapid loss of large trees or a failure to recruit new large trees or both. We quantify this problem in a globally iconic ecosystem in southeastern Australia--forests dominated by the world's tallest angiosperms, Mountain Ash (Eucalyptus regnans). Tree, stand and landscape-level factors influencing the death and collapse of large living cavity trees and the decay and collapse of dead trees with cavities are documented using a suite of long-term datasets gathered between 1983 and 2011. The historical rate of tree mortality on unburned sites between 1997 and 2011 was >14% with a mortality spike in the driest period (2006-2009). Following a major wildfire in 2009, 79% of large living trees with cavities died and 57-100% of large dead trees were destroyed on burned sites. Repeated measurements between 1997 and 2011 revealed no recruitment of any new large trees with cavities on any of our unburned or burned sites. Transition probability matrices of large trees with cavities through increasingly decayed condition states projects a severe shortage of large trees with cavities by 2039 that will continue until at least 2067. This large cavity tree crisis in Mountain Ash forests is a product of: (1) the prolonged time required (>120 years) for initiation of cavities; and (2) repeated past wildfires and widespread logging operations. These latter factors have resulted in all landscapes being dominated by stands ≤72 years and just 1.16% of forest being unburned and unlogged. We discuss how the features that make Mountain Ash forests vulnerable to a decline in large tree abundance are shared with many forest types worldwide.
Energy requirements of tire pulling.
Fredriksen, Per M; Mamen, Asgeir
2017-10-01
We have investigated the effect using walking poles and pulling tires at 4 and 6 km·h-1 (1.11 and 1.67 m·s-1) speeds on oxygen uptake (V̇O2) and heart rate. Eleven subjects, 6 males, with a mean (SD) age of 25.2 (6.9) years participated in field tests involving walking without poles, walking with poles and tire pulling with poles. Increasing the load caused the largest increases in energy demand, more than 4 MET. Speed increase also caused substantial energy increase, approximately 4 MET. Increasing the inclination only modestly increased the oxygen uptake, approximately 2 MET. In both level walking and uphill walking, using poles marginally increased oxygen uptake compared to working without poles. Pulling one tire (12.5 kg) required an oxygen uptake of 27 (4) mL·kg-1·min-1 at 4 km·h-1 and 0% inclination. Adding one more tire (6 kg) drove the oxygen uptake further up to 39 (4) mL·kg-1·min-1. This is close to the requirement of level running at 10.5 km·h-1. Pulling both tires at 6 km·h-1 and 5% inclination required a V̇O2 of 54 (6) mL·kg-1·min-1, equal to running uphill at 5% inclination and 12.5 km·h-1 speed. Heart rate rose comparably with oxygen uptake. At 4 km·h-1 and 0% inclination the increase was 29 bpm, from 134 (21) to 163 (22) bpm when going from pulling one tire to two tires. In the hardest exercise, 6 km·h-1 and 5% inclination, heart rate reached 174 (14) bpm. The study showed that tire pulling even at slow speeds has an energy requirement that is so large that the activity may be feasible as endurance training.
Countermeasure for Radiation Protection and Repair
NASA Technical Reports Server (NTRS)
2008-01-01
Exposure to ionizing radiation during long-duration space missions is expected to cause short-term illness and increase long-term risk of cancer for astronauts. Radiation-induced free radicals overload the antioxidant defense mechanisms and lead to cellular damage at the membrane, enzyme, and chromosome levels. A large number of radioprotective agents were screened, but most had significant side effects. But there is increasing evidence that significant radioprotective benefit is achieved by increasing the dietary intake of foods with high antioxidant potential. Early plant-growing systems for space missions will be limited in both size and volume to minimize power and mass requirements. These systems will be well suited to producing plants containing high concentrations of bioprotective antioxidants. This project explored whether the production of bioprotective compounds could be increased by altering the lighting system, without increasing the space or power requirements for production, and evaluated the effects of environmental conditions (light quantity, light quality, and carbon dioxide [CO2] concentration) on the production of bioprotective compounds in lettuce, which provide a biological countermeasure for radiation exposure. The specific deliverables were to develop a database of bioprotectant compounds in plants that are suitable for use on longduration space missions, develop protocols for maintaining and increasing bioprotectant production under light emitting diodes (LEDs), recommend lighting requirements to produce dietary countermeasures of radiation, and publish results in the Journal of the American Society for Horticultural Science.
30 CFR 49.40 - Requirements for large coal mines.
Code of Federal Regulations, 2011 CFR
2011-07-01
... 30 Mineral Resources 1 2011-07-01 2011-07-01 false Requirements for large coal mines. 49.40 Section 49.40 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR EDUCATION AND TRAINING MINE RESCUE TEAMS Mine Rescue Teams for Underground Coal Mines § 49.40 Requirements for large coal...
30 CFR 49.40 - Requirements for large coal mines.
Code of Federal Regulations, 2012 CFR
2012-07-01
... 30 Mineral Resources 1 2012-07-01 2012-07-01 false Requirements for large coal mines. 49.40 Section 49.40 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR EDUCATION AND TRAINING MINE RESCUE TEAMS Mine Rescue Teams for Underground Coal Mines § 49.40 Requirements for large coal...
30 CFR 49.40 - Requirements for large coal mines.
Code of Federal Regulations, 2013 CFR
2013-07-01
... 30 Mineral Resources 1 2013-07-01 2013-07-01 false Requirements for large coal mines. 49.40 Section 49.40 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR EDUCATION AND TRAINING MINE RESCUE TEAMS Mine Rescue Teams for Underground Coal Mines § 49.40 Requirements for large coal...
30 CFR 49.40 - Requirements for large coal mines.
Code of Federal Regulations, 2010 CFR
2010-07-01
... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Requirements for large coal mines. 49.40 Section 49.40 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR EDUCATION AND TRAINING MINE RESCUE TEAMS Mine Rescue Teams for Underground Coal Mines § 49.40 Requirements for large coal...
30 CFR 49.40 - Requirements for large coal mines.
Code of Federal Regulations, 2014 CFR
2014-07-01
... 30 Mineral Resources 1 2014-07-01 2014-07-01 false Requirements for large coal mines. 49.40 Section 49.40 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR EDUCATION AND TRAINING MINE RESCUE TEAMS Mine Rescue Teams for Underground Coal Mines § 49.40 Requirements for large coal...
Ward, Megan; Johnson, Steven D; Zalucki, Myron P
2013-04-01
One of the essential requirements for an introduced plant species to become invasive is an ability to reproduce outside the native range, particularly when initial populations are small. If a reproductive Allee effect is operating, plants in small populations will have reduced reproductive success relative to plants in larger populations. Alternatively, if plants in small populations experience less competition for pollination than those in large populations, they may actually have higher levels of reproductive success than plants in large populations. To resolve this uncertainty, we investigated how the per capita fecundity of plants was affected by population size in three invasive milkweed species. Field surveys of seed production in natural populations of different sizes but similar densities were conducted for three pollinator-dependent invasive species, namely Asclepias curassavica, Gomphocarpus fruticosus and G. physocarpus. Additionally, supplemental hand-pollinations were performed in small and large populations in order to determine whether reproductive output was limited by pollinator activity in these populations. Reproductive Allee effects were not detected in any of the study species. Instead, plants in small populations exhibited remarkably high levels of reproductive output compared to those in large populations. Increased fruit production following supplemental hand-pollinations suggested that the lower reproductive output of naturally pollinated plants in large populations is a consequence of pollen limitation rather than limitation due to abiotic resources. This is consistent with increased intraspecific competition for pollination amongst plants in large populations. It is likely that the invasion of these milkweed species in Australia has been enhanced because plants in small founding populations experience less intraspecific competition for pollinators than those in large populations, and thus have the ability to produce copious amounts of seeds.
NASA Astrophysics Data System (ADS)
Ross, M. N.; Toohey, D.
2008-12-01
Emissions from solid and liquid propellant rocket engines reduce global stratospheric ozone levels. Currently ~ one kiloton of payloads are launched into earth orbit annually by the global space industry. Stratospheric ozone depletion from present day launches is a small fraction of the ~ 4% globally averaged ozone loss caused by halogen gases. Thus rocket engine emissions are currently considered a minor, if poorly understood, contributor to ozone depletion. Proposed space-based geoengineering projects designed to mitigate climate change would require order of magnitude increases in the amount of material launched into earth orbit. The increased launches would result in comparable increases in the global ozone depletion caused by rocket emissions. We estimate global ozone loss caused by three space-based geoengineering proposals to mitigate climate change: (1) mirrors, (2) sunshade, and (3) space-based solar power (SSP). The SSP concept does not directly engineer climate, but is touted as a mitigation strategy in that SSP would reduce CO2 emissions. We show that launching the mirrors or sunshade would cause global ozone loss between 2% and 20%. Ozone loss associated with an economically viable SSP system would be at least 0.4% and possibly as large as 3%. It is not clear which, if any, of these levels of ozone loss would be acceptable under the Montreal Protocol. The large uncertainties are mainly caused by a lack of data or validated models regarding liquid propellant rocket engine emissions. Our results offer four main conclusions. (1) The viability of space-based geoengineering schemes could well be undermined by the relatively large ozone depletion that would be caused by the required rocket launches. (2) Analysis of space- based geoengineering schemes should include the difficult tradeoff between the gain of long-term (~ decades) climate control and the loss of short-term (~ years) deep ozone loss. (3) The trade can be properly evaluated only if our understanding of the stratospheric impact of rocket emissions is significantly improved. (4) Such an improved understanding requires a concerted effort of research including new in situ measurements in a variety of rocket plumes and a multi-scale modeling program similar in scope to the effort required to address the climate and ozone impacts of aircraft emissions.
Emergency egress requirements for Space Station Freedom
NASA Technical Reports Server (NTRS)
Ray, Paul S.
1990-01-01
There is a real concern regarding the requirements for safe emergency egress from the Space Station Freedom (SSF). The possible causes of emergency are depressurization due to breach of the station hull by space debris, meteoroids, seal failure, or vent failure; chemical toxicity; and a large fire. The objectives of the current study are to identify the tasks required to be performed in emergencies, establish the time required to perform these tasks, and to review the human equipment interface in emergencies. It was found that a fixed time value specified for egress has shifted focus from the basic requirements of safe egress, that in some situations the crew members may not be able to complete the emergency egress tasks in three minutes without sacrificing more than half of the station, and that increased focus should be given to human factors aspects of space station design.
Anabolic-androgenic steroids and decision making: Probability and effort discounting in male rats.
Wallin, Kathryn G; Alves, Jasmin M; Wood, Ruth I
2015-07-01
Anabolic-androgenic steroid (AAS) abuse is implicated in maladaptive behaviors such as increased aggression and risk taking. Impaired judgment due to changes in the mesocorticolimbic dopamine system may contribute to these behavioral changes. While AAS are known to influence dopamine function in mesocorticolimbic circuitry, the effects on decision making are unknown. This was the focus of the present study. Adolescent male Long-Evans rats were treated chronically with high-dose testosterone (7.5 mg/kg) or vehicle (13% cyclodextrin in water), and tested for cost/benefit decision making in two discounting paradigms. Rats chose between a small reward (1 sugar pellet) and a large discounted reward (3 or 4 pellets). Probability discounting (PD) measures sensitivity to reward uncertainty by decreasing the probability (100, 75, 50, 25, 0%) of receiving the large reward in successive blocks of each daily session. Effort discounting (ED) measures sensitivity to a work cost by increasing the lever presses required to earn the large reward (1, 2, 5, 10, 15 presses). In PD, testosterone-treated rats selected the large/uncertain reward significantly less than vehicle-treated controls. However, during ED, testosterone-treated rats selected the large/high effort reward significantly more than controls. These studies show that testosterone has divergent effects on different aspects of decision making. Specifically, testosterone increases aversion to uncertainty but decreases sensitivity to the output of effort for reward. These results have implications for understanding maladaptive behavioral changes in human AAS users. Copyright © 2015 Elsevier Ltd. All rights reserved.
Large-baseline InSAR for precise topographic mapping: a framework for TanDEM-X large-baseline data
NASA Astrophysics Data System (ADS)
Pinheiro, Muriel; Reigber, Andreas; Moreira, Alberto
2017-09-01
The global Digital Elevation Model (DEM) resulting from the TanDEM-X mission provides information about the world topography with outstanding precision. In fact, performance analysis carried out with the already available data have shown that the global product is well within the requirements of 10 m absolute vertical accuracy and 2 m relative vertical accuracy for flat to moderate terrain. The mission's science phase took place from October 2014 to December 2015. During this phase, bistatic acquisitions with across-track separation between the two satellites up to 3.6 km at the equator were commanded. Since the relative vertical accuracy of InSAR derived elevation models is, in principle, inversely proportional to the system baseline, the TanDEM-X science phase opened the doors for the generation of elevation models with improved quality with respect to the standard product. However, the interferometric processing of the large-baseline data is troublesome due to the increased volume decorrelation and very high frequency of the phase variations. Hence, in order to fully profit from the increased baseline, sophisticated algorithms for the interferometric processing, and, in particular, for the phase unwrapping have to be considered. This paper proposes a novel dual-baseline region-growing framework for the phase unwrapping of the large-baseline interferograms. Results from two experiments with data from the TanDEM-X science phase are discussed, corroborating the expected increased level of detail of the large-baseline DEMs.
An optimal FFT-based anisotropic power spectrum estimator
NASA Astrophysics Data System (ADS)
Hand, Nick; Li, Yin; Slepian, Zachary; Seljak, Uroš
2017-07-01
Measurements of line-of-sight dependent clustering via the galaxy power spectrum's multipole moments constitute a powerful tool for testing theoretical models in large-scale structure. Recent work shows that this measurement, including a moving line-of-sight, can be accelerated using Fast Fourier Transforms (FFTs) by decomposing the Legendre polynomials into products of Cartesian vectors. Here, we present a faster, optimal means of using FFTs for this measurement. We avoid redundancy present in the Cartesian decomposition by using a spherical harmonic decomposition of the Legendre polynomials. With this method, a given multipole of order l requires only 2l+1 FFTs rather than the (l+1)(l+2)/2 FFTs of the Cartesian approach. For the hexadecapole (l = 4), this translates to 40% fewer FFTs, with increased savings for higher l. The reduction in wall-clock time enables the calculation of finely-binned wedges in P(k,μ), obtained by computing multipoles up to a large lmax and combining them. This transformation has a number of advantages. We demonstrate that by using non-uniform bins in μ, we can isolate plane-of-sky (angular) systematics to a narrow bin at 0μ simeq while eliminating the contamination from all other bins. We also show that the covariance matrix of clustering wedges binned uniformly in μ becomes ill-conditioned when combining multipoles up to large values of lmax, but that the problem can be avoided with non-uniform binning. As an example, we present results using lmax=16, for which our procedure requires a factor of 3.4 fewer FFTs than the Cartesian method, while removing the first μ bin leads only to a 7% increase in statistical error on f σ8, as compared to a 54% increase with lmax=4.
Weegman, Bradley P.; Nash, Peter; Carlson, Alexandra L.; Voltzke, Kristin J.; Geng, Zhaohui; Jahani, Marjan; Becker, Benjamin B.; Papas, Klearchos K.; Firpo, Meri T.
2013-01-01
Cellular therapies are emerging as a standard approach for the treatment of several diseases. However, realizing the promise of cellular therapies across the full range of treatable disorders will require large-scale, controlled, reproducible culture methods. Bioreactor systems offer the scale-up and monitoring needed, but standard stirred bioreactor cultures do not allow for the real-time regulation of key nutrients in the medium. In this study, β-TC6 insulinoma cells were aggregated and cultured for 3 weeks as a model of manufacturing a mammalian cell product. Cell expansion rates and medium nutrient levels were compared in static, stirred suspension bioreactors (SSB), and continuously fed (CF) SSB. While SSB cultures facilitated increased culture volumes, no increase in cell yields were observed, partly due to limitations in key nutrients, which were consumed by the cultures between feedings, such as glucose. Even when glucose levels were increased to prevent depletion between feedings, dramatic fluctuations in glucose levels were observed. Continuous feeding eliminated fluctuations and improved cell expansion when compared with both static and SSB culture methods. Further improvements in growth rates were observed after adjusting the feed rate based on calculated nutrient depletion, which maintained physiological glucose levels for the duration of the expansion. Adjusting the feed rate in a continuous medium replacement system can maintain the consistent nutrient levels required for the large-scale application of many cell products. Continuously fed bioreactor systems combined with nutrient regulation can be used to improve the yield and reproducibility of mammalian cells for biological products and cellular therapies and will facilitate the translation of cell culture from the research lab to clinical applications. PMID:24204645
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hand, Nick; Seljak, Uroš; Li, Yin
Measurements of line-of-sight dependent clustering via the galaxy power spectrum's multipole moments constitute a powerful tool for testing theoretical models in large-scale structure. Recent work shows that this measurement, including a moving line-of-sight, can be accelerated using Fast Fourier Transforms (FFTs) by decomposing the Legendre polynomials into products of Cartesian vectors. Here, we present a faster, optimal means of using FFTs for this measurement. We avoid redundancy present in the Cartesian decomposition by using a spherical harmonic decomposition of the Legendre polynomials. With this method, a given multipole of order ℓ requires only 2ℓ+1 FFTs rather than the (ℓ+1)(ℓ+2)/2 FFTsmore » of the Cartesian approach. For the hexadecapole (ℓ = 4), this translates to 40% fewer FFTs, with increased savings for higher ℓ. The reduction in wall-clock time enables the calculation of finely-binned wedges in P ( k ,μ), obtained by computing multipoles up to a large ℓ{sub max} and combining them. This transformation has a number of advantages. We demonstrate that by using non-uniform bins in μ, we can isolate plane-of-sky (angular) systematics to a narrow bin at 0μ ≅ while eliminating the contamination from all other bins. We also show that the covariance matrix of clustering wedges binned uniformly in μ becomes ill-conditioned when combining multipoles up to large values of ℓ{sub max}, but that the problem can be avoided with non-uniform binning. As an example, we present results using ℓ{sub max}=16, for which our procedure requires a factor of 3.4 fewer FFTs than the Cartesian method, while removing the first μ bin leads only to a 7% increase in statistical error on f σ{sub 8}, as compared to a 54% increase with ℓ{sub max}=4.« less
A distributed computing approach to mission operations support. [for spacecraft
NASA Technical Reports Server (NTRS)
Larsen, R. L.
1975-01-01
Computing mission operation support includes orbit determination, attitude processing, maneuver computation, resource scheduling, etc. The large-scale third-generation distributed computer network discussed is capable of fulfilling these dynamic requirements. It is shown that distribution of resources and control leads to increased reliability, and exhibits potential for incremental growth. Through functional specialization, a distributed system may be tuned to very specific operational requirements. Fundamental to the approach is the notion of process-to-process communication, which is effected through a high-bandwidth communications network. Both resource-sharing and load-sharing may be realized in the system.
Disease control in hatchery fish
Fish, F.F.
1947-01-01
The method described herein has been extensively tested, both in the laboratory and at the producing hatchery, over a period of several years. Once familiarity with the details of application have been mastered, th8 reduction in effort required to treat fish is amazing. For example, two men have treated 20 large ponds containing several million fish, in one morning with no significant increase in mortality of the fish, whereas a crew of eight men required a full day to treat a single similar pond by hand dipping the fish with a subsequent loss approximating 50 percent of the stock.
Studying three-phase supply in school
NASA Astrophysics Data System (ADS)
Singhal, Amit Kumar; Arun, P.
2009-07-01
The power distributions of nearly all major countries have accepted three-phase distribution as a standard. With increasing power requirements of instrumentation today even a small physics laboratory requires a three-phase supply. While physics students are given an introduction to this in passing, no experimental work is done with three-phase supply due to the possibility of accidents while working with such large power. We believe a conceptual understanding of three-phase supply would be useful for physics students, with hands-on experience using a simple circuit that can be assembled even in a high school laboratory.
Gao, Johnway [Richland, WA; Skeen, Rodney S [Pendleton, OR
2002-05-28
The present invention is a pulse spilling self-aerator (PSSA) that has the potential to greatly lower the installation, operation, and maintenance cost associated with aerating and mixing aqueous solutions. Currently, large quantities of low-pressure air are required in aeration systems to support many biochemical production processes and wastewater treatment plants. Oxygen is traditionally supplied and mixed by a compressor or blower and a mechanical agitator. These systems have high-energy requirements and high installation and maintenance costs. The PSSA provides a mixing and aeration capability that can increase operational efficiency and reduce overall cost.
Determination of $sup 241$Am in soil using an automated nuclear radiation measurement laboratory
DOE Office of Scientific and Technical Information (OSTI.GOV)
Engstrom, D.E.; White, M.G.; Dunaway, P.B.
The recent completion of REECo's Automated Laboratory and associated software systems has provided a significant increase in capability while reducing manpower requirements. The system is designed to perform gamma spectrum analyses on the large numbers of samples required by the current Nevada Applied Ecology Group (NAEG) and Plutonium Distribution Inventory Program (PDIP) soil sampling programs while maintaining sufficient sensitivities as defined by earlier investigations of the same type. The hardware and systems are generally described in this paper, with emphasis being placed on spectrum reduction and the calibration procedures used for soil samples. (auth)
AQBE — QBE Style Queries for Archetyped Data
NASA Astrophysics Data System (ADS)
Sachdeva, Shelly; Yaginuma, Daigo; Chu, Wanming; Bhalla, Subhash
Large-scale adoption of electronic healthcare applications requires semantic interoperability. The new proposals propose an advanced (multi-level) DBMS architecture for repository services for health records of patients. These also require query interfaces at multiple levels and at the level of semi-skilled users. In this regard, a high-level user interface for querying the new form of standardized Electronic Health Records system has been examined in this study. It proposes a step-by-step graphical query interface to allow semi-skilled users to write queries. Its aim is to decrease user effort and communication ambiguities, and increase user friendliness.
Hybrid Propulsion Technology Program
NASA Technical Reports Server (NTRS)
Jensen, G. E.; Holzman, A. L.
1990-01-01
Future launch systems of the United States will require improvements in booster safety, reliability, and cost. In order to increase payload capabilities, performance improvements are also desirable. The hybrid rocket motor (HRM) offers the potential for improvements in all of these areas. The designs are presented for two sizes of hybrid boosters, a large 4.57 m (180 in.) diameter booster duplicating the Advanced Solid Rocket Motor (ASRM) vacuum thrust-time profile and smaller 2.44 m (96 in.), one-quater thrust level booster. The large booster would be used in tandem, while eight small boosters would be used to achieve the same total thrust. These preliminary designs were generated as part of the NASA Hybrid Propulsion Technology Program. This program is the first phase of an eventual three-phaes program culminating in the demonstration of a large subscale engine. The initial trade and sizing studies resulted in preferred motor diameters, operating pressures, nozzle geometry, and fuel grain systems for both the large and small boosters. The data were then used for specific performance predictions in terms of payload and the definition and selection of the requirements for the major components: the oxidizer feed system, nozzle, and thrust vector system. All of the parametric studies were performed using realistic fuel regression models based upon specific experimental data.
A Computational Study of the Effect of Winglets on the Performance of Micro-Aerial Vehicles
NASA Astrophysics Data System (ADS)
Mönttinen, Jarmo T.; Reed, Helen L.; Squires, Kyle D.; Saric, William S.
2003-11-01
Since mid-1990's an increased interest in developing Micro-Aerial Vehicles (MAVs) has been expressed by military and civilian entities. An MAV is required to have a nominal maximum dimension of 6 inches and to fly at 5 to 20 m/s, which leads to chord Reynolds numbers of 50,000 to 200,000. Despite a wide variety of research projects in universities, companies, and government agencies, the MAV-design problem remains unsolved to a satisfactory manner. The current study uses the Finite Volume solver Cobalt to computationally investigate the effect of winglets on the performance of MAVs. Historically the effectiveness of winglets is addressed in terms of drag reduction. For MAVs, the increase in lift obtained through the addition of winglets is more important as this increase is required to enable low-speed flight. The current results show that winglets can lead to an increase in lift that is sufficiently large to improve the lift-to-drag-ratio as well despite the increased form drag that is typically larger than the decrease in the induced drag.
NASA Technical Reports Server (NTRS)
Yeager, W. T., Jr.; Young, W. H., Jr.; Mantay, W. R.
1974-01-01
An investigation was conducted in the Langley full-scale tunnel to measure the performance of several helicopter tail-rotor/fin configurations with regard to directional control problems encountered at low speeds in ground effect. Tests were conducted at wind azimuths of 0 deg to 360 deg in increments of 30 deg and 60 deg and at wind speeds from 0 to 35 knots. The results indicate that at certain combinations of wind speed and wind azimuth, large increases in adverse fin force require correspondingly large increases in the tail-rotor thrust, collective pitch, and power required to maintain yaw trim. Changing the tail-rotor direction of rotation to top blade aft for either a pusher tail rotor (tail-rotor wake blowing away from fin) or a tractor tail rotor (tail-rotor wake blowing against fin) will alleviate this problem. For a pusher tail rotor at 180 deg wind azimuth, increases in the fin/tail-rotor gap were not found to have any significant influence on the overall vehicle directional control capability. Changing the tail rotor to a higher position was found to improve tail-rotor performance for a fin-off configuration at a wind azimuth of 180 deg. A V-tail configuration with a pusher tail rotor with top blade aft direction of rotation was found to be the best configuration with regard to overall directional control capability.
Obstetric fistula in low and middle income countries.
Capes, Tracy; Ascher-Walsh, Charles; Abdoulaye, Idrissa; Brodman, Michael
2011-01-01
Vesicovaginal fistula secondary to obstructed labor continues to be an all-too-common occurrence in underdeveloped nations throughout Africa and Asia. Vesicovaginal fistula remains largely an overlooked problem in developing nations as it affects the most marginalized members of society: young, poor, illiterate women who live in remote areas. The formation of obstetric fistula is a result of complex interactions of social, biologic, and economic influences. The key underlying causes of fistula are the combination of a lack of functional emergency obstetric care, poverty, illiteracy, and low status of women. In order to prevent fistula, some strategies include creation of governmental policy aimed toward reducing maternal mortality/morbidity and increasing availability of skilled obstetric care, as well as attempts to increase awareness about its prevention and treatment among policymakers, service providers, and communities. Whereas prevention will require the widespread development of infrastructure within these developing countries, treatment of fistula is an act which can be done "in the now." Treatment and subsequent reintegration of fistula patients requires a team of specialists including surgeons, nurses, midwives, and social workers, which is largely unavailable in developing countries. However, there is increasing support for training of fistula surgeons through standardized programs as well as establishment of rehabilitation centers in many nations. The eradication of fistula is dependent upon building programs that target both prevention and treatment. © 2011 Mount Sinai School of Medicine.
The effect of brain size evolution on feeding propensity, digestive efficiency, and juvenile growth.
Kotrschal, Alexander; Corral-Lopez, Alberto; Szidat, Sönke; Kolm, Niclas
2015-11-01
One key hypothesis in the study of brain size evolution is the expensive tissue hypothesis; the idea that increased investment into the brain should be compensated by decreased investment into other costly organs, for instance the gut. Although the hypothesis is supported by both comparative and experimental evidence, little is known about the potential changes in energetic requirements or digestive traits following such evolutionary shifts in brain and gut size. Organisms may meet the greater metabolic requirements of larger brains despite smaller guts via increased food intake or better digestion. But increased investment in the brain may also hamper somatic growth. To test these hypotheses we here used guppy (Poecilia reticulata) brain size selection lines with a pronounced negative association between brain and gut size and investigated feeding propensity, digestive efficiency (DE), and juvenile growth rate. We did not find any difference in feeding propensity or DE between large- and small-brained individuals. Instead, we found that large-brained females had slower growth during the first 10 weeks after birth. Our study provides experimental support that investment into larger brains at the expense of gut tissue carries costs that are not necessarily compensated by a more efficient digestive system. © 2015 The Author(s). Evolution published by Wiley Periodicals, Inc. on behalf of The Society for the Study of Evolution.
Gupta, Vipul; Feng, Kent; Cheruvu, Pavan; Boyer, Nathan; Yeghiazarians, Yerem; Ports, Thomas A; Zimmet, Jeffrey; Shunk, Kendrick; Boyle, Andrew J
2014-09-01
Recent advances in technology have led to an increase in the use of bilateral femoral artery access and the requirement for large-bore access. Optimal access is in the common femoral artery (CFA), rather than higher (in the external iliac artery) or lower (in one of the branches of the CFA). However, there is a paucity of data in the literature about the relationship between bifurcation level of one CFA and the contralateral CFA. To define the prevalence of high bifurcation of the CFA and the relationship between bifurcation level on both sides, we performed a retrospective analysis of all patients with bilateral femoral angiography. From 4880 femoral angiograms performed at UCSF cardiac catheterization laboratory between 2005-2013, a total of 273 patients had bilateral femoral angiograms. The prevalence of low/normal, high, and very-high femoral bifurcations was 70%, 26%, and 4%, respectively, with no difference between sides. A high or very-high bifurcation significantly increased the likelihood of a high bifurcation on the contralateral side (odds ratio >3.0). Multivariable logistic regression analysis revealed age, gender, self-reported race, height, weight, and body mass index were not predictive of high or very-high bifurcations on either side. In conclusion, high femoral artery bifurcations are common and increase the likelihood of a high bifurcation of the contralateral femoral artery.
Preliminary study of a large span-distributed-load flying-wing cargo airplane concept
NASA Technical Reports Server (NTRS)
Jernell, L. S.
1978-01-01
An aircraft capable of transporting containerized cargo over intercontinental distances is analyzed. The specifications for payload weight, density, and dimensions in essence configure the wing and establish unusually low values of wing loading and aspect ratio. The structural weight comprises only about 18 percent of the design maximum gross weight. Although the geometric aspect ratio is 4.53, the winglet effect of the wing-tip-mounted vertical tails, increase the effective aspect ratio to approximately 7.9. Sufficient control power to handle the large rolling moment of inertia dictates a relatively high minimum approach velocity of 315 km/hr (170 knots). The airplane has acceptable spiral, Dutch roll, and roll-damping modes. A hardened stability augmentation system is required. The most significant noise source is that of the airframe. However, for both take-off and approach, the levels are below the FAR-36 limit of 108 db. The design mission fuel efficiency is approximately 50 percent greater than that of the most advanced, currently operational, large freighter aircraft. The direct operating cost is significantly lower than that of current freighters, the advantage increasing as fuel price increases.
Preliminary study of a large span-distributed-load flying-wing cargo airplane concept
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jernell, L.S.
1978-05-01
An aircraft capable of transporting containerized cargo over intercontinental distances is analyzed. The specifications for payload weight, density, and dimensions in essence configure the wing and establish unusually low values of wing loading and aspect ratio. The structural weight comprises only about 18 percent of the design maximum gross weight. Although the geometric aspect ratio is 4.53, the winglet effect of the wing-tip-mounted vertical tails, increase the effective aspect ratio to approximately 7.9. Sufficient control power to handle the large rolling moment of inertia dictates a relatively high minimum approach velocity of 315 km/hr (170 knots). The airplane has acceptablemore » spiral, Dutch roll, and roll-damping modes. A hardened stability augmentation system is required. The most significant noise source is that of the airframe. However, for both take-off and approach, the levels are below the FAR-36 limit of 108 db. The design mission fuel efficiency is approximately 50 percent greater than that of the most advanced, currently operational, large freighter aircraft. The direct operating cost is significantly lower than that of current freighters, the advantage increasing as fuel price increases.« less
Use of Advanced Solar Cells for Commercial Communication Satellites
NASA Technical Reports Server (NTRS)
Bailey, Sheila G.; Landis, Geoffrey A.
1995-01-01
The current generation of communications satellites are located primarily in geosynchronous Earth orbit (GEO). Over the next decade, however, a new generation of communications satellites will be built and launched, designed to provide a world-wide interconnection of portable telephones. For this mission, the satellites must be positioned in lower polar and near-polar orbits. To provide complete coverage, large numbers of satellites will be required. Because the required number of satellites decreases as the orbital altitude is increased, fewer satellites would be required if the orbit chosen were raised from low to intermediate orbit. However, in intermediate orbits, satellites encounter significant radiation due to trapped electrons and protons. Radiation tolerant solar cells may be necessary to make such satellites feasible. We analyze the amount of radiation encountered in low and intermediate polar orbits at altitudes of interest to next-generation communication satellites, calculate the expected degradation for silicon, GaAs, and InP solar cells, and show that the lifetimes can be significantly increased by use of advanced solar cells.
Use of advanced solar cells for commerical communication satellites
NASA Astrophysics Data System (ADS)
Landis, Geoffrey A.; Bailey, Sheila G.
1995-01-01
The current generation of communications satellites are located primarily in geosynchronous Earth orbit (GEO). Over the next decade, however, a new generation of communications satellites will be built and launched, designed to provide a world-wide interconnection of portable telephones. For this mission, the satellites must be positioned in lower polar- and near-polar orbits. To provide complete coverage, large numbers of satellites will be required. Because of the required number of satellites decreases as the orbital altitude is increased, fewer satellites would be required if the orbit chosen were raised from Low to intermediate orbit. However, in intermediate orbits, satellites encounter significant radiation due to trapped electrons and protons. Radiation tolerant solar cells may be necessary to make such satellites feasible. We analyze the amount of radiation encountered in low and intermediate polar orbits at altitudes of interest to next-generation communication satellites, calculate the expected degradation for silicon, GaAs, and InP solar cells, and show that the lifetimes can be significantly increased by use of advanced solar cells.
Use of advanced solar cells for commercial communication satellites
NASA Astrophysics Data System (ADS)
Bailey, Sheila G.; Landis, Geoffrey A.
1995-03-01
The current generation of communications satellites are located primarily in geosynchronous Earth orbit (GEO). Over the next decade, however, a new generation of communications satellites will be built and launched, designed to provide a world-wide interconnection of portable telephones. For this mission, the satellites must be positioned in lower polar and near-polar orbits. To provide complete coverage, large numbers of satellites will be required. Because the required number of satellites decreases as the orbital altitude is increased, fewer satellites would be required if the orbit chosen were raised from low to intermediate orbit. However, in intermediate orbits, satellites encounter significant radiation due to trapped electrons and protons. Radiation tolerant solar cells may be necessary to make such satellites feasible. We analyze the amount of radiation encountered in low and intermediate polar orbits at altitudes of interest to next-generation communication satellites, calculate the expected degradation for silicon, GaAs, and InP solar cells, and show that the lifetimes can be significantly increased by use of advanced solar cells.
Hardie, Jason; Spruston, Nelson
2009-03-11
Long-term potentiation (LTP) requires postsynaptic depolarization that can result from EPSPs paired with action potentials or larger EPSPs that trigger dendritic spikes. We explored the relative contribution of these sources of depolarization to LTP induction during synaptically driven action potential firing in hippocampal CA1 pyramidal neurons. Pairing of a weak test input with a strong input resulted in large LTP (approximately 75% increase) when the weak and strong inputs were both located in the apical dendrites. This form of LTP did not require somatic action potentials. When the strong input was located in the basal dendrites, the resulting LTP was smaller (< or =25% increase). Pairing the test input with somatically evoked action potentials mimicked this form of LTP. Thus, back-propagating action potentials may contribute to modest LTP, but local synaptic depolarization and/or dendritic spikes mediate a stronger form of LTP that requires spatial proximity of the associated synaptic inputs.
Water requirements of selected industries
,; Mussey, Orville D.; Conklin, Howard L.; Durfor, Charles N.; Otts, Louis Ethelbert; Walling, Faulkner B.
1955-01-01
The early industries in America generally were established when and where demands for the products of industry arose. Most of the early industries were so located that their increasing requirements for transportation, raw materials, market, labor, and water supply could be satisfied economically. Many of these original plant locations have continued as modern industrial centers and their output has increased manyfold in meeting the demands of our growing Nation. The recent and current industrial expansion and the trend toward the growth of chemical industries, many Of which are heavy users of water, has resulted in a tremendous increase in the total withdrawal of water for industrial use as well as a large increase in the per capita use of water. This increase in industrial water requirement has strained the capacity of the developed water supplies in many areas, and in some instances the adequacy of the potential water supplies is questionable. The Geological Survey is engaged in preparing and publishing a series of reports describing the developed and undeveloped water resources of many important industrial areas. This work was started initially at the request of the National Securities Resources Board as a means to insure that water supplies are adequate for our rapidly expanding industrial development. Although many factors contribute to establishing the feasibility or even the limits of future industrial development, the one relating to available water supply is extremely important. A knowledge of the water requirements of various industries is valuable therefore in planning the logical development in any area where water supply is a critical factor. Thus far very little suitable information on the water requirements of our major industries is available for general planning. An inventory of unit water-use values in industry therefore would be generally helpful and also might tend to stimulate water-conservation methods. To obtain such information, investigations
Sun, Xiaobo; Gao, Jingjing; Jin, Peng; Eng, Celeste; Burchard, Esteban G; Beaty, Terri H; Ruczinski, Ingo; Mathias, Rasika A; Barnes, Kathleen; Wang, Fusheng; Qin, Zhaohui S
2018-06-01
Sorted merging of genomic data is a common data operation necessary in many sequencing-based studies. It involves sorting and merging genomic data from different subjects by their genomic locations. In particular, merging a large number of variant call format (VCF) files is frequently required in large-scale whole-genome sequencing or whole-exome sequencing projects. Traditional single-machine based methods become increasingly inefficient when processing large numbers of files due to the excessive computation time and Input/Output bottleneck. Distributed systems and more recent cloud-based systems offer an attractive solution. However, carefully designed and optimized workflow patterns and execution plans (schemas) are required to take full advantage of the increased computing power while overcoming bottlenecks to achieve high performance. In this study, we custom-design optimized schemas for three Apache big data platforms, Hadoop (MapReduce), HBase, and Spark, to perform sorted merging of a large number of VCF files. These schemas all adopt the divide-and-conquer strategy to split the merging job into sequential phases/stages consisting of subtasks that are conquered in an ordered, parallel, and bottleneck-free way. In two illustrating examples, we test the performance of our schemas on merging multiple VCF files into either a single TPED or a single VCF file, which are benchmarked with the traditional single/parallel multiway-merge methods, message passing interface (MPI)-based high-performance computing (HPC) implementation, and the popular VCFTools. Our experiments suggest all three schemas either deliver a significant improvement in efficiency or render much better strong and weak scalabilities over traditional methods. Our findings provide generalized scalable schemas for performing sorted merging on genetics and genomics data using these Apache distributed systems.
Gao, Jingjing; Jin, Peng; Eng, Celeste; Burchard, Esteban G; Beaty, Terri H; Ruczinski, Ingo; Mathias, Rasika A; Barnes, Kathleen; Wang, Fusheng
2018-01-01
Abstract Background Sorted merging of genomic data is a common data operation necessary in many sequencing-based studies. It involves sorting and merging genomic data from different subjects by their genomic locations. In particular, merging a large number of variant call format (VCF) files is frequently required in large-scale whole-genome sequencing or whole-exome sequencing projects. Traditional single-machine based methods become increasingly inefficient when processing large numbers of files due to the excessive computation time and Input/Output bottleneck. Distributed systems and more recent cloud-based systems offer an attractive solution. However, carefully designed and optimized workflow patterns and execution plans (schemas) are required to take full advantage of the increased computing power while overcoming bottlenecks to achieve high performance. Findings In this study, we custom-design optimized schemas for three Apache big data platforms, Hadoop (MapReduce), HBase, and Spark, to perform sorted merging of a large number of VCF files. These schemas all adopt the divide-and-conquer strategy to split the merging job into sequential phases/stages consisting of subtasks that are conquered in an ordered, parallel, and bottleneck-free way. In two illustrating examples, we test the performance of our schemas on merging multiple VCF files into either a single TPED or a single VCF file, which are benchmarked with the traditional single/parallel multiway-merge methods, message passing interface (MPI)–based high-performance computing (HPC) implementation, and the popular VCFTools. Conclusions Our experiments suggest all three schemas either deliver a significant improvement in efficiency or render much better strong and weak scalabilities over traditional methods. Our findings provide generalized scalable schemas for performing sorted merging on genetics and genomics data using these Apache distributed systems. PMID:29762754
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jones, D.W.
This paper identifies a number of developments which are prominent during the urbanization of a country and which have particularly strong implications for energy use. Concomitant with urbanization, the industrial composition of the economy's production shifts, with reductions in agriculture and increases in the importance of primary metals, chemicals, and cement, all of which are relatively energy-intensive sectors. Evidence from India indicates that the movement of a worker from agriculture to the least energy-intensive urban activity other than services will quadruple per worker production energy requirements. Next, population concentration associated with urbanization facilitates increases in the scale of production whichmore » in turn encourages the substitution of modern energy for traditional fuels and requires energy for longer deliveries. Also, concentrated, off-farm populations require processing and delivery of food, which are not required for largely agricultural countries. Domestic activity changes send activities which were formerly conducted in the household with little or no energy use, outside, usually into firms, where fuels are used. Urban households also use considerably more transportation than do rural households. Evidence from Hong Kong indicates that pure urban density increases encourage substitutions of modern energy for traditional fuels. Finally, increased real incomes associated with urbanization increase energy consumption, with an elasticity of roughly unity. Aggregate cross-sectional data evidence from sixty developing countries was used to examine the overall magnitude of the effects of urbanization and associated developmental changes on per capita energy use. Controlling for industrial structure, per capita income (per capita gross domestic product), and several other variables, a one-percent increase in urbanization will cause a one-half percent increase in per capita energy use. 81 refs., 5 figs., 63 tabs.« less
NASA Astrophysics Data System (ADS)
Neumann, Jay; Parlato, Russell; Tracy, Gregory; Randolph, Max
2015-09-01
Focal plane alignment for large format arrays and faster optical systems require enhanced precision methodology and stability over temperature. The increase in focal plane array size continues to drive the alignment capability. Depending on the optical system, the focal plane flatness of less than 25μm (.001") is required over transition temperatures from ambient to cooled operating temperatures. The focal plane flatness requirement must also be maintained in airborne or launch vibration environments. This paper addresses the challenge of the detector integration into the focal plane module and housing assemblies, the methodology to reduce error terms during integration and the evaluation of thermal effects. The driving factors influencing the alignment accuracy include: datum transfers, material effects over temperature, alignment stability over test, adjustment precision and traceability to NIST standard. The FPA module design and alignment methodology reduces the error terms by minimizing the measurement transfers to the housing. In the design, the proper material selection requires matched coefficient of expansion materials minimizes both the physical shift over temperature as well as lowering the stress induced into the detector. When required, the co-registration of focal planes and filters can achieve submicron relative positioning by applying precision equipment, interferometry and piezoelectric positioning stages. All measurements and characterizations maintain traceability to NIST standards. The metrology characterizes the equipment's accuracy, repeatability and precision of the measurements.
A Nationwide Overview of Sight-Singing Requirements of Large-Group Choral Festivals
ERIC Educational Resources Information Center
Norris, Charles E.
2004-01-01
The purpose of this study was to examine sight-singing requirements at junior and senior high school large-group ratings-based choral festivals throughout the United States. Responses to the following questions were sought from each state: (1) Are there ratings-based large-group choral festivals? (2) Is sight-singing a requirement? (3) Are there…
Climate Golden Age or Greenhouse Gas Dark Age Legacy?
NASA Astrophysics Data System (ADS)
Carter, P.
2016-12-01
Relying on the IPCC Assessments, this paper assesses legacy from total committed global warming over centuries, correlated with comprehensive projected impacts. Socio-economic inertia, climate system inertia, atmospheric greenhouse gas (GHG) concentrations, amplifying feedback emissions, and unmasking of cooling aerosols are determinants. Stabilization of global temperature (and ocean acidification for CO2) requires emissions of "long lived greenhouse gases" to be "about zero," including feedbacks. "The feedback … is positive" this century; many large feedback sources tend to be self- and inter-reinforcing. Only timely total conversion of all fossil fuel power to clean, virtually zero-carbon renewable power can achieve virtual zero carbon emissions. This results in multiple, increasing benefits for the entire world population of today's and all future generations, as laid out here. Conversions of methane- and nitrous oxide-emitting sources have large benefits. Without timely conversion to virtual zero emissions, the global climate and ocean disruptions are predicted to become progressively more severe and practically irreversible. "Continued emission of greenhouse gases will increase the likelihood of severe, pervasive and irreversible impacts for people and ecosystems." Crop yields in all main food-producing regions are projected to decline progressively with rising temperature (as proxy to multiple adverse effects) (AR5). Ocean heating, acidification, and de-oxygenation are projected to increase under all scenarios, as is species extinction. The legacy for humanity depends on reducing long-lived global emissions fast enough to virtual zero. Today's surface warming with unprecedented and accelerating atmospheric GHG concentrations requires an immediate response. The only IPCC scenario to possibly meet this and not exceed 2ºC by and after 2100 is the best-case RCP2.6, which requires CO2 eq. emissions to peak right away and decline at the latest by 2020.
Houdelet, Marcel; Galinski, Anna; Holland, Tanja; Wenzel, Kathrin; Schillberg, Stefan; Buyel, Johannes Felix
2017-04-01
Transient expression systems allow the rapid production of recombinant proteins in plants. Such systems can be scaled up to several hundred kilograms of biomass, making them suitable for the production of pharmaceutical proteins required at short notice, such as emergency vaccines. However, large-scale transient expression requires the production of recombinant Agrobacterium tumefaciens strains with the capacity for efficient gene transfer to plant cells. The complex media often used for the cultivation of this species typically include animal-derived ingredients that can contain human pathogens, thus conflicting with the requirements of good manufacturing practice (GMP). We replaced all the animal-derived components in yeast extract broth (YEB) cultivation medium with soybean peptone, and then used a design-of-experiments approach to optimize the medium composition, increasing the biomass yield while maintaining high levels of transient expression in subsequent infiltration experiments. The resulting plant peptone Agrobacterium medium (PAM) achieved a two-fold increase in OD 600 compared to YEB medium during a 4-L batch fermentation lasting 18 h. Furthermore, the yields of the monoclonal antibody 2G12 and the fluorescent protein DsRed were maintained when the cells were cultivated in PAM rather than YEB. We have thus demonstrated a simple, efficient and scalable method for medium optimization that reduces process time and costs. The final optimized medium for the cultivation of A. tumefaciens completely lacks animal-derived components, thus facilitating the GMP-compliant large-scale transient expression of recombinant proteins in plants. © 2017 The Authors. Biotechnology Journal published by WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Facilitating large-scale clinical trials: in Asia.
Choi, Han Yong; Ko, Jae-Wook
2010-01-01
The number of clinical trials conducted in Asian countries has started to increase as a result of expansion of the pharmaceutical market in this area. There is a growing opportunity for large-scale clinical trials because of the large number of patients, significant market potential, good quality of data, and the cost effective and qualified medical infrastructure. However, for carrying out large-scale clinical trials in Asia, there are several major challenges, including the quality control of data, budget control, laboratory validation, monitoring capacity, authorship, staff training, and nonstandard treatment that need to be considered. There are also several difficulties in collaborating on international trials in Asia because Asia is an extremely diverse continent. The major challenges are language differences, diversity of patterns of disease, and current treatments, a large gap in the experience with performing multinational trials, and regulatory differences among the Asian countries. In addition, there are also differences in the understanding of global clinical trials, medical facilities, indemnity assurance, and culture, including food and religion. To make regional and local data provide evidence for efficacy through the standardization of these differences, unlimited effort is required. At this time, there are no large clinical trials led by urologists in Asia, but it is anticipated that the role of urologists in clinical trials will continue to increase. Copyright © 2010 Elsevier Inc. All rights reserved.
Influence of forces acting on side of machine on precision machining of large diameter holes
NASA Astrophysics Data System (ADS)
Fedorenko, M. A.; Bondarenko, J. A.; Sanina, T. M.
2018-03-01
One of the most important factors that increase efficiency, durability and reliability of rotating units is precision installation, preventive maintenance work, timely replacing of a failed or worn components and assemblies. These works should be carried out in the operation of the equipment, as the downtime in many cases leads to large financial losses. Stop of one unit of an industrial enterprise can interrupt the technological chain of production, resulting in a possible stop of the entire equipment. Improving the efficiency and optimization of the repair process increases accuracy of installation work when installing equipment, conducting restoration under operating conditions relevant for enterprises of different industries because it eliminates dismantling the equipment, sending it to maintenance, the expectation of equipment return, the new installation with the required quality and accuracy of repair.
NASA Astrophysics Data System (ADS)
Kino, Hisashi; Fukushima, Takafumi; Tanaka, Tetsu
2018-04-01
Charge-trapping memory requires the increase of bit density per cell and a larger memory window for lower-power operation. A tunnel field-effect transistor (TFET) can achieve to increase the bit density per cell owing to its steep subthreshold slope. In addition, a TFET structure has an asymmetric structure, which is promising for achieving a larger memory window. A TFET with the N-type gate shows a higher electric field between the P-type source and the N-type gate edge than the conventional FET structure. This high electric field enables large amounts of charges to be injected into the charge storage layer. In this study, we fabricated silicon-oxide-nitride-oxide-semiconductor (SONOS) memory devices with the TFET structure and observed a steep subthreshold slope and a larger memory window.
Nutritional Biochemistry of Spaceflight
NASA Technical Reports Server (NTRS)
Smith, Scott M.
2000-01-01
Adequate nutrition is critical for crew health and safety during spaceflight. To ensure adequate nutrition, the nutrient requirements need to be both accurate and available from the spaceflight food system. The existing nutritional requirements for extended-duration spaceflight have been defined largely by extrapolation from ground-based research. However, nutritional requirements are influenced by most of the physiological consequences of spaceflight, including loss of lean, adipose, and bone tissue; changes in blood composition; and increased risk of renal stone formation. This review focuses on key areas where information has been gained in recent years: dietary intake and energy metabolism, bone health, fluid and electrolyte homeostasis, and hematological changes. Areas in which specific nutrients have the potential to serve as countermeasures to the negative effects of spaceflight are also reviewed. Dietary Intake
Atmospheric rendezvous feasibility study
NASA Technical Reports Server (NTRS)
Schaezler, A. D.
1972-01-01
A study was carried out to determine the feasibility of using atmospheric rendezvous to increase the efficiency of space transportation and to determine the most effective implementation. It is concluded that atmospheric rendezvous is feasible and can be utilized in a space transportation system to reduce size of the orbiter vehicle, provide a powered landing with go-around capability for every mission, and achieve lateral range performance that exceeds requirements. A significantly lighter booster and reduced launch fuel requirements are additional benefits that can be realized with a system that includes a large subsonic airplane for recovery of the orbiter. Additional reduction in booster size is possible if the airplane is designed for recovery of the booster by towing. An airplane about the size of the C-5A is required.
A framework for improving resident research participation and scholarly output.
Manring, M M; Panzo, Julia A; Mayerson, Joel L
2014-01-01
The Accreditation Council for Graduate Medical Education requires that "faculty should encourage and support residents in scholarly activities." There are no guidelines, however, to illustrate how this should be done, and only a small number of published reports offer examples of successful efforts to spur resident research. We sought to improve our residents' participation in scholarly activities. We describe a multifaceted program to quickly build resident scholarship at an orthopaedic department. Large academic medical center in the Midwestern United States. An experienced medical editor was recruited to assist faculty and mentor residents in coordinating research projects and to direct publishing activity. Additional publishing requirements were added to the resident curriculum beyond those already required by the Accreditation Council for Graduate Medical Education. Residents were required to select a faculty research mentor to guide all research projects toward a manuscript suitable for submission to a peer-reviewed journal. Activities were monitored by the editor and the resident coordinator. Over 4 years, total department peer-reviewed publications increased from 33 to 163 annually. Despite a decrease in resident complement, the number of peer-reviewed publications with a resident author increased from 6 in 2009 to 53 in 2012. The addition of an experienced medical editor, changes in program requirements, and an increased commitment to promotion of resident research across the faculty led to a dramatic increase in resident publications. Our changes may be a model for other programs that have the financial resources and faculty commitment necessary to achieve a rapid turnaround. Copyright © 2014 Association of Program Directors in Surgery. Published by Elsevier Inc. All rights reserved.
Common modular avionics - Partitioning and design philosophy
NASA Astrophysics Data System (ADS)
Scott, D. M.; Mulvaney, S. P.
The design objectives and definition criteria for common modular hardware that will perform digital processing functions in multiple avionic subsystems are examined. In particular, attention is given to weapon system-level objectives, such as increased supportability, reduced life cycle costs, and increased upgradability. These objectives dictate the following overall modular design goals: reduce test equipment requirements; have a large number of subsystem applications; design for architectural growth; and standardize for technology transparent implementations. Finally, specific partitioning criteria are derived on the basis of the weapon system-level objectives and overall design goals.
Environmental microbiology as related to planetary quarantine
NASA Technical Reports Server (NTRS)
Iflug, I. J.
1971-01-01
The results of studies to determine the effect of soil particle size on the survival time at 125 C of the microflora associated with these particles are discussed. The data suggest that longer survival times exist for the microflora associated with larger particles. The studies indicate that microorganisms associated with soil are difficult to kill and that organisms associated with large particles are harder to kill than those associated with small particles. Sterlization requirements increase as the level of contamination increases. Soil particles and their accompanying microflora are the most critical contaminants.
Technical fixes and Climate Change: Optimizing for Risks and Consequences
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rasch, Philip J.
2010-09-16
Scientists and society in general are becoming increasingly concerned about the risks of climate change from the emission of greenhouse gases [IPCC, 2007]. Yet emissions continue to increase [Raupach et al., 2007], and reductions soon enough to avoid large and undesirable impacts requires a near revolutionary global transformation of energy and transportation systems [Hoffert et al., 1998]. The size of the transformation and lack of an effective societal response has motivated some to explore other quite controversial strategies to mitigate some of the planetary consequences of these emissions.
Clustangles: An Open Library for Clustering Angular Data.
Sargsyan, Karen; Hua, Yun Hao; Lim, Carmay
2015-08-24
Dihedral angles are good descriptors of the numerous conformations visited by large, flexible systems, but their analysis requires directional statistics. A single package including the various multivariate statistical methods for angular data that accounts for the distinct topology of such data does not exist. Here, we present a lightweight standalone, operating-system independent package called Clustangles to fill this gap. Clustangles will be useful in analyzing the ever-increasing number of structures in the Protein Data Bank and clustering the copious conformations from increasingly long molecular dynamics simulations.
Preliminary Assessment of Microwave Readout Multiplexing Factor
DOE Office of Scientific and Technical Information (OSTI.GOV)
Croce, Mark Philip; Koehler, Katrina Elizabeth; Rabin, Michael W.
2017-01-23
Ultra-high resolution microcalorimeter gamma spectroscopy is a new non-destructive assay technology for measurement of plutonium isotopic composition, with the potential to reduce total measurement uncertainty to a level competitive with destructive analysis methods [1-4]. Achieving this level of performance in practical applications requires not only the energy resolution now routinely achieved with transition-edge sensor microcalorimeter arrays (an order of magnitude better than for germanium detectors) but also high throughput. Microcalorimeter gamma spectrometers have not yet achieved detection efficiency and count rate capability that is comparable to germanium detectors, largely because of limits from existing readout technology. Microcalorimeter detectors must bemore » operated at low temperature to achieve their exceptional energy resolution. Although the typical 100 mK operating temperatures can be achieved with reliable, cryogen-free systems, the cryogenic complexity and heat load from individual readout channels for large sensor arrays is prohibitive. Multiplexing is required for practical systems. The most mature multiplexing technology at present is time-division multiplexing (TDM) [3, 5-6]. In TDM, the sensor outputs are switched by applying bias current to one SQUID amplifier at a time. Transition-edge sensor (TES) microcalorimeter arrays as large as 256 pixels have been developed for X-ray and gamma-ray spectroscopy using TDM technology. Due to bandwidth limits and noise scaling, TDM is limited to a maximum multiplexing factor of approximately 32-40 sensors on one readout line [8]. Increasing the size of microcalorimeter arrays above the kilopixel scale, required to match the throughput of germanium detectors, requires the development of a new readout technology with a much higher multiplexing factor.« less
Carbone, Chris; Teacher, Amber; Rowcliffe, J. Marcus
2007-01-01
Mammalian carnivores fall into two broad dietary groups: smaller carnivores (<20 kg) that feed on very small prey (invertebrates and small vertebrates) and larger carnivores (>20 kg) that specialize in feeding on large vertebrates. We develop a model that predicts the mass-related energy budgets and limits of carnivore size within these groups. We show that the transition from small to large prey can be predicted by the maximization of net energy gain; larger carnivores achieve a higher net gain rate by concentrating on large prey. However, because it requires more energy to pursue and subdue large prey, this leads to a 2-fold step increase in energy expenditure, as well as increased intake. Across all species, energy expenditure and intake both follow a three-fourths scaling with body mass. However, when each dietary group is considered individually they both display a shallower scaling. This suggests that carnivores at the upper limits of each group are constrained by intake and adopt energy conserving strategies to counter this. Given predictions of expenditure and estimates of intake, we predict a maximum carnivore mass of approximately a ton, consistent with the largest extinct species. Our approach provides a framework for understanding carnivore energetics, size, and extinction dynamics. PMID:17227145
Cowan, Natasha; Goldenberg, Alina; Basu, Pallavi; Eilers, Robert; Hau, Jennifer; I Brian Jiang, Shang
2018-05-01
Clinically large cutaneous tumors and those with aggressive subclinical extension (ASE) often require wider margins and increased operative time during Mohs micrographic surgery (MMS). Our goal is to improve dermatologic surgeons' counseling information on complication risks for aggressive tumors. To examine the incidence of postoperative complications in MMS patients, with a focus on differences between aggressive and non-aggressive tumors. We performed a retrospective cross-sectional chart review of 4151 MMS cases at the University of California, San Diego. A postoperative complication was defined as an adverse event directly related to MMS reported within 6 weeks of the procedure. Clinically, large tumors had 50 times the odds of postoperative complication as compared to all other tumors (P less than 0.001). ASE was not found to be significantly associated with higher rates of postoperative complications when controlled for other factors. Clinically, large tumors may be at higher risk for complications following MMS due to their increased size and need for repair with methods other than linear closures. Tumors with ASE were not found to be at higher risk for postoperative complications. J Drugs Dermatol. 2018;17(5):511-515.
NASA Astrophysics Data System (ADS)
Blackman, Eric G.; Subramanian, Kandaswamy
2013-02-01
The extent to which large-scale magnetic fields are susceptible to turbulent diffusion is important for interpreting the need for in situ large-scale dynamos in astrophysics and for observationally inferring field strengths compared to kinetic energy. By solving coupled evolution equations for magnetic energy and magnetic helicity in a system initialized with isotropic turbulence and an arbitrarily helical large-scale field, we quantify the decay rate of the latter for a bounded or periodic system. The magnetic energy associated with the non-helical large-scale field decays at least as fast as the kinematically estimated turbulent diffusion rate, but the decay rate of the helical part depends on whether the ratio of its magnetic energy to the turbulent kinetic energy exceeds a critical value given by M1, c = (k1/k2)2, where k1 and k2 are the wavenumbers of the large and forcing scales. Turbulently diffusing helical fields to small scales while conserving magnetic helicity requires a rapid increase in total magnetic energy. As such, only when the helical field is subcritical can it so diffuse. When supercritical, it decays slowly, at a rate determined by microphysical dissipation even in the presence of macroscopic turbulence. In effect, turbulent diffusion of such a large-scale helical field produces small-scale helicity whose amplification abates further turbulent diffusion. Two curious implications are that (1) standard arguments supporting the need for in situ large-scale dynamos based on the otherwise rapid turbulent diffusion of large-scale fields require re-thinking since only the large-scale non-helical field is so diffused in a closed system. Boundary terms could however provide potential pathways for rapid change of the large-scale helical field. (2) Since M1, c ≪ 1 for k1 ≪ k2, the presence of long-lived ordered large-scale helical fields as in extragalactic jets do not guarantee that the magnetic field dominates the kinetic energy.
Data Intensive Analysis of Biomolecular Simulations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Straatsma, TP; Soares, Thereza A.
2007-12-01
The advances in biomolecular modeling and simulation made possible by the availability of increasingly powerful high performance computing resources is extending molecular simulations to biological more relevant system size and time scales. At the same time, advances in simulation methodologies are allowing more complex processes to be described more accurately. These developments make a systems approach to computational structural biology feasible, but this will require a focused emphasis on the comparative analysis of the increasing number of molecular simulations that are being carried out for biomolecular systems with more realistic models, multi-component environments, and for longer simulation times. Just asmore » in the case of the analysis of the large data sources created by the new high-throughput experimental technologies, biomolecular computer simulations contribute to the progress in biology through comparative analysis. The continuing increase in available protein structures allows the comparative analysis of the role of structure and conformational flexibility in protein function, and is the foundation of the discipline of structural bioinformatics. This creates the opportunity to derive general findings from the comparative analysis of molecular dynamics simulations of a wide range of proteins, protein-protein complexes and other complex biological systems. Because of the importance of protein conformational dynamics for protein function, it is essential that the analysis of molecular trajectories is carried out using a novel, more integrative and systematic approach. We are developing a much needed rigorous computer science based framework for the efficient analysis of the increasingly large data sets resulting from molecular simulations. Such a suite of capabilities will also provide the required tools for access and analysis of a distributed library of generated trajectories. Our research is focusing on the following areas: (1) the development of an efficient analysis framework for very large scale trajectories on massively parallel architectures, (2) the development of novel methodologies that allow automated detection of events in these very large data sets, and (3) the efficient comparative analysis of multiple trajectories. The goal of the presented work is the development of new algorithms that will allow biomolecular simulation studies to become an integral tool to address the challenges of post-genomic biological research. The strategy to deliver the required data intensive computing applications that can effectively deal with the volume of simulation data that will become available is based on taking advantage of the capabilities offered by the use of large globally addressable memory architectures. The first requirement is the design of a flexible underlying data structure for single large trajectories that will form an adaptable framework for a wide range of analysis capabilities. The typical approach to trajectory analysis is to sequentially process trajectories time frame by time frame. This is the implementation found in molecular simulation codes such as NWChem, and has been designed in this way to be able to run on workstation computers and other architectures with an aggregate amount of memory that would not allow entire trajectories to be held in core. The consequence of this approach is an I/O dominated solution that scales very poorly on parallel machines. We are currently using an approach of developing tools specifically intended for use on large scale machines with sufficient main memory that entire trajectories can be held in core. This greatly reduces the cost of I/O as trajectories are read only once during the analysis. In our current Data Intensive Analysis (DIANA) implementation, each processor determines and skips to the entry within the trajectory that typically will be available in multiple files and independently from all other processors read the appropriate frames.« less
Data Intensive Analysis of Biomolecular Simulations
DOE Office of Scientific and Technical Information (OSTI.GOV)
Straatsma, TP
2008-03-01
The advances in biomolecular modeling and simulation made possible by the availability of increasingly powerful high performance computing resources is extending molecular simulations to biological more relevant system size and time scales. At the same time, advances in simulation methodologies are allowing more complex processes to be described more accurately. These developments make a systems approach to computational structural biology feasible, but this will require a focused emphasis on the comparative analysis of the increasing number of molecular simulations that are being carried out for biomolecular systems with more realistic models, multi-component environments, and for longer simulation times. Just asmore » in the case of the analysis of the large data sources created by the new high-throughput experimental technologies, biomolecular computer simulations contribute to the progress in biology through comparative analysis. The continuing increase in available protein structures allows the comparative analysis of the role of structure and conformational flexibility in protein function, and is the foundation of the discipline of structural bioinformatics. This creates the opportunity to derive general findings from the comparative analysis of molecular dynamics simulations of a wide range of proteins, protein-protein complexes and other complex biological systems. Because of the importance of protein conformational dynamics for protein function, it is essential that the analysis of molecular trajectories is carried out using a novel, more integrative and systematic approach. We are developing a much needed rigorous computer science based framework for the efficient analysis of the increasingly large data sets resulting from molecular simulations. Such a suite of capabilities will also provide the required tools for access and analysis of a distributed library of generated trajectories. Our research is focusing on the following areas: (1) the development of an efficient analysis framework for very large scale trajectories on massively parallel architectures, (2) the development of novel methodologies that allow automated detection of events in these very large data sets, and (3) the efficient comparative analysis of multiple trajectories. The goal of the presented work is the development of new algorithms that will allow biomolecular simulation studies to become an integral tool to address the challenges of post-genomic biological research. The strategy to deliver the required data intensive computing applications that can effectively deal with the volume of simulation data that will become available is based on taking advantage of the capabilities offered by the use of large globally addressable memory architectures. The first requirement is the design of a flexible underlying data structure for single large trajectories that will form an adaptable framework for a wide range of analysis capabilities. The typical approach to trajectory analysis is to sequentially process trajectories time frame by time frame. This is the implementation found in molecular simulation codes such as NWChem, and has been designed in this way to be able to run on workstation computers and other architectures with an aggregate amount of memory that would not allow entire trajectories to be held in core. The consequence of this approach is an I/O dominated solution that scales very poorly on parallel machines. We are currently using an approach of developing tools specifically intended for use on large scale machines with sufficient main memory that entire trajectories can be held in core. This greatly reduces the cost of I/O as trajectories are read only once during the analysis. In our current Data Intensive Analysis (DIANA) implementation, each processor determines and skips to the entry within the trajectory that typically will be available in multiple files and independently from all other processors read the appropriate frames.« less
Federal Register 2010, 2011, 2012, 2013, 2014
2012-03-12
... Collection; Comment Request; Gear-Marking Requirement for Atlantic Large Whale Take Reduction Plan AGENCY... of large whales, especially right whales, due to incidental entanglement in the United States (U.S... Large Whale Take Reduction Plan (ALWTRP), developed under the authority of the Marine Mammal Protection...
Challenges in engineering large customized bone constructs.
Forrestal, David P; Klein, Travis J; Woodruff, Maria A
2017-06-01
The ability to treat large tissue defects with customized, patient-specific scaffolds is one of the most exciting applications in the tissue engineering field. While an increasing number of modestly sized tissue engineering solutions are making the transition to clinical use, successfully scaling up to large scaffolds with customized geometry is proving to be a considerable challenge. Managing often conflicting requirements of cell placement, structural integrity, and a hydrodynamic environment supportive of cell culture throughout the entire thickness of the scaffold has driven the continued development of many techniques used in the production, culturing, and characterization of these scaffolds. This review explores a range of technologies and methods relevant to the design and manufacture of large, anatomically accurate tissue-engineered scaffolds with a focus on the interaction of manufactured scaffolds with the dynamic tissue culture fluid environment. Biotechnol. Bioeng. 2017;114: 1129-1139. © 2016 Wiley Periodicals, Inc. © 2016 Wiley Periodicals, Inc.
Template Interfaces for Agile Parallel Data-Intensive Science
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ramakrishnan, Lavanya; Gunter, Daniel; Pastorello, Gilerto Z.
Tigres provides a programming library to compose and execute large-scale data-intensive scientific workflows from desktops to supercomputers. DOE User Facilities and large science collaborations are increasingly generating large enough data sets that it is no longer practical to download them to a desktop to operate on them. They are instead stored at centralized compute and storage resources such as high performance computing (HPC) centers. Analysis of this data requires an ability to run on these facilities, but with current technologies, scaling an analysis to an HPC center and to a large data set is difficult even for experts. Tigres ismore » addressing the challenge of enabling collaborative analysis of DOE Science data through a new concept of reusable "templates" that enable scientists to easily compose, run and manage collaborative computational tasks. These templates define common computation patterns used in analyzing a data set.« less
Response of high-rise and base-isolated buildings to a hypothetical M w 7.0 blind thrust earthquake
Heaton, T.H.; Hall, J.F.; Wald, D.J.; Halling, M.W.
1995-01-01
High-rise flexible-frame buildings are commonly considered to be resistant to shaking from the largest earthquakes. In addition, base isolation has become increasingly popular for critical buildings that should still function after an earthquake. How will these two types of buildings perform if a large earthquake occurs beneath a metropolitan area? To answer this question, we simulated the near-source ground motions of a Mw 7.0 thrust earthquake and then mathematically modeled the response of a 20-story steel-frame building and a 3-story base-isolated building. The synthesized ground motions were characterized by large displacement pulses (up to 2 meters) and large ground velocities. These ground motions caused large deformation and possible collapse of the frame building, and they required exceptional measures in the design of the base-isolated building if it was to remain functional.
Dodge, Hiroko H.; Zhu, Jian; Woltjer, Randy; Nelson, Peter T.; Bennett, David A.; Cairns, Nigel J.; Fardo, David W.; Kaye, Jeffrey A.; Lyons, Deniz-Erten; Mattek, Nora; Schneider, Julie A; Silbert, Lisa C.; Xiong, Chengjie; Yu, Lei; Schmitt, Frederick A.; Kryscio, Richard J.; Abner, Erin L.
2016-01-01
Introduction Presence of cerebrovascular pathology may increase the risk of clinical diagnosis of AD. Methods We examined excess risk of incident clinical diagnosis of AD (probable and possible AD) posed by the presence of lacunes and large infarcts beyond AD pathology using data from the Statistical Modelling of Aging and Risk of Transition (SMART) study, a consortium of longitudinal cohort studies with over 2000 autopsies. We created six mutually exclusive pathology patterns combining three levels of AD pathology (low, moderate or high AD pathology) and two levels of vascular pathology (without lacunes and large infarcts or with lacunes and/or large infarcts). Results The coexistence of lacunes and large infarcts results in higher likelihood of clinical diagnosis of AD only when AD pathology burden is low. Discussion Our results reinforce the diagnostic importance of AD pathology in clinical AD. Further harmonization of assessment approaches for vascular pathologies is required. PMID:28017827
ERIC Educational Resources Information Center
Walker, Alicia; Bush, Amy; Sanchagrin, Ken; Holland, Jonathon
2017-01-01
This study examined possible ways to increase student engagement in small sections of a large, introductory-level, required university course. Research shows that cooperative group learning boosts achievement through fostering better interpersonal relationships between students. Cooperative group learning is an evidence-based instructional…
Contamination Study of Micro Pulsed Plasma Thruster
2008-03-01
missions ranging from nano -satellites to large spacecrafts requiring precision placement1. As an unfortunate side effect, the exhaust plume induces...the electrodes9. The electrodes have often been arranged coaxially . The system consists of electrodes, capacitors as energy storage units, a spark...increasing interest in the so-called micro- and nano -satellites, which are highly maneuverable and have lower cost. These small satellites are aimed to
ERIC Educational Resources Information Center
Pufpaff, Lisa A.; Mcintosh, Constance E.; Thomas, Cynthia; Elam, Megan; Irwin, Mary Kay
2015-01-01
The number of students with special healthcare needs (SHCN) and severe disabilities in public schools in the United States has steadily increased in recent years, largely due to the changing landscape of public health relative to advances in medicine and medical technology. The specialized care required for these students often necessitates…
Development of Traction Drive Motors for the Toyota Hybrid System
NASA Astrophysics Data System (ADS)
Kamiya, Munehiro
Toyota Motor Corporation developed in 2005 a new hybrid system for a large SUV. This system included the new development of a high-speed traction drive motor achieving a significant increase in power weight ratio. This paper provides an overview of the hybrid system, discusses the characteristics required of a traction drive motor, and presents the technologies employed in the developed motor.
ERIC Educational Resources Information Center
Metzler, Carol W.; Sanders, Matthew R.; Rusby, Julie C.; Crowley, Ryann N.
2012-01-01
Within a public health approach to improving parenting, the mass media offer a potentially more efficient and affordable format for directly reaching a large number of parents with evidence-based parenting information than do traditional approaches to parenting interventions that require delivery by a practitioner. Little is known, however, about…
Christine E. Turner; William H. Romme; Jim Chew; Mark E. Miller; Lisa Floyd-Hanna Leavesley; George San Miguel; Neil Cobb; Richard Zirbes; Roland Viger; Kirsten Ironside
2008-01-01
The rapid pace of social and environmental changes over the past few decades has presented significant challenges for the people who manage our public lands. Demographic shifts have placed large populations in close proximity to public lands and have resulted in increased public scrutiny of decision making on those lands. Natural resource managers are required to make...
Effect of Carboxymethylation on the Rheological Properties of Hyaluronan
Wendling, Rian J.; Christensen, Amanda M.; Quast, Arthur D.; Atzet, Sarah K.; Mann, Brenda K.
2016-01-01
Chemical modifications made to hyaluronan to enable covalent crosslinking to form a hydrogel or to attach other molecules may alter the physical properties as well, which have physiological importance. Here we created carboxymethyl hyaluronan (CMHA) with varied degree of modification and investigated the effect on the viscosity of CMHA solutions. Viscosity decreased initially as modification increased, with a minimum viscosity for about 30–40% modification. This was followed by an increase in viscosity around 45–50% modification. The pH of the solution had a variable effect on viscosity, depending on the degree of carboxymethyl modification and buffer. The presence of phosphates in the buffer led to decreased viscosity. We also compared large-scale production lots of CMHA to lab-scale and found that large-scale required extended reaction times to achieve the same degree of modification. Finally, thiolated CMHA was disulfide crosslinked to create hydrogels with increased viscosity and shear-thinning aspects compared to CMHA solutions. PMID:27611817
NASA Technical Reports Server (NTRS)
Matthews, Elaine; Walter, B.; Bogner, J.; Sarma, D.; Portney, B.; Hansen, James (Technical Monitor)
2000-01-01
In situ measurements of atmospheric methane concentrations begun in the early 1980s show decadal trends, as well as large interannual variations, in growth rate. Recent research indicates that while wetlands can explain several of the large growth anomalies for individual years, the decadal trend may be the combined effect of increasing sinks, due to increases in tropospheric OH, and stabilizing sources. We discuss new 20-year histories of annual, global source strengths for all major methane sources, i.e., natural wetlands, rice cultivation, ruminant animals, landfills, fossil fuels, and biomass burning, and present estimates of the temporal pattern of the sink required to reconcile these sources and atmospheric concentrations over the time period. Analysis of the individual emission sources, together with model-derived estimates of the OH sink strength, indicates that the growth rate of atmospheric methane observed over the last 20 years can only be explained by a combination of changes in source emissions and an increasing tropospheric sink.
Tanger, Paul; Klassen, Stephen; Mojica, Julius P.; ...
2017-02-21
In order to ensure food security in the face of population growth, decreasing water and land for agriculture, and increasing climate variability, crop yields must increase faster than the current rates. Increased yields will require implementing novel approaches in genetic discovery and breeding. We demonstrate the potential of field-based high throughput phenotyping (HTP) on a large recombinant population of rice to identify genetic variation underlying important traits. We find that detecting quantitative trait loci (QTL) with HTP phenotyping is as accurate and effective as traditional labor- intensive measures of flowering time, height, biomass, grain yield, and harvest index. Furthermore, geneticmore » mapping in this population, derived from a cross of an modern cultivar (IR64) with a landrace (Aswina), identified four alleles with negative effect on grain yield that are fixed in IR64, demonstrating the potential for HTP of large populations as a strategy for the second green revolution.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tanger, Paul; Klassen, Stephen; Mojica, Julius P.
In order to ensure food security in the face of population growth, decreasing water and land for agriculture, and increasing climate variability, crop yields must increase faster than the current rates. Increased yields will require implementing novel approaches in genetic discovery and breeding. We demonstrate the potential of field-based high throughput phenotyping (HTP) on a large recombinant population of rice to identify genetic variation underlying important traits. We find that detecting quantitative trait loci (QTL) with HTP phenotyping is as accurate and effective as traditional labor- intensive measures of flowering time, height, biomass, grain yield, and harvest index. Furthermore, geneticmore » mapping in this population, derived from a cross of an modern cultivar (IR64) with a landrace (Aswina), identified four alleles with negative effect on grain yield that are fixed in IR64, demonstrating the potential for HTP of large populations as a strategy for the second green revolution.« less
Invisible water, visible impact: groundwater use and Indian agriculture under climate change
Zaveri, Esha; Grogan, Danielle S.; Fisher-Vanden, Karen; ...
2016-08-03
India is one of the world's largest food producers, making the sustainability of its agricultural system of global significance. Groundwater irrigation underpins India's agriculture, currently boosting crop production by enough to feed 170 million people. Groundwater overexploitation has led to drastic declines in groundwater levels, threatening to push this vital resource out of reach for millions of small-scale farmers who are the backbone of India's food security. Historically, losing access to groundwater has decreased agricultural production and increased poverty. We take a multidisciplinary approach to assess climate change challenges facing India's agricultural system, and to assess the effectiveness of large-scalemore » water infrastructure projects designed to meet these challenges. We find that even in areas that experience climate change induced precipitation increases, expansion of irrigated agriculture will require increasing amounts of unsustainable groundwater. Finally, the large proposed national river linking project has limited capacity to alleviate groundwater stress. Thus, without intervention, poverty and food insecurity in rural India is likely to worsen.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sullivan, Lauren, L.; Johnson, Brenda, L.; Brudvig, Lars, A.
Habitat corridors, a common management strategy for increasing connectivity in fragmented landscapes, have experimentally validated positive influences on species movement and diversity. However, long-standing concerns that corridors could negatively impact native species by spreading antagonists, such as disease, remain largely untested. Using a large-scale, replicated experiment, we evaluated whether corridors increase the incidence of plant parasites. We found that corridor impacts varied with parasite dispersal mode. Connectivity provided by corridors increased incidence of biotically dispersed parasites (galls on Solidago odora) but not of abiotically dispersed parasites (foliar fungi on S. odora and three Lespedeza spp.). Both biotically and abiotically dispersedmore » parasites responded to edge effects, but the direction of responses varied across species. Although our results require additional tests for generality to other species and landscapes, they suggest that, when establishing conservation corridors, managers should focus on mitigating two potential negative effects: the indirect effects of narrow corridors in creating edges and direct effects of corridors in enhancing connectivity of biotically dispersed parasites.« less
Invisible water, visible impact: groundwater use and Indian agriculture under climate change
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zaveri, Esha; Grogan, Danielle S.; Fisher-Vanden, Karen
India is one of the world's largest food producers, making the sustainability of its agricultural system of global significance. Groundwater irrigation underpins India's agriculture, currently boosting crop production by enough to feed 170 million people. Groundwater overexploitation has led to drastic declines in groundwater levels, threatening to push this vital resource out of reach for millions of small-scale farmers who are the backbone of India's food security. Historically, losing access to groundwater has decreased agricultural production and increased poverty. We take a multidisciplinary approach to assess climate change challenges facing India's agricultural system, and to assess the effectiveness of large-scalemore » water infrastructure projects designed to meet these challenges. We find that even in areas that experience climate change induced precipitation increases, expansion of irrigated agriculture will require increasing amounts of unsustainable groundwater. Finally, the large proposed national river linking project has limited capacity to alleviate groundwater stress. Thus, without intervention, poverty and food insecurity in rural India is likely to worsen.« less
Can dispersal mode predict corridor effects on plant parasites?
Sullivan, Lauren L; Johnson, Brenda L; Brudvig, Lars A; Haddad, Nick M
2011-08-01
Habitat corridors, a common management strategy for increasing connectivity in fragmented landscapes, have experimentally validated positive influences on species movement and diversity. However, long-standing concerns that corridors could negatively impact native species by spreading antagonists, such as disease, remain largely untested. Using a large-scale, replicated experiment, we evaluated whether corridors increase the incidence of plant parasites. We found that corridor impacts varied with parasite dispersal mode. Connectivity provided by corridors increased incidence of biotically dispersed parasites (galls on Solidago odora) but not of abiotically dispersed parasites (foliar fungi on S. odora and three Lespedeza spp.). Both biotically and abiotically dispersed parasites responded to edge effects, but the direction of responses varied across species. Although our results require additional tests for generality to other species and landscapes, they suggest that, when establishing conservation corridors, managers should focus on mitigating two potential negative effects: the indirect effects of narrow corridors in creating edges and direct effects of corridors in enhancing connectivity of biotically dispersed parasites.
Invisible water, visible impact: groundwater use and Indian agriculture under climate change
NASA Astrophysics Data System (ADS)
Zaveri, Esha; Grogan, Danielle S.; Fisher-Vanden, Karen; Frolking, Steve; Lammers, Richard B.; Wrenn, Douglas H.; Prusevich, Alexander; Nicholas, Robert E.
2016-08-01
India is one of the world’s largest food producers, making the sustainability of its agricultural system of global significance. Groundwater irrigation underpins India’s agriculture, currently boosting crop production by enough to feed 170 million people. Groundwater overexploitation has led to drastic declines in groundwater levels, threatening to push this vital resource out of reach for millions of small-scale farmers who are the backbone of India’s food security. Historically, losing access to groundwater has decreased agricultural production and increased poverty. We take a multidisciplinary approach to assess climate change challenges facing India’s agricultural system, and to assess the effectiveness of large-scale water infrastructure projects designed to meet these challenges. We find that even in areas that experience climate change induced precipitation increases, expansion of irrigated agriculture will require increasing amounts of unsustainable groundwater. The large proposed national river linking project has limited capacity to alleviate groundwater stress. Thus, without intervention, poverty and food insecurity in rural India is likely to worsen.
NASA Technical Reports Server (NTRS)
Handley, Thomas H., Jr.; Preheim, Larry E.
1990-01-01
Data systems requirements in the Earth Observing System (EOS) Space Station Freedom (SSF) eras indicate increasing data volume, increased discipline interplay, higher complexity and broader data integration and interpretation. A response to the needs of the interdisciplinary investigator is proposed, considering the increasing complexity and rising costs of scientific investigation. The EOS Data Information System, conceived to be a widely distributed system with reliable communication links between central processing and the science user community, is described. Details are provided on information architecture, system models, intelligent data management of large complex databases, and standards for archiving ancillary data, using a research library, a laboratory and collaboration services.
Hypogonadism: Therapeutic Risks, Benefits, and Outcomes.
Sigalos, John T; Pastuszak, Alexander W; Khera, Mohit
2018-03-01
Hypogonadism is a common condition defined by the presence of low serum testosterone levels and hypogonadal symptoms, and most commonly treated using testosterone therapy (TTh). The accuracy of diagnosis and appropriateness of treatment, along with proper follow-up, are increasingly important given the large increase in testosterone prescriptions and the recent concern for cardiovascular (CV) risk associated with TTh. In March of 2015, the US Food and Drug Administration required that testosterone product labels disclose a potential CV risk, despite the evidence base for this association being weak and inconclusive. However, TTh may improve CV outcomes rather than increase risks. Copyright © 2017 Elsevier Inc. All rights reserved.
Large area crop inventory experiment crop assessment subsystem software requirements document
NASA Technical Reports Server (NTRS)
1975-01-01
The functional data processing requirements are described for the Crop Assessment Subsystem of the Large Area Crop Inventory Experiment. These requirements are used as a guide for software development and implementation.
Impact of Alternative Processes for Aluminum Production on Energy Requirements
NASA Astrophysics Data System (ADS)
Grjotheim, Kai; Welch, Barry
1981-09-01
Increasing prices and the shortage of large blocks of electrical energy have given greater impetus to the search for viable alternative processes for aluminum production. These include electrolysis of aluminum chloride, sulfide, and nitride; carbothermal reduction of either the ore or alumina; and disproportioning reactions of either aluminum sulfide or the monochloride route. Common to all these processes are the starting material—an ore containing aluminum oxide—and the final product—the metal. Thus, the thermodynamic cycle will invariably dictate similar theoretical energy requirements for the three processes. In practice, however, the achievable efficiencies and, more noticeably, the proportion of electrical to carbothermal energy required for the various stages of operation can vary. The present status of these alternative processes indicates that while alternative routes, such as the Alcoa-AlCl3-Smelting Process, show distinct potential for reducing electrical energy requirements, they offer little chance of reducing overall energy requirements. Furthermore, because of more stringent purity requirements, any gains made may be at the expense of production costs.
Potential climatic impacts and reliability of large-scale offshore wind farms
NASA Astrophysics Data System (ADS)
Wang, Chien; Prinn, Ronald G.
2011-04-01
The vast availability of wind power has fueled substantial interest in this renewable energy source as a potential near-zero greenhouse gas emission technology for meeting future world energy needs while addressing the climate change issue. However, in order to provide even a fraction of the estimated future energy needs, a large-scale deployment of wind turbines (several million) is required. The consequent environmental impacts, and the inherent reliability of such a large-scale usage of intermittent wind power would have to be carefully assessed, in addition to the need to lower the high current unit wind power costs. Our previous study (Wang and Prinn 2010 Atmos. Chem. Phys. 10 2053) using a three-dimensional climate model suggested that a large deployment of wind turbines over land to meet about 10% of predicted world energy needs in 2100 could lead to a significant temperature increase in the lower atmosphere over the installed regions. A global-scale perturbation to the general circulation patterns as well as to the cloud and precipitation distribution was also predicted. In the later study reported here, we conducted a set of six additional model simulations using an improved climate model to further address the potential environmental and intermittency issues of large-scale deployment of offshore wind turbines for differing installation areas and spatial densities. In contrast to the previous land installation results, the offshore wind turbine installations are found to cause a surface cooling over the installed offshore regions. This cooling is due principally to the enhanced latent heat flux from the sea surface to lower atmosphere, driven by an increase in turbulent mixing caused by the wind turbines which was not entirely offset by the concurrent reduction of mean wind kinetic energy. We found that the perturbation of the large-scale deployment of offshore wind turbines to the global climate is relatively small compared to the case of land-based installations. However, the intermittency caused by the significant seasonal wind variations over several major offshore sites is substantial, and demands further options to ensure the reliability of large-scale offshore wind power. The method that we used to simulate the offshore wind turbine effect on the lower atmosphere involved simply increasing the ocean surface drag coefficient. While this method is consistent with several detailed fine-scale simulations of wind turbines, it still needs further study to ensure its validity. New field observations of actual wind turbine arrays are definitely required to provide ultimate validation of the model predictions presented here.
Infrastructure stability surveillance with high resolution InSAR
NASA Astrophysics Data System (ADS)
Balz, Timo; Düring, Ralf
2017-02-01
The construction of new infrastructure in largely unknown and difficult environments, as it is necessary for the construction of the New Silk Road, can lead to a decreased stability along the construction site, leading to an increase in landslide risk and deformation caused by surface motion. This generally requires a thorough pre-analysis and consecutive surveillance of the deformation patterns to ensure the stability and safety of the infrastructure projects. Interferometric SAR (InSAR) and the derived techniques of multi-baseline InSAR are very powerful tools for a large area observation of surface deformation patterns. With InSAR and deriver techniques, the topographic height and the surface motion can be estimated for large areas, making it an ideal tool for supporting the planning, construction, and safety surveillance of new infrastructure elements in remote areas.
Tipping elements in the Arctic marine ecosystem.
Duarte, Carlos M; Agustí, Susana; Wassmann, Paul; Arrieta, Jesús M; Alcaraz, Miquel; Coello, Alexandra; Marbà, Núria; Hendriks, Iris E; Holding, Johnna; García-Zarandona, Iñigo; Kritzberg, Emma; Vaqué, Dolors
2012-02-01
The Arctic marine ecosystem contains multiple elements that present alternative states. The most obvious of which is an Arctic Ocean largely covered by an ice sheet in summer versus one largely devoid of such cover. Ecosystems under pressure typically shift between such alternative states in an abrupt, rather than smooth manner, with the level of forcing required for shifting this status termed threshold or tipping point. Loss of Arctic ice due to anthropogenic climate change is accelerating, with the extent of Arctic sea ice displaying increased variance at present, a leading indicator of the proximity of a possible tipping point. Reduced ice extent is expected, in turn, to trigger a number of additional tipping elements, physical, chemical, and biological, in motion, with potentially large impacts on the Arctic marine ecosystem.
Software Reliability Issues Concerning Large and Safety Critical Software Systems
NASA Technical Reports Server (NTRS)
Kamel, Khaled; Brown, Barbara
1996-01-01
This research was undertaken to provide NASA with a survey of state-of-the-art techniques using in industrial and academia to provide safe, reliable, and maintainable software to drive large systems. Such systems must match the complexity and strict safety requirements of NASA's shuttle system. In particular, the Launch Processing System (LPS) is being considered for replacement. The LPS is responsible for monitoring and commanding the shuttle during test, repair, and launch phases. NASA built this system in the 1970's using mostly hardware techniques to provide for increased reliability, but it did so often using custom-built equipment, which has not been able to keep up with current technologies. This report surveys the major techniques used in industry and academia to ensure reliability in large and critical computer systems.
Disaster and Contingency Planning for Scientific Shared Resource Cores
Wilkerson, Amy
2016-01-01
Progress in biomedical research is largely driven by improvements, innovations, and breakthroughs in technology, accelerating the research process, and an increasingly complex collaboration of both clinical and basic science. This increasing sophistication has driven the need for centralized shared resource cores (“cores”) to serve the scientific community. From a biomedical research enterprise perspective, centralized resource cores are essential to increased scientific, operational, and cost effectiveness; however, the concentration of instrumentation and resources in the cores may render them highly vulnerable to damage from severe weather and other disasters. As such, protection of these assets and the ability to recover from a disaster is increasingly critical to the mission and success of the institution. Therefore, cores should develop and implement both disaster and business continuity plans and be an integral part of the institution’s overall plans. Here we provide an overview of key elements required for core disaster and business continuity plans, guidance, and tools for developing these plans, and real-life lessons learned at a large research institution in the aftermath of Superstorm Sandy. PMID:26848285
NASA Astrophysics Data System (ADS)
Feldmann, Daniel; Bauer, Christian; Wagner, Claus
2018-03-01
We present results from direct numerical simulations (DNS) of turbulent pipe flow at shear Reynolds numbers up to Reτ = 1500 using different computational domains with lengths up to ?. The objectives are to analyse the effect of the finite size of the periodic pipe domain on large flow structures in dependency of Reτ and to assess a minimum ? required for relevant turbulent scales to be captured and a minimum Reτ for very large-scale motions (VLSM) to be analysed. Analysing one-point statistics revealed that the mean velocity profile is invariant for ?. The wall-normal location at which deviations occur in shorter domains changes strongly with increasing Reτ from the near-wall region to the outer layer, where VLSM are believed to live. The root mean square velocity profiles exhibit domain length dependencies for pipes shorter than 14R and 7R depending on Reτ. For all Reτ, the higher-order statistical moments show only weak dependencies and only for the shortest domain considered here. However, the analysis of one- and two-dimensional pre-multiplied energy spectra revealed that even for larger ?, not all physically relevant scales are fully captured, even though the aforementioned statistics are in good agreement with the literature. We found ? to be sufficiently large to capture VLSM-relevant turbulent scales in the considered range of Reτ based on our definition of an integral energy threshold of 10%. The requirement to capture at least 1/10 of the global maximum energy level is justified by a 14% increase of the streamwise turbulence intensity in the outer region between Reτ = 720 and 1500, which can be related to VLSM-relevant length scales. Based on this scaling anomaly, we found Reτ⪆1500 to be a necessary minimum requirement to investigate VLSM-related effects in pipe flow, even though the streamwise energy spectra does not yet indicate sufficient scale separation between the most energetic and the very long motions.
Vibration of a Singly-curved Thin Shell Reflector with a Unidirectional Tension Field
NASA Technical Reports Server (NTRS)
Williams, R. Brett; Klein, Kerry J.; Agnes, Gregory S.
2006-01-01
Increased science requirements for space-based instruments over the past few decades have lead to the increased popularity of deployable space structures constructed from thin, lightweight films. Such structures offer both low mass and the ability to be stowed inside conventional launch vehicles. The analysis in this work pertains to large, singly-curved lightweight deployable reflectors commonly used in radar antennas and space telescopes. These types of systems, which can vary a great deal in size, often have frequency requirement that must be met. This work discusses two missions that utilize this type of aperture technology, and then develops a Rayleigh-Ritz model that predicts the natural frequencies and mode shapes for a (nearly) flat and singly-curved reflector with unidirectional in-plane loading. The results are compared with NASTRAN analyses.
The impact of supercomputers on experimentation: A view from a national laboratory
NASA Technical Reports Server (NTRS)
Peterson, V. L.; Arnold, J. O.
1985-01-01
The relative roles of large scale scientific computers and physical experiments in several science and engineering disciplines are discussed. Increasing dependence on computers is shown to be motivated both by the rapid growth in computer speed and memory, which permits accurate numerical simulation of complex physical phenomena, and by the rapid reduction in the cost of performing a calculation, which makes computation an increasingly attractive complement to experimentation. Computer speed and memory requirements are presented for selected areas of such disciplines as fluid dynamics, aerodynamics, aerothermodynamics, chemistry, atmospheric sciences, astronomy, and astrophysics, together with some examples of the complementary nature of computation and experiment. Finally, the impact of the emerging role of computers in the technical disciplines is discussed in terms of both the requirements for experimentation and the attainment of previously inaccessible information on physical processes.
A database management capability for Ada
NASA Technical Reports Server (NTRS)
Chan, Arvola; Danberg, SY; Fox, Stephen; Landers, Terry; Nori, Anil; Smith, John M.
1986-01-01
The data requirements of mission critical defense systems have been increasing dramatically. Command and control, intelligence, logistics, and even weapons systems are being required to integrate, process, and share ever increasing volumes of information. To meet this need, systems are now being specified that incorporate data base management subsystems for handling storage and retrieval of information. It is expected that a large number of the next generation of mission critical systems will contain embedded data base management systems. Since the use of Ada has been mandated for most of these systems, it is important to address the issues of providing data base management capabilities that can be closely coupled with Ada. A comprehensive distributed data base management project has been investigated. The key deliverables of this project are three closely related prototype systems implemented in Ada. These three systems are discussed.
Student Perceptions of General Education Requirements at a Large Public University: No Surprises?
ERIC Educational Resources Information Center
Thompson, Clarissa A.; Eodice, Michele; Tran, Phuoc
2015-01-01
The current study surveyed students' knowledge of and perceptions about general education requirements at a large research-intensive university. Findings revealed that students harbored misconceptions about general education requirements and illuminated the reasons why students were choosing to take required general education courses at other…
Eyeglass Large Aperture, Lightweight Space Optics FY2000 - FY2002 LDRD Strategic Initiative
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hyde, R
2003-02-10
A series of studies by the Air Force, the National Reconnaissance Office and NASA have identified the critical role played by large optics in fulfilling many of the space related missions of these agencies. Whether it is the Next Generation Space Telescope for NASA, high resolution imaging systems for NRO, or beam weaponry for the Air Force, the diameter of the primary optic is central to achieving high resolution (imaging) or a small spot size on target (lethality). While the detailed requirements differ for each application (high resolution imaging over the visible and near-infrared for earth observation, high damage thresholdmore » but single-wavelength operation for directed energy), the challenges of a large, lightweight primary optic which is space compatible and operates with high efficiency are the same. The advantage of such large optics to national surveillance applications is that it permits these observations to be carried-out with much greater effectiveness than with smaller optics. For laser weapons, the advantage is that it permits more tightly focused beams which can be leveraged into either greater effective range, reduced laser power, and/or smaller on-target spot-sizes; weapon systems can be made either much more effective or much less expensive. This application requires only single-wavelength capability, but places an emphasis upon robust, rapidly targetable optics. The advantages of large aperture optics to astronomy are that it increases the sensitivity and resolution with which we can view the universe. This can be utilized either for general purpose astronomy, allowing us to examine greater numbers of objects in more detail and at greater range, or it can enable the direct detection and detailed examination of extra-solar planets. This application requires large apertures (for both light-gathering and resolution reasons), with broad-band spectral capability, but does not emphasize either large fields-of-view or pointing agility. Despite differences in their requirements and implementations, the fundamental difficulty in utilizing large aperture optics is the same for all of these applications: It is extremely difficult to design large aperture space optics which are both optically precise and can meet the practical requirements for launch and deployment in space. At LLNL we have developed a new concept (Eyeglass) which uses large diffractive optics to solve both of these difficulties; greatly reducing both the mass and the tolerance requirements for large aperture optics. During previous LDRD-supported research, we developed this concept, built and tested broadband diffractive telescopes, and built 50 cm aperture diffraction-limited diffractive lenses (the largest in the world). This work is fully described in UCRL-ID-136262, Eyeglass: A Large Aperture Space Telescope. However, there is a large gap between optical proof-of-principle with sub-meter apertures, and actual 50 meter space telescopes. This gap is far too large (both in financial resources and in spacecraft expertise) to be filled internally at LLNL; implementation of large aperture diffractive space telescopes must be done externally using non-LLNL resources and expertise. While LLNL will never become the primary contractor and integrator for large space optical systems, our natural role is to enable these devices by developing the capability of producing very large diffractive optics. Accordingly, the purpose of the Large Aperture, Lightweight Space Optics Strategic Initiative was to develop the technology to fabricate large, lightweight diffractive lenses. The additional purpose of this Strategic Initiative was, of course, to demonstrate this lens-fabrication capability in a fashion compellingly enough to attract the external support necessary to continue along the path to full-scale space-based telescopes. During this 3 year effort (FY2000-FY2002) we have developed the capability of optically smoothing and diffractively-patterning thin meter-sized sheets of glass into lens panels. We have also developed alignment and seaming techniques which allow individual lens panels to be assembled together, forming a much larger, segmented, diffractive lens. The capabilities provided by this LDRD-supported developmental effort were then demonstrated by the fabrication and testing of a lightweight, 5 meter aperture, diffractive lens.« less
Climate science and famine early warning
Verdin, James P.; Funk, Chris; Senay, Gabriel B.; Choularton, R.
2005-01-01
Food security assessment in sub-Saharan Africa requires simultaneous consideration of multiple socio-economic and environmental variables. Early identification of populations at risk enables timely and appropriate action. Since large and widely dispersed populations depend on rainfed agriculture and pastoralism, climate monitoring and forecasting are important inputs to food security analysis. Satellite rainfall estimates (RFE) fill in gaps in station observations, and serve as input to drought index maps and crop water balance models. Gridded rainfall time-series give historical context, and provide a basis for quantitative interpretation of seasonal precipitation forecasts. RFE are also used to characterize flood hazards, in both simple indices and stream flow models. In the future, many African countries are likely to see negative impacts on subsistence agriculture due to the effects of global warming. Increased climate variability is forecast, with more frequent extreme events. Ethiopia requires special attention. Already facing a food security emergency, troubling persistent dryness has been observed in some areas, associated with a positive trend in Indian Ocean sea surface temperatures. Increased African capacity for rainfall observation, forecasting, data management and modelling applications is urgently needed. Managing climate change and increased climate variability require these fundamental technical capacities if creative coping strategies are to be devised.
Process Performance of Optima XEx Single Wafer High Energy Implanter
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kim, J. H.; Yoon, Jongyoon; Kondratenko, S.
2011-01-07
To meet the process requirements for well formation in future CMOS memory production, high energy implanters require more robust angle, dose, and energy control while maintaining high productivity. The Optima XEx high energy implanter meets these requirements by integrating a traditional LINAC beamline with a robust single wafer handling system. To achieve beam angle control, Optima XEx can control both the horizontal and vertical beam angles to within 0.1 degrees using advanced beam angle measurement and correction. Accurate energy calibration and energy trim functions accelerate process matching by eliminating energy calibration errors. The large volume process chamber and UDC (upstreammore » dose control) using faraday cups outside of the process chamber precisely control implant dose regardless of any chamber pressure increase due to PR (photoresist) outgassing. An optimized RF LINAC accelerator improves reliability and enables singly charged phosphorus and boron energies up to 1200 keV and 1500 keV respectively with higher beam currents. A new single wafer endstation combined with increased beam performance leads to overall increased productivity. We report on the advanced performance of Optima XEx observed during tool installation and volume production at an advanced memory fab.« less
Climate science and famine early warning.
Verdin, James; Funk, Chris; Senay, Gabriel; Choularton, Richard
2005-11-29
Food security assessment in sub-Saharan Africa requires simultaneous consideration of multiple socio-economic and environmental variables. Early identification of populations at risk enables timely and appropriate action. Since large and widely dispersed populations depend on rainfed agriculture and pastoralism, climate monitoring and forecasting are important inputs to food security analysis. Satellite rainfall estimates (RFE) fill in gaps in station observations, and serve as input to drought index maps and crop water balance models. Gridded rainfall time-series give historical context, and provide a basis for quantitative interpretation of seasonal precipitation forecasts. RFE are also used to characterize flood hazards, in both simple indices and stream flow models. In the future, many African countries are likely to see negative impacts on subsistence agriculture due to the effects of global warming. Increased climate variability is forecast, with more frequent extreme events. Ethiopia requires special attention. Already facing a food security emergency, troubling persistent dryness has been observed in some areas, associated with a positive trend in Indian Ocean sea surface temperatures. Increased African capacity for rainfall observation, forecasting, data management and modelling applications is urgently needed. Managing climate change and increased climate variability require these fundamental technical capacities if creative coping strategies are to be devised.
Climate science and famine early warning
Verdin, James; Funk, Chris; Senay, Gabriel; Choularton, Richard
2005-01-01
Food security assessment in sub-Saharan Africa requires simultaneous consideration of multiple socio-economic and environmental variables. Early identification of populations at risk enables timely and appropriate action. Since large and widely dispersed populations depend on rainfed agriculture and pastoralism, climate monitoring and forecasting are important inputs to food security analysis. Satellite rainfall estimates (RFE) fill in gaps in station observations, and serve as input to drought index maps and crop water balance models. Gridded rainfall time-series give historical context, and provide a basis for quantitative interpretation of seasonal precipitation forecasts. RFE are also used to characterize flood hazards, in both simple indices and stream flow models. In the future, many African countries are likely to see negative impacts on subsistence agriculture due to the effects of global warming. Increased climate variability is forecast, with more frequent extreme events. Ethiopia requires special attention. Already facing a food security emergency, troubling persistent dryness has been observed in some areas, associated with a positive trend in Indian Ocean sea surface temperatures. Increased African capacity for rainfall observation, forecasting, data management and modelling applications is urgently needed. Managing climate change and increased climate variability require these fundamental technical capacities if creative coping strategies are to be devised. PMID:16433101
Witt, Cordelie E; Arbabi, Saman; Nathens, Avery B; Vavilala, Monica S; Rivara, Frederick P
2017-04-01
The implications of childhood obesity on pediatric trauma outcomes are not clearly established. Anthropomorphic data were recently added to the National Trauma Data Bank (NTDB) Research Datasets, enabling a large, multicenter evaluation of the effect of obesity on pediatric trauma patients. Children ages 2 to 19years who required hospitalization for traumatic injury were identified in the 2013-2014 NTDB Research Datasets. Age and gender-specific body mass indices (BMI) were calculated. Outcomes included injury patterns, operative procedures, complications, and hospital utilization parameters. Data from 149,817 pediatric patients were analyzed; higher BMI percentiles were associated with significantly more extremity injuries, and fewer injuries to the head, abdomen, thorax and spine (p values <0.001). On multivariable analysis, higher BMI percentiles were associated with significantly increased likelihood of death, deep venous thrombosis, pulmonary embolus and pneumonia; although there was no difference in risk of overall complications. Obese children also had significantly longer lengths of stay and more frequent ventilator requirement. Among children admitted after trauma, increased BMI percentile is associated with increased risk of death and potentially preventable complications. These findings suggest that obese children may require different management than nonobese counterparts to prevent complications. Level III; prognosis study. Copyright © 2017 Elsevier Inc. All rights reserved.
Towards real-time quantitative optical imaging for surgery
NASA Astrophysics Data System (ADS)
Gioux, Sylvain
2017-07-01
There is a pressing clinical need to provide image guidance during surgery. Currently, assessment of tissue that needs to be resected or avoided is performed subjectively leading to a large number of failures, patient morbidity and increased healthcare cost. Because near-infrared (NIR) optical imaging is safe, does not require contact, and can provide relatively deep information (several mm), it offers unparalleled capabilities for providing image guidance during surgery. In this work, we introduce a novel concept that enables the quantitative imaging of endogenous molecular information over large fields-of-view. Because this concept can be implemented in real-time, it is amenable to provide video-rate endogenous information during surgery.
DNA-encoded chemistry: enabling the deeper sampling of chemical space.
Goodnow, Robert A; Dumelin, Christoph E; Keefe, Anthony D
2017-02-01
DNA-encoded chemical library technologies are increasingly being adopted in drug discovery for hit and lead generation. DNA-encoded chemistry enables the exploration of chemical spaces four to five orders of magnitude more deeply than is achievable by traditional high-throughput screening methods. Operation of this technology requires developing a range of capabilities including aqueous synthetic chemistry, building block acquisition, oligonucleotide conjugation, large-scale molecular biological transformations, selection methodologies, PCR, sequencing, sequence data analysis and the analysis of large chemistry spaces. This Review provides an overview of the development and applications of DNA-encoded chemistry, highlighting the challenges and future directions for the use of this technology.
Robotic tape library system level testing at NSA: Present and planned
NASA Technical Reports Server (NTRS)
Shields, Michael F.
1994-01-01
In the present of declining Defense budgets, increased pressure has been placed on the DOD to utilize Commercial Off the Shelf (COTS) solutions to incrementally solve a wide variety of our computer processing requirements. With the rapid growth in processing power, significant expansion of high performance networking, and the increased complexity of applications data sets, the requirement for high performance, large capacity, reliable and secure, and most of all affordable robotic tape storage libraries has greatly increased. Additionally, the migration to a heterogeneous, distributed computing environment has further complicated the problem. With today's open system compute servers approaching yesterday's supercomputer capabilities, the need for affordable, reliable secure Mass Storage Systems (MSS) has taken on an ever increasing importance to our processing center's ability to satisfy operational mission requirements. To that end, NSA has established an in-house capability to acquire, test, and evaluate COTS products. Its goal is to qualify a set of COTS MSS libraries, thereby achieving a modicum of standardization for robotic tape libraries which can satisfy our low, medium, and high performance file and volume serving requirements. In addition, NSA has established relations with other Government Agencies to complete this in-house effort and to maximize our research, testing, and evaluation work. While the preponderance of the effort is focused at the high end of the storage ladder, considerable effort will be extended this year and next at the server class or mid range storage systems.
Shape Tracking of a Dexterous Continuum Manipulator Utilizing Two Large Deflection Shape Sensors
Farvardin, Amirhossein; Grupp, Robert; Murphy, Ryan J.; Taylor, Russell H.; Iordachita, Iulian
2016-01-01
Dexterous continuum manipulators (DCMs) can largely increase the reachable region and steerability for minimally and less invasive surgery. Many such procedures require the DCM to be capable of producing large deflections. The real-time control of the DCM shape requires sensors that accurately detect and report large deflections. We propose a novel, large deflection, shape sensor to track the shape of a 35 mm DCM designed for a less invasive treatment of osteolysis. Two shape sensors, each with three fiber Bragg grating sensing nodes is embedded within the DCM, and the sensors’ distal ends fixed to the DCM. The DCM centerline is computed using the centerlines of each sensor curve. An experimental platform was built and different groups of experiments were carried out, including free bending and three cases of bending with obstacles. For each experiment, the DCM drive cable was pulled with a precise linear slide stage, the DCM centerline was calculated, and a 2D camera image was captured for verification. The reconstructed shape created with the shape sensors is compared with the ground truth generated by executing a 2D–3D registration between the camera image and 3D DCM model. Results show that the distal tip tracking accuracy is 0.40 ± 0.30 mm for the free bending and 0.61 ± 0.15 mm, 0.93 ± 0.05 mm and 0.23 ± 0.10 mm for three cases of bending with obstacles. The data suggest FBG arrays can accurately characterize the shape of large-deflection DCMs. PMID:27761103
Large Space Antenna Systems Technology, 1984
NASA Technical Reports Server (NTRS)
Boyer, W. J. (Compiler)
1985-01-01
Papers are presented which provide a comprehensive review of space missions requiring large antenna systems and of the status of key technologies required to enable these missions. Topic areas include mission applications for large space antenna systems, large space antenna structural systems, materials and structures technology, structural dynamics and control technology, electromagnetics technology, large space antenna systems and the space station, and flight test and evaluation.
Phosphorus: a limiting nutrient for humanity?
Elser, James J
2012-12-01
Phosphorus is a chemical element that is essential to life because of its role in numerous key molecules, including DNA and RNA; indeed, organisms require large amounts of P to grow rapidly. However, the supply of P from the environment is often limiting to production, including to crops. Thus, large amounts of P are mined annually to produce fertilizer that is applied in support of the 'Green Revolution.' However, much of this fertilizer eventually ends up in rivers, lakes and oceans where it causes costly eutrophication. Furthermore, given increasing human population, expanding meat consumption, and proliferating bioenergy pressures, concerns have recently been raised about the long-term geological, economic, and geopolitical viability of mined P for fertilizer production. Together, these issues highlight the non-sustainable nature of current human P use. To achieve P sustainability, farms need to become more efficient in how they use P while society as a whole must develop technologies and practices to recycle P from the food chain. Such large-scale changes will probably require a radical restructuring of the entire food system, highlighting the need for prompt but sustained action. Copyright © 2012 Elsevier Ltd. All rights reserved.
Inadequacy the Health System in Serbia and Corrupt Institutions
Dickov, Veselin
2012-01-01
Rapid changes in the health system require a new trained professionals who fully understand the processes of health and organizational problems and have the knowledge and skills that enable them to manage health care services. Health services to their largely rests on a system of solidarity and “socialism”, and only partly on market principle, and more than in other sectors of the economy requires individuals who are able to bridge that gap. Realize savings in the system that one side is not profitable, on the other hand is able to swallow a huge media arts is that simply needs to learn–just relying on common sense and intuition that no longer helps. The increase in costs. Advances in medicine and technology, and discovery of new drugs, namely, the almost daily increase the costs of diagnosis and treatment. Advances in medicine prolongs life expectancy by increasing the number of patients, especially those with chronic diseases, the biggest consumer of drugs and frequent guests hospital. PMID:23678330
NASA Astrophysics Data System (ADS)
Soja, Amber; Westberg, David; Stackhouse, Paul, Jr.; McRae, Douglas; Jin, Ji-Zhong; Sukhinin, Anatoly
2010-05-01
Fire is the dominant disturbance that precipitates ecosystem change in boreal regions, and fire is largely under the control of weather and climate. Fire frequency, fire severity, area burned and fire season length are predicted to increase in boreal regions under current climate change scenarios. Therefore, changes in fire regimes have the potential to compel ecological change, moving ecosystems more quickly towards equilibrium with a new climate. The ultimate goal of this research is to assess the viability of large-scale (1°) data to be used to define fire weather danger and fire regimes, so that large-scale data can be confidently used to predict future fire regimes using large-scale fire weather data, like that available from current Intergovernmental Panel on Climate Change (IPCC) climate change scenarios. In this talk, we intent to: (1) evaluate Fire Weather Indices (FWI) derived using reanalysis and interpolated station data; (2) discuss the advantages and disadvantages of using these distinct data sources; and (3) highlight established relationships between large-scale fire weather data, area burned, active fires and ecosystems burned. Specifically, the Canadian Forestry Service (CFS) Fire Weather Index (FWI) will be derived using: (1) NASA Goddard Earth Observing System version 4 (GEOS-4) large-scale reanalysis and NASA Global Precipitation Climatology Project (GPCP) data; and National Climatic Data Center (NCDC) surface station-interpolated data. Requirements of the FWI are local noon surface-level air temperature, relative humidity, wind speed, and daily (noon-noon) rainfall. GEOS-4 reanalysis and NCDC station-interpolated fire weather indices are generally consistent spatially, temporally and quantitatively. Additionally, increased fire activity coincides with increased FWI ratings in both data products. Relationships have been established between large-scale FWI to area burned, fire frequency, ecosystem types, and these can be use to estimate historic and future fire regimes.
NASA Astrophysics Data System (ADS)
Ni, Wei-Tou; Han, Sen; Jin, Tao
2016-11-01
With the LIGO announcement of the first direct detection of gravitational waves (GWs), the GW Astronomy was formally ushered into our age. After one-hundred years of theoretical investigation and fifty years of experimental endeavor, this is a historical landmark not just for physics and astronomy, but also for industry and manufacturing. The challenge and opportunity for industry is precision and innovative manufacturing in large size - production of large and homogeneous optical components, optical diagnosis of large components, high reflectance dielectric coating on large mirrors, manufacturing of components for ultrahigh vacuum of large volume, manufacturing of high attenuating vibration isolation system, production of high-power high-stability single-frequency lasers, production of high-resolution positioning systems etc. In this talk, we address the requirements and methods to satisfy these requirements. Optical diagnosis of large optical components requires large phase-shifting interferometer; the 1.06 μm Phase Shifting Interferometer for testing LIGO optics and the recently built 24" phase-shifting Interferometer in Chengdu, China are examples. High quality mirrors are crucial for laser interferometric GW detection, so as for ring laser gyroscope, high precision laser stabilization via optical cavities, quantum optomechanics, cavity quantum electrodynamics and vacuum birefringence measurement. There are stringent requirements on the substrate materials and coating methods. For cryogenic GW interferometer, appropriate coating on sapphire or silicon are required for good thermal and homogeneity properties. Large ultrahigh vacuum components and high attenuating vibration system together with an efficient metrology system are required and will be addressed. For space interferometry, drag-free technology and weak-light manipulation technology are must. Drag-free technology is well-developed. Weak-light phase locking is demonstrated in the laboratories while weak-light manipulation technology still needs developments.
Remote Earth Sciences data collection using ACTS
NASA Technical Reports Server (NTRS)
Evans, Robert H.
1992-01-01
Given the focus on global change and the attendant scope of such research, we anticipate significant growth of requirements for investigator interaction, processing system capabilities, and availability of data sets. The increased complexity of global processes requires interdisciplinary teams to address them; the investigators will need to interact on a regular basis; however, it is unlikely that a single institution will house sufficient investigators with the required breadth of skills. The complexity of the computations may also require resources beyond those located within a single institution; this lack of sufficient computational resources leads to a distributed system located at geographically dispersed institutions. Finally the combination of long term data sets like the Pathfinder datasets and the data to be gathered by new generations of satellites such as SeaWiFS and MODIS-N yield extra-ordinarily large amounts of data. All of these factors combine to increase demands on the communications facilities available; the demands are generating requirements for highly flexible, high capacity networks. We have been examining the applicability of the Advanced Communications Technology Satellite (ACTS) to address the scientific, computational, and, primarily, communications questions resulting from global change research. As part of this effort three scenarios for oceanographic use of ACTS have been developed; a full discussion of this is contained in Appendix B.
Critical assessment and ramifications of a purported marine trophic cascade
NASA Astrophysics Data System (ADS)
Grubbs, R. Dean; Carlson, John K.; Romine, Jason G.; Curtis, Tobey H.; McElroy, W. David; McCandless, Camilla T.; Cotton, Charles F.; Musick, John A.
2016-02-01
When identifying potential trophic cascades, it is important to clearly establish the trophic linkages between predators and prey with respect to temporal abundance, demographics, distribution, and diet. In the northwest Atlantic Ocean, the depletion of large coastal sharks was thought to trigger a trophic cascade whereby predation release resulted in increased cownose ray abundance, which then caused increased predation on and subsequent collapse of commercial bivalve stocks. These claims were used to justify the development of a predator-control fishery for cownose rays, the “Save the Bay, Eat a Ray” fishery, to reduce predation on commercial bivalves. A reexamination of data suggests declines in large coastal sharks did not coincide with purported rapid increases in cownose ray abundance. Likewise, the increase in cownose ray abundance did not coincide with declines in commercial bivalves. The lack of temporal correlations coupled with published diet data suggest the purported trophic cascade is lacking the empirical linkages required of a trophic cascade. Furthermore, the life history parameters of cownose rays suggest they have low reproductive potential and their populations are incapable of rapid increases. Hypothesized trophic cascades should be closely scrutinized as spurious conclusions may negatively influence conservation and management decisions.
Critical assessment and ramifications of a purported marine trophic cascade
Grubbs, R. Dean; Carlson, John K; Romine, Jason G.; Curtis, Tobey H; McElroy, W. David; McCandless, Camilla T; Cotton, Charles F; Musick, John A.
2016-01-01
When identifying potential trophic cascades, it is important to clearly establish the trophic linkages between predators and prey with respect to temporal abundance, demographics, distribution, and diet. In the northwest Atlantic Ocean, the depletion of large coastal sharks was thought to trigger a trophic cascade whereby predation release resulted in increased cownose ray abundance, which then caused increased predation on and subsequent collapse of commercial bivalve stocks. These claims were used to justify the development of a predator-control fishery for cownose rays, the “Save the Bay, Eat a Ray” fishery, to reduce predation on commercial bivalves. A reexamination of data suggests declines in large coastal sharks did not coincide with purported rapid increases in cownose ray abundance. Likewise, the increase in cownose ray abundance did not coincide with declines in commercial bivalves. The lack of temporal correlations coupled with published diet data suggest the purported trophic cascade is lacking the empirical linkages required of a trophic cascade. Furthermore, the life history parameters of cownose rays suggest they have low reproductive potential and their populations are incapable of rapid increases. Hypothesized trophic cascades should be closely scrutinized as spurious conclusions may negatively influence conservation and management decisions.
Critical assessment and ramifications of a purported marine trophic cascade
Grubbs, R. Dean; Carlson, John K.; Romine, Jason G.; Curtis, Tobey H.; McElroy, W. David; McCandless, Camilla T.; Cotton, Charles F.; Musick, John A.
2016-01-01
When identifying potential trophic cascades, it is important to clearly establish the trophic linkages between predators and prey with respect to temporal abundance, demographics, distribution, and diet. In the northwest Atlantic Ocean, the depletion of large coastal sharks was thought to trigger a trophic cascade whereby predation release resulted in increased cownose ray abundance, which then caused increased predation on and subsequent collapse of commercial bivalve stocks. These claims were used to justify the development of a predator-control fishery for cownose rays, the “Save the Bay, Eat a Ray” fishery, to reduce predation on commercial bivalves. A reexamination of data suggests declines in large coastal sharks did not coincide with purported rapid increases in cownose ray abundance. Likewise, the increase in cownose ray abundance did not coincide with declines in commercial bivalves. The lack of temporal correlations coupled with published diet data suggest the purported trophic cascade is lacking the empirical linkages required of a trophic cascade. Furthermore, the life history parameters of cownose rays suggest they have low reproductive potential and their populations are incapable of rapid increases. Hypothesized trophic cascades should be closely scrutinized as spurious conclusions may negatively influence conservation and management decisions. PMID:26876514
Optimized star sensors laboratory calibration method using a regularization neural network.
Zhang, Chengfen; Niu, Yanxiong; Zhang, Hao; Lu, Jiazhen
2018-02-10
High-precision ground calibration is essential to ensure the performance of star sensors. However, the complex distortion and multi-error coupling have brought great difficulties to traditional calibration methods, especially for large field of view (FOV) star sensors. Although increasing the complexity of models is an effective way to improve the calibration accuracy, it significantly increases the demand for calibration data. In order to achieve high-precision calibration of star sensors with large FOV, a novel laboratory calibration method based on a regularization neural network is proposed. A multi-layer structure neural network is designed to represent the mapping of the star vector and the corresponding star point coordinate directly. To ensure the generalization performance of the network, regularization strategies are incorporated into the net structure and the training algorithm. Simulation and experiment results demonstrate that the proposed method can achieve high precision with less calibration data and without any other priori information. Compared with traditional methods, the calibration error of the star sensor decreased by about 30%. The proposed method can satisfy the precision requirement for large FOV star sensors.
Global warming of the mantle at the origin of flood basalts over supercontinents
NASA Astrophysics Data System (ADS)
Coltice, N.; Phillips, B. R.; Bertrand, H.; Ricard, Y.; Rey, P.
2007-05-01
Continents episodically cluster together into a supercontinent, eventually breaking up with intense magmatic activity supposedly caused by mantle plumes (Morgan, 1983; Richards et al., 1989; Condie, 2004). The breakup of Pangea, the last supercontinent, was accompanied by the emplacement of the largest known continental flood basalt, the Central Atlantic Magmatic Province, which caused massive extinctions at the Triassic-Jurassic boundary (Marzoli et al., 1999). However, there is little support for a plume origin for this catastrophic event (McHone, 2000). On the basis of convection modeling in an internally heated mantle, this paper shows that continental aggregation promotes large-scale melting without requiring the involvement of plumes. When only internal heat sources in the mantle are considered, the formation of a supercontinent causes the enlargement of flow wavelength and a subcontinental increase in temperature as large as 100 °C. This temperature increase may lead to large-scale melting without the involvement of plumes. Our results suggest the existence of two distinct types of continental flood basalts, caused by plume or by mantle global warming.
Silver nanorod structures for metal enhanced fluorescence
NASA Astrophysics Data System (ADS)
Badshah, Mohsin Ali; Lu, Xun; Ju, Jonghyun; Kim, Seok-min
2016-09-01
Fluorescence based detection is a commonly used methodology in biotechnology and medical diagnostics. Metalenhanced fluorescence (MEF) becomes a promising strategy to improve the sensitivity of fluorescence detection, where fluorophores coupling with surface plasmon on metallic structures results fluorescence enhancement. To apply the MEF methodology in real medical diagnostics, especially for protein or DNA microarray detection, a large area (e.g., slide glass, 75 × 25 mm2) with uniform metallic nanostructures is required. In this study, we fabricated a large area MEF substrates using oblique angle deposition (OAD), which is a single step, inexpensive large area fabrication method of nanostructures. To optimize the morphological effect, Ag-nanorods with various lengths were fabricated on the conventional slide glass substrates. Streptavidin-Cy5 dissolved in buffer solution with different concentration (100ng/ml 100μg/ml) were applied to MEF substrates using a pipette, and the fluorescence signals were measured. The enhancement factor increased with the increase in length of Ag-nanorods and maximum enhancement factor 91x was obtained from Ag-nanorods 750nm length compare to bare glass due to higher surface Plasmon effect.
Gamma-burst emission from neutron-star accretion
NASA Technical Reports Server (NTRS)
Colgate, S. A.; Petschek, A. G.; Sarracino, R.
1983-01-01
A model for emission of the hard photons of gamma bursts is presented. The model assumes accretion at nearly the Eddington limited rate onto a neutron star without a magnetic field. Initially soft photons are heated as they are compressed between the accreting matter and the star. A large electric field due to relatively small charge separation is required to drag electrons into the star with the nuclei against the flux of photons leaking out through the accreting matter. The photon number is not increased substantially by Bremsstrahlung or any other process. It is suggested that instability in an accretion disc might provide the infalling matter required.
Otorhinolaryngological aspects of sleep-related breathing disorders
Virk, Jagdeep S.
2016-01-01
Snoring and obstructive sleep apnoea (OSA) are disorders within a wide spectrum of sleep-related breathing disorders (SRBD). Given the obesity epidemic, these conditions will become increasingly prevalent and continue to serve as a large economic burden. A thorough clinical evaluation and appropriate investigations will allow stratification of patients into appropriate treatment groups. A multidisciplinary team is required to manage these patients. Patient selection is critical in ensuring successful surgical and non-surgical outcomes. A wide range of options are available and further long term prospective studies, with standardised data capture and outcome goals, are required to evaluate the most appropriate techniques and long term success rates. PMID:26904262
NASA Astrophysics Data System (ADS)
Cao, Xuesong; Jiang, Ling; Hu, Ruimin
2006-10-01
Currently, the applications of surveillance system have been increasingly widespread. But there are few surveillance platforms that can meet the requirement of large-scale, cross-regional, and flexible surveillance business. In the paper, we present a distributed surveillance system platform to improve safety and security of the society. The system is constructed by an object-oriented middleware called as Internet Communications Engine (ICE). This middleware helps our platform to integrate a lot of surveillance resource of the society and accommodate diverse range of surveillance industry requirements. In the follow sections, we will describe in detail the design concepts of system and introduce traits of ICE.
Interagency telemetry arraying for Voyager-Neptune encounter
NASA Technical Reports Server (NTRS)
Brown, D. W.; Brundage, W. D.; Ulvestad, J. S.; Kent, S. S.; Bartos, K. P.
1990-01-01
The reception capability of the Deep Space Network (DSN) has been improved over the years by increasing both the size and number of antennas at each complex to meet spacecraft-support requirements. However, even more aperture was required for the final planetary encounters of the Voyager 2 spacecraft. This need was met by arraying one radio astronomy observatory with the DSN complex in the United States and another with the complex in Australia. Following a review of augmentation for the Uranus encounter, both the preparation at the National Radio Astronomy (NRAO) Very Large Array (VLA) and the Neptune encounter results for the Parkes-Canberra and VLA-Goldstone arrays are presented.
An assessment of future computer system needs for large-scale computation
NASA Technical Reports Server (NTRS)
Lykos, P.; White, J.
1980-01-01
Data ranging from specific computer capability requirements to opinions about the desirability of a national computer facility are summarized. It is concluded that considerable attention should be given to improving the user-machine interface. Otherwise, increased computer power may not improve the overall effectiveness of the machine user. Significant improvement in throughput requires highly concurrent systems plus the willingness of the user community to develop problem solutions for that kind of architecture. An unanticipated result was the expression of need for an on-going cross-disciplinary users group/forum in order to share experiences and to more effectively communicate needs to the manufacturers.
Xenon Purification Research and Development for the LZ Dark Matter Experiment
NASA Astrophysics Data System (ADS)
Pech, Katherin
2013-04-01
The LZ Experiment is a next generation dark matter detector based on the current LUX detector design, with a 7-ton active volume. Although many research and development breakthroughs were achieved for the 350 kg LUX detector, the large volume scaling required for LZ presents a new set of design challenges that need to be overcome. Because the search for WIMP-like dark matter requires ultra low background experiments, the xenon target material in the LZ detector must meet purity specifications beyond what is commercially available. This challenge is two-fold. The xenon must contain extremely low amounts of electronegative impurities such as oxygen, which attenuate the charge signal. Additionally, it must also have very little of the inert isotope Kr-85, a beta-emitter that can obscure the dark matter signal in the detector volume. The purity requirements for the LUX experiment have been achieved, but the factor of 20 scaling in volume for LZ and increased demands for sensitivity mean that new research and development work must be done to increase our xenon purification capabilities. This talk will focus on the efforts being done at Case Western Reserve University to meet these strict purity requirements for the LZ Experiment.
Financing commercial RLVs: Considering government incentives
NASA Astrophysics Data System (ADS)
Greenberg, Joel S.
1997-01-01
There appears to be a national goal to achieve a commercial space transportation industry that provides launch services utilizing a fleet of reusable launch vehicles (RLVs). Because of the combination of large required investment, inadequate rate of return, and perceived high risk, industry has indicated that this goal may not be achievable without government support. What form of government support will likely be necessary? Government programs and policies can effect private sector investment decisions by reducing risk perceptions, reducing capital requirements, and increasing expected rates of return. Different programs and policies will have different impacts. For example, tax policies will affect expected return on investment but are likely to have little or no effect on risk perceptions and magnitude of required investment, whereas anchor tenancy is likely to alter risk perceptions and may increase expected rates of return. This paper is concerned with the development of an approach that may be used to identify packages of government incentives that may be required to influence private sector investment decisions so as to achieve the desired goal of a commercial space transportation industry that provides launch services utilizing a fleet of RLVs. The paper discusses the relationship of government incentive programs and policies to the RLV investment decision.
Impact of Large Scale Energy Efficiency Programs On Consumer Tariffs and Utility Finances in India
DOE Office of Scientific and Technical Information (OSTI.GOV)
Abhyankar, Nikit; Phadke, Amol
2011-01-20
Large-scale EE programs would modestly increase tariffs but reduce consumers' electricity bills significantly. However, the primary benefit of EE programs is a significant reduction in power shortages, which might make these programs politically acceptable even if tariffs increase. To increase political support, utilities could pursue programs that would result in minimal tariff increases. This can be achieved in four ways: (a) focus only on low-cost programs (such as replacing electric water heaters with gas water heaters); (b) sell power conserved through the EE program to the market at a price higher than the cost of peak power purchase; (c) focusmore » on programs where a partial utility subsidy of incremental capital cost might work and (d) increase the number of participant consumers by offering a basket of EE programs to fit all consumer subcategories and tariff tiers. Large scale EE programs can result in consistently negative cash flows and significantly erode the utility's overall profitability. In case the utility is facing shortages, the cash flow is very sensitive to the marginal tariff of the unmet demand. This will have an important bearing on the choice of EE programs in Indian states where low-paying rural and agricultural consumers form the majority of the unmet demand. These findings clearly call for a flexible, sustainable solution to the cash-flow management issue. One option is to include a mechanism like FAC in the utility incentive mechanism. Another sustainable solution might be to have the net program cost and revenue loss built into utility's revenue requirement and thus into consumer tariffs up front. However, the latter approach requires institutionalization of EE as a resource. The utility incentive mechanisms would be able to address the utility disincentive of forgone long-run return but have a minor impact on consumer benefits. Fundamentally, providing incentives for EE programs to make them comparable to supply-side investments is a way of moving the electricity sector toward a model focused on providing energy services rather than providing electricity.« less
Watermarking techniques for electronic delivery of remote sensing images
NASA Astrophysics Data System (ADS)
Barni, Mauro; Bartolini, Franco; Magli, Enrico; Olmo, Gabriella
2002-09-01
Earth observation missions have recently attracted a growing interest, mainly due to the large number of possible applications capable of exploiting remotely sensed data and images. Along with the increase of market potential, the need arises for the protection of the image products. Such a need is a very crucial one, because the Internet and other public/private networks have become preferred means of data exchange. A critical issue arising when dealing with digital image distribution is copyright protection. Such a problem has been largely addressed by resorting to watermarking technology. A question that obviously arises is whether the requirements imposed by remote sensing imagery are compatible with existing watermarking techniques. On the basis of these motivations, the contribution of this work is twofold: assessment of the requirements imposed by remote sensing applications on watermark-based copyright protection, and modification of two well-established digital watermarking techniques to meet such constraints. More specifically, the concept of near-lossless watermarking is introduced and two possible algorithms matching such a requirement are presented. Experimental results are shown to measure the impact of watermark introduction on a typical remote sensing application, i.e., unsupervised image classification.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Johnson, Michael J.; Go, David B., E-mail: dgo@nd.edu; Department of Chemical and Biomolecular Engineering, University of Notre Dame, Notre Dame, Indianapolis 46556
To generate a gas discharge (plasma) in atmospheric air requires an electric field that exceeds the breakdown threshold of ∼30 kV/cm. Because of safety, size, or cost constraints, the large applied voltages required to generate such fields are often prohibitive for portable applications. In this work, piezoelectric transformers are used to amplify a low input applied voltage (<30 V) to generate breakdown in air without the need for conventional high-voltage electrical equipment. Piezoelectric transformers (PTs) use their inherent electromechanical resonance to produce a voltage amplification, such that the surface of the piezoelectric exhibits a large surface voltage that can generate corona-like dischargesmore » on its corners or on adjacent electrodes. In the proper configuration, these discharges can be used to generate a bulk air flow called an ionic wind. In this work, PT-driven discharges are characterized by measuring the discharge current and the velocity of the induced ionic wind with ionic winds generated using input voltages as low as 7 V. The characteristics of the discharge change as the input voltage increases; this modifies the resonance of the system and subsequent required operating parameters.« less
Comparing memory-efficient genome assemblers on stand-alone and cloud infrastructures.
Kleftogiannis, Dimitrios; Kalnis, Panos; Bajic, Vladimir B
2013-01-01
A fundamental problem in bioinformatics is genome assembly. Next-generation sequencing (NGS) technologies produce large volumes of fragmented genome reads, which require large amounts of memory to assemble the complete genome efficiently. With recent improvements in DNA sequencing technologies, it is expected that the memory footprint required for the assembly process will increase dramatically and will emerge as a limiting factor in processing widely available NGS-generated reads. In this report, we compare current memory-efficient techniques for genome assembly with respect to quality, memory consumption and execution time. Our experiments prove that it is possible to generate draft assemblies of reasonable quality on conventional multi-purpose computers with very limited available memory by choosing suitable assembly methods. Our study reveals the minimum memory requirements for different assembly programs even when data volume exceeds memory capacity by orders of magnitude. By combining existing methodologies, we propose two general assembly strategies that can improve short-read assembly approaches and result in reduction of the memory footprint. Finally, we discuss the possibility of utilizing cloud infrastructures for genome assembly and we comment on some findings regarding suitable computational resources for assembly.
Theory and computation of optimal low- and medium-thrust transfers
NASA Technical Reports Server (NTRS)
Chuang, C.-H.
1994-01-01
This report presents two numerical methods considered for the computation of fuel-optimal, low-thrust orbit transfers in large numbers of burns. The origins of these methods are observations made with the extremal solutions of transfers in small numbers of burns; there seems to exist a trend such that the longer the time allowed to perform an optimal transfer the less fuel that is used. These longer transfers are obviously of interest since they require a motor of low thrust; however, we also find a trend that the longer the time allowed to perform the optimal transfer the more burns are required to satisfy optimality. Unfortunately, this usually increases the difficulty of computation. Both of the methods described use small-numbered burn solutions to determine solutions in large numbers of burns. One method is a homotopy method that corrects for problems that arise when a solution requires a new burn or coast arc for optimality. The other method is to simply patch together long transfers from smaller ones. An orbit correction problem is solved to develop this method. This method may also lead to a good guidance law for transfer orbits with long transfer times.
Local wavelet transform: a cost-efficient custom processor for space image compression
NASA Astrophysics Data System (ADS)
Masschelein, Bart; Bormans, Jan G.; Lafruit, Gauthier
2002-11-01
Thanks to its intrinsic scalability features, the wavelet transform has become increasingly popular as decorrelator in image compression applications. Throuhgput, memory requirements and complexity are important parameters when developing hardware image compression modules. An implementation of the classical, global wavelet transform requires large memory sizes and implies a large latency between the availability of the input image and the production of minimal data entities for entropy coding. Image tiling methods, as proposed by JPEG2000, reduce the memory sizes and the latency, but inevitably introduce image artefacts. The Local Wavelet Transform (LWT), presented in this paper, is a low-complexity wavelet transform architecture using a block-based processing that results in the same transformed images as those obtained by the global wavelet transform. The architecture minimizes the processing latency with a limited amount of memory. Moreover, as the LWT is an instruction-based custom processor, it can be programmed for specific tasks, such as push-broom processing of infinite-length satelite images. The features of the LWT makes it appropriate for use in space image compression, where high throughput, low memory sizes, low complexity, low power and push-broom processing are important requirements.
Mapping of Residues Forming the Voltage Sensor of the Voltage-Dependent Anion-Selective Channel
NASA Astrophysics Data System (ADS)
Thomas, Lorie; Blachly-Dyson, Elizabeth; Colombini, Marco; Forte, Michael
1993-06-01
Voltage-gated ion-channel proteins contain "voltage-sensing" domains that drive the conformational transitions between open and closed states in response to changes in transmembrane voltage. We have used site-directed mutagenesis to identify residues affecting the voltage sensitivity of a mitochondrial channel, the voltage-dependent anion-selective channel (VDAC). Although charge changes at many sites had no effect, at other sites substitutions that increased positive charge also increased the steepness of voltage dependance and substitutions that decreased positive charge decreased voltage dependance by an appropriate amount. In contrast to the plasma membrane K^+ and Na^+ channels, these residues are distributed over large parts of the VDAC protein. These results have been used to define the conformational transitions that accompany voltage gating of an ion channel. This gating mechanism requires the movement of large portions of the VDAC protein through the membrane.
Life in the fast lane: high-throughput chemistry for lead generation and optimisation.
Hunter, D
2001-01-01
The pharmaceutical industry has come under increasing pressure due to regulatory restrictions on the marketing and pricing of drugs, competition, and the escalating costs of developing new drugs. These forces can be addressed by the identification of novel targets, reductions in the development time of new drugs, and increased productivity. Emphasis has been placed on identifying and validating new targets and on lead generation: the response from industry has been very evident in genomics and high throughput screening, where new technologies have been applied, usually coupled with a high degree of automation. The combination of numerous new potential biological targets and the ability to screen large numbers of compounds against many of these targets has generated the need for large diverse compound collections. To address this requirement, high-throughput chemistry has become an integral part of the drug discovery process. Copyright 2002 Wiley-Liss, Inc.
NASA Astrophysics Data System (ADS)
Neklyudov, A. A.; Savenkov, V. N.; Sergeyez, A. G.
1984-06-01
Memories are improved by increasing speed or the memory volume on a single chip. The most effective means for increasing speeds in bipolar memories are current control circuits with the lowest extraction times for a specific power consumption (1/4 pJ/bit). The control current circuitry involves multistage current switches and circuits accelerating transient processes in storage elements and links. Circuit principles for the design of bipolar memories with maximum speeds for an assigned minimum of circuit topology are analyzed. Two main classes of storage with current control are considered: the ECL type and super-integrated injection type storage with data capacities of N = 1/4 and N 4/16, respectively. The circuits reduce logic voltage differentials and the volumes of lexical and discharge buses and control circuit buses. The limiting speed is determined by the antiinterference requirements of the memory in storage and extraction modes.
Frictional velocity-weakening in landslides on Earth and on other planetary bodies.
Lucas, Antoine; Mangeney, Anne; Ampuero, Jean Paul
2014-03-04
One of the ultimate goals in landslide hazard assessment is to predict maximum landslide extension and velocity. Despite much work, the physical processes governing energy dissipation during these natural granular flows remain uncertain. Field observations show that large landslides travel over unexpectedly long distances, suggesting low dissipation. Numerical simulations of landslides require a small friction coefficient to reproduce the extension of their deposits. Here, based on analytical and numerical solutions for granular flows constrained by remote-sensing observations, we develop a consistent method to estimate the effective friction coefficient of landslides. This method uses a constant basal friction coefficient that reproduces the first-order landslide properties. We show that friction decreases with increasing volume or, more fundamentally, with increasing sliding velocity. Inspired by frictional weakening mechanisms thought to operate during earthquakes, we propose an empirical velocity-weakening friction law under a unifying phenomenological framework applicable to small and large landslides observed on Earth and beyond.
Big Data and Health Economics: Strengths, Weaknesses, Opportunities and Threats.
Collins, Brendan
2016-02-01
'Big data' is the collective name for the increasing capacity of information systems to collect and store large volumes of data, which are often unstructured and time stamped, and to analyse these data by using regression and other statistical techniques. This is a review of the potential applications of big data and health economics, using a SWOT (strengths, weaknesses, opportunities, threats) approach. In health economics, large pseudonymized databases, such as the planned care.data programme in the UK, have the potential to increase understanding of how drugs work in the real world, taking into account adherence, co-morbidities, interactions and side effects. This 'real-world evidence' has applications in individualized medicine. More routine and larger-scale cost and outcomes data collection will make health economic analyses more disease specific and population specific but may require new skill sets. There is potential for biomonitoring and lifestyle data to inform health economic analyses and public health policy.
Potential productivity benefits of float-zone versus Czochralski crystal growth
NASA Technical Reports Server (NTRS)
Abe, T.
1985-01-01
Efficient mass production of single-crystal silicon is necessary for the efficient silicon solar arrays needed in the coming decade. However, it is anticipated that there will be difficulty growing such volumes of crystals using conventional Czochralski (Cz) methods. While the productivity of single crystals might increase with a crystal diameter increase, there are two obstacles to the mass production of large diameter Czochralski crystals, the long production cycle due to slow growth rate and the high heat requirements of the furnaces. Also counterproductive would be the large resistivity gradient along the growth direction of the crystals due to impurity concentration. Comparison between Float zone (FZ) and Cz crystal growth on the basis of a crystal 150 mm in diameter is on an order of two to four times in favor of the FZ method. This advantage results from high growth rates and steady-state growth while maintaining a dislocation-free condition and impurity segregation.
Evaluation of ultra-low background materials for uranium and thorium using ICP-MS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hoppe, E. W.; Overman, N. R.; LaFerriere, B. D.
2013-08-08
An increasing number of physics experiments require low background materials for their construction. The presence of Uranium and Thorium and their progeny in these materials present a variety of unwanted background sources for these experiments. The sensitivity of the experiments continues to drive the necessary levels of detection ever lower as well. This requirement for greater sensitivity has rendered direct radioassay impractical in many cases requiring large quantities of material, frequently many kilograms, and prolonged counting times, often months. Other assay techniques have been employed such as Neutron Activation Analysis but this requires access to expensive facilities and instrumentation andmore » can be further complicated and delayed by the formation of unwanted radionuclides. Inductively Coupled Plasma Mass Spectrometry (ICP-MS) is a useful tool and recent advancements have increased the sensitivity particularly in the elemental high mass range of U and Th. Unlike direct radioassay, ICP-MS is a destructive technique since it requires the sample to be in liquid form which is aspirated into a high temperature plasma. But it benefits in that it usually requires a very small sample, typically about a gram. This paper discusses how a variety of low background materials such as copper, polymers, and fused silica are made amenable to ICP-MS assay and how the arduous task of maintaining low backgrounds of U and Th is achieved.« less
Evaluation of Ultra-Low Background Materials for Uranium and Thorium Using ICP-MS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hoppe, Eric W.; Overman, Nicole R.; LaFerriere, Brian D.
2013-08-08
An increasing number of physics experiments require low background materials for their construction. The presence of Uranium and Thorium and their progeny in these materials present a variety of unwanted background sources for these experiments. The sensitivity of the experiments continues to drive the necessary levels of detection ever lower as well. This requirement for greater sensitivity has rendered direct radioassay impractical in many cases requiring large quantities of material, frequently many kilograms, and prolonged counting times, often months. Other assay techniques have been employed such as Neutron Activation Analysis but this requires access to expensive facilities and instrumentation andmore » can be further complicated and delayed by the formation of unwanted radionuclides. Inductively Coupled Plasma Mass Spectrometry (ICP-MS) is a useful tool and recent advancements have increased the sensitivity particularly in the elemental high mass range of U and Th. Unlike direct radioassay, ICP-MS is a destructive technique since it requires the sample to be in liquid form which is aspirated into a high temperature plasma. But it benefits in that it usually requires a very small sample, typically about a gram. Here we will discuss how a variety of low background materials such as copper, polymers, and fused silica are made amenable to ICP-MS assay and how the arduous task of maintaining low backgrounds of U and Th is achieved.« less
Hazard, Brittany; Zhang, Xiaoqin; Naemeh, Mahmoudreza; Hamilton, M. Kristina; Rust, Bret; Raybould, Helen E.; Newman, John W.; Martin, Roy; Dubcovsky, Jorge
2016-01-01
Increased amylose in wheat (Triticum ssp.) starch is associated with increased resistant starch, a fermentable dietary fiber. Fermentation of resistant starch in the large intestine produces short-chain fatty acids that are associated with human health benefits. Since wheat foods are an important component of the human diet, increases in amylose and resistant starch in wheat grains have the potential to deliver health benefits to a large number of people. In three replicated field trials we found that mutations in starch branching enzyme II genes (SBEIIa and SBEIIb) in both A and B genomes (SBEIIa/b-AB) of durum wheat [T. turgidum L. subsp. durum (Desf.) Husn.] resulted in large increases of amylose and resistant starch content. The presence of these four mutations was also associated with an average 5% reduction in kernel weight (P = 0.0007) and 15% reduction in grain yield (P = 0.06) compared to the wild type. Complete milling and pasta quality analysis showed that the mutant lines have an acceptable quality with positive effects on pasta firmness and negative effects on semolina extraction and pasta color. Positive fermentation responses were detected in rats (Rattus spp.) fed with diets incorporating mutant wheat flour. This study quantifies benefits and limitations associated with the deployment of the SBEIIa/b-AB mutations in durum wheat and provides the information required to develop realistic strategies to deploy durum wheat varieties with increased levels of amylose and resistant starch. PMID:27134286
Identifying western yellow-billed cuckoo breeding habitat with a dual modelling approach
Johnson, Matthew J.; Hatten, James R.; Holmes, Jennifer A.; Shafroth, Patrick B.
2017-01-01
The western population of the yellow-billed cuckoo (Coccyzus americanus) was recently listed as threatened under the federal Endangered Species Act. Yellow-billed cuckoo conservation efforts require the identification of features and area requirements associated with high quality, riparian forest habitat at spatial scales that range from nest microhabitat to landscape, as well as lower-suitability areas that can be enhanced or restored. Spatially explicit models inform conservation efforts by increasing ecological understanding of a target species, especially at landscape scales. Previous yellow-billed cuckoo modelling efforts derived plant-community maps from aerial photography, an expensive and oftentimes inconsistent approach. Satellite models can remotely map vegetation features (e.g., vegetation density, heterogeneity in vegetation density or structure) across large areas with near perfect repeatability, but they usually cannot identify plant communities. We used aerial photos and satellite imagery, and a hierarchical spatial scale approach, to identify yellow-billed cuckoo breeding habitat along the Lower Colorado River and its tributaries. Aerial-photo and satellite models identified several key features associated with yellow-billed cuckoo breeding locations: (1) a 4.5 ha core area of dense cottonwood-willow vegetation, (2) a large native, heterogeneously dense forest (72 ha) around the core area, and (3) moderately rough topography. The odds of yellow-billed cuckoo occurrence decreased rapidly as the amount of tamarisk cover increased or when cottonwood-willow vegetation was limited. We achieved model accuracies of 75–80% in the project area the following year after updating the imagery and location data. The two model types had very similar probability maps, largely predicting the same areas as high quality habitat. While each model provided unique information, a dual-modelling approach provided a more complete picture of yellow-billed cuckoo habitat requirements and will be useful for management and conservation activities.
Variable Acceleration Force Calibration System (VACS)
NASA Technical Reports Server (NTRS)
Rhew, Ray D.; Parker, Peter A.; Johnson, Thomas H.; Landman, Drew
2014-01-01
Conventionally, force balances have been calibrated manually, using a complex system of free hanging precision weights, bell cranks, and/or other mechanical components. Conventional methods may provide sufficient accuracy in some instances, but are often quite complex and labor-intensive, requiring three to four man-weeks to complete each full calibration. To ensure accuracy, gravity-based loading is typically utilized. However, this often causes difficulty when applying loads in three simultaneous, orthogonal axes. A complex system of levers, cranks, and cables must be used, introducing increased sources of systematic error, and significantly increasing the time and labor intensity required to complete the calibration. One aspect of the VACS is a method wherein the mass utilized for calibration is held constant, and the acceleration is changed to thereby generate relatively large forces with relatively small test masses. Multiple forces can be applied to a force balance without changing the test mass, and dynamic forces can be applied by rotation or oscillating acceleration. If rotational motion is utilized, a mass is rigidly attached to a force balance, and the mass is exposed to a rotational field. A large force can be applied by utilizing a large rotational velocity. A centrifuge or rotating table can be used to create the rotational field, and fixtures can be utilized to position the force balance. The acceleration may also be linear. For example, a table that moves linearly and accelerates in a sinusoidal manner may also be utilized. The test mass does not have to move in a path that is parallel to the ground, and no re-leveling is therefore required. Balance deflection corrections may be applied passively by monitoring the orientation of the force balance with a three-axis accelerometer package. Deflections are measured during each test run, and adjustments with respect to the true applied load can be made during the post-processing stage. This paper will present the development and testing of the VASC concept.
ERIC Educational Resources Information Center
Zhang, Xuesong; Dorn, Bradley
2012-01-01
Agile development has received increasing interest both in industry and academia due to its benefits in developing software quickly, meeting customer needs, and keeping pace with the rapidly changing requirements. However, agile practices and scrum in particular have been mainly tested in mid- to large-size projects. In this paper, we present…
Health Effects of Embedded Depleted Uranium Fragments.
1996-11-15
18]. The metabolic (approximately 99.8%). DU emits alpha (a), beta models required to estimate the chemical and radio- (3), and weak gamma ( y ...for the large-scale screening of patients who choosing an animal model for carcinogenesis stud- may be at increased risk of developing cancer . ies...appears warranted biomarkers for neoplastic tissues. since it is likely to provide better assurance of detecting changes that are associated with Cancer
Laboratory Spectrometer for Wear Metal Analysis of Engine Lubricants.
1986-04-01
analysis, the acid digestion technique for sample pretreatment is the best approach available to date because of its relatively large sample size (1000...microliters or more). However, this technique has two major shortcomings limiting its application: (1) it requires the use of hydrofluoric acid (a...accuracy. Sample preparation including filtration or acid digestion may increase analysis times by 20 minutes or more. b. Repeatability In the analysis
A Holistic Management Architecture for Large-Scale Adaptive Networks
2007-09-01
transmission and processing overhead required for management. The challenges of building models to describe dynamic systems are well-known to the field of...increases the challenge of finding a simple approach to assessing the state of the network. Moreover, the performance state of one network link may be... challenging . These obstacles indicate the need for a less comprehensive-analytical, more systemic-holistic approach to managing networks. This approach might
ERIC Educational Resources Information Center
Hayes, Ann Milligan
2015-01-01
Over the last two decades, there has been renewed interest in formative assessment, in large part due to the increasing pressures and prevalence of "high stakes" summative assessments. As states try to meet the requirements of the No Child Left Behind law, teachers and administrators are realizing that formative assessment offers an…
An Introduction to Military Satellite Communications.
1987-04-01
in the range 220 - 400 MHz, and hence shared with a variety of terrestrial systems (including TV). Helix antennas are commonly employed, both on the...downlink). A common helix antenna is used for both transmit and receive, with a multiplexer unit separating the two. This antenna is relatively large...into account; these increase sharply as the TWT is driven into saturation, and the controller will determine any required back-off (and will perhaps
Increasing Efficiency of Fecal Coliform Testing Through EPA-Approved Alternate Method Colilert*-18
NASA Technical Reports Server (NTRS)
Cornwell, Brian
2017-01-01
The 21 SM 9221 E multiple-tube fermentation method for fecal coliform analysis requires a large time and reagent investment for the performing laboratory. In late 2010, the EPA approved an alternative procedure for the determination of fecal coliforms designated as Colilert*-18. However, as of late 2016, only two VELAP-certified laboratories in the Commonwealth of Virginia have been certified in this method.
Preliminary analysis of the implications of natural radiations on geostationary operations
NASA Technical Reports Server (NTRS)
Wilson, J. W.; Denn, F. M.
1976-01-01
The natural radiations present at geostationary orbit are discussed. Low-level galactic cosmic rays are important for careers spending a year or more at geostationary altitude. Trapped radiation will on occasion require interruption of extravehicular activity (EVA). The spacesuit shield requirements are strongly affected by the number of interruptions allowed. EVA cannot proceed during a large solar event and maximum allowable doses are exceeded in a few hours unless a heavily shielded area is provided. A shelter of 10 g/sq cm with personal shielding for the eyes and testes would contain exposure to within the presently accepted exposure constraints. Since radiation levels can increase unexpectedly to serious levels, an onboard radiation monitoring system with rate and integration capabilities is required for both surface-dose and depth-dose monitoring.
Using artificial neural networks (ANN) for open-loop tomography
NASA Astrophysics Data System (ADS)
Osborn, James; De Cos Juez, Francisco Javier; Guzman, Dani; Butterley, Timothy; Myers, Richard; Guesalaga, Andres; Laine, Jesus
2011-09-01
The next generation of adaptive optics (AO) systems require tomographic techniques in order to correct for atmospheric turbulence along lines of sight separated from the guide stars. Multi-object adaptive optics (MOAO) is one such technique. Here, we present a method which uses an artificial neural network (ANN) to reconstruct the target phase given off-axis references sources. This method does not require any input of the turbulence profile and is therefore less susceptible to changing conditions than some existing methods. We compare our ANN method with a standard least squares type matrix multiplication method (MVM) in simulation and find that the tomographic error is similar to the MVM method. In changing conditions the tomographic error increases for MVM but remains constant with the ANN model and no large matrix inversions are required.
Optics Requirements For The Generation-X X-Ray Telescope
NASA Technical Reports Server (NTRS)
O'Dell, S. .; Elsner, R. F.; Kolodziejczak, J. J.; Ramsey, B. D.; Weisskopf, M. C.; Zhang, W. W.; Content, D. A.; Petre, R.; Saha, T. T.; Reid, P. B.;
2008-01-01
US, European, and Japanese space agencies each now operate successful X-ray missions -- NASA s Chandra, ESA s XMM-Newton, and JAXA s Suzaku observatories. Recently these agencies began a collaboration to develop the next major X-ray astrophysics facility -- the International X-ray Observatory (IXO) -- for launch around 2020. IXO will provide an order-of-magnitude increase in effective area, while maintaining good (but not sub-arcsecond) angular resolution. X-ray astronomy beyond IXO will require optics with even larger aperture areas and much better angular resolution. We are currently conducting a NASA strategic mission concept study to identify technology issues and to formulate a technology roadmap for a mission -- Generation-X (Gen-X) -- to provide these capabilities. Achieving large X-ray collecting areas in a space observatory requires extremely lightweight mirrors.
NASA Astrophysics Data System (ADS)
Deng, Mingfeng; Chen, Ningsheng; Ding, Haitao
2018-02-01
The Parlung Zangbo Basin in the southeastern Tibet Plateau is affected by the summer monsoon from the Indian Ocean, which produces large rainfall gradients in the basin. Rainfall data during 2012-2015 from five new meteorological stations are used to analyse the rainfall characteristics. The daily rainfall, rainfall duration, mean rainfall intensity, and peak rainfall intensity are consistent, but sometimes contrasting. For example, these values decrease with increasing altitude, and the gradient is large downstream and small upstream, respectively. Moreover, the rainfall intensity peaks between 01:00 and 06:00 and increases during the afternoon. Based on the analysis of 14 debris flow cases in the basin, differences in the rainfall threshold differ depending on the location as sediment varieties. The sediment in the middle portions of the basin is wet and well structured; thus, long-duration, high-intensity rainfall is required to generate debris flows. Ravels in the upstream area are arid and not well structured, and short-duration rainfall is required to trigger debris flows. Between the above two locations, either long-duration, low-intensity rainfall or short-duration, high-intensity rainfall could provoke debris flows. Clearly, differences in rainfall characteristics and rainfall thresholds that are associated with the location must be considered in debris flow monitoring and warnings.
State of the Art and Development Trends of the Digital Radiography Systems for Cargo Inspection
NASA Astrophysics Data System (ADS)
Udod, V.; Van, J.; Osipov, S.; Chakhlov, S.; Temnik, A.
2016-01-01
Increasing requirements for technical parameters of inspection digital radiography systems are caused by increasing incidences of terrorism, drug trafficking and explosives via variety of transport. These requirements have determined research for new technical solutions that enable to ensure the safety of passengers and cargos in real-time. The main efforts in the analyzed method of testing are aimed at the creation of new and modernization of operated now systems of digital radiography as a whole and their main components and elements in particular. The number of these main components and elements includes sources of X-ray recording systems and transformation of radiometric information as well as algorithms and software that implements these algorithms for processing, visualization and results interpretation of inspection. Recent developments of X-ray units and betatrons used for inspection of small- and large-sized objects that are made from different materials are deserve special attention. The most effective X-ray detectors are a line and a radiometric detector matrix based on various scintillators. The most promising methods among the algorithms of material identification of testing objects are dual-energy methods. The article describes various models of digital radiography systems applied in Russia and abroad to inspection of baggage, containers, vehicles and large trucks.
The future of very large subsonic transports
NASA Technical Reports Server (NTRS)
Justice, R. Steven; Hays, Anthony P.; Parrott, Ed L.
1996-01-01
The Very Large Subsonic Transport (VLST) is a multi-use commercial passenger, commercial cargo, and military airlifter roughly 50% larger than the current Lockheed C-5 and Boeing 747. Due to the large size and cost of the VLST, it is unlikely that the commercial market can support more than one aircraft production line, while declining defense budgets will not support a dedicated military VLST. A successful VLST must therefore meet airline requirements for more passenger and cargo capacity on congested routes into slot-limited airports and also provide a cost effective heavy airlift capacity to support the overseas deployment of US military forces. A successful VLST must satisfy three key missions: commercial passenger service with nominal seating capacity at a minimum of 650 passengers with a range capability of 7,000 to 10,000 miles; commercial air cargo service for containerized cargo to support global manufacturing of high value added products, 'just-in-time' parts delivery, and the general globalization of trade; and military airlift with adequate capacity to load current weapon systems, with minimal break-down, over global ranges (7,000 to 10,000 miles) required to reach the operational theater without need of overseas bases and midair refueling. The development of the VLST poses some technical issues specific to large aircraft, but also key technologies applicable to a wide range of subsonic transport aircraft. Key issues and technologies unique to the VLST include: large composite structures; dynamic control of a large, flexible structure; aircraft noise requirements for aircraft over 850,000 pounds; and increased aircraft separation due to increased wake vortex generation. Other issues, while not unique to the VLST, will critically impact the ability to build an efficient and affordable aircraft include: active control systems: Fly-By-Light/Power-By-Wire (FBL/PBW); high lift systems; flight deck associate systems; laminar flow; emergency egress; and modular design. The VLST will encounter severe restrictions on weight, ground flotation, span, length, and door height to operate at current airports/bases, gates, and cargo loading systems. One option under consideration is for a sea-based VLST, either a conventional seaplane or Wing-In-Ground effect (WIG) vehicle, which would allow greater operational flexibility, while introducing other design challenges such as water impact loads and salt-water corrosion. Lockheed Martin is currently developing a floatplane version of the C-130 Hercules which will provide experience with a modern sea-based aircraft. In addition to its own ongoing research activities, Lockheed Martin is also participating in the NASA Advanced Subsonic Technology, High Speed Research (HSR), and other programs which address some of the technologies needed for the VLST. The VLST will require NASA and US aerospace companies to work together to develop new capabilities and technologies for make the VLST a viable part of transportation beyond 2000.
Antenna Electronics Concept for the Next-Generation Very Large Array
NASA Astrophysics Data System (ADS)
Beasley, Anthony J.; Jackson, Jim; Selina, Robert
2017-01-01
The National Radio Astronomy Observatory (NRAO), in collaboration with its international partners, completed two major projects over the past decade: the sensitivity upgrade for the Karl Jansky Very Large Array (VLA) and the construction of the Atacama Large Millimeter/Sub-Millimeter Array (ALMA). The NRAO is now considering the scientific potential and technical feasibility of a next-generation VLA (ngVLA) with an emphasis on thermal imaging at milli-arcsecond resolution. The preliminary goals for the ngVLA are to increase both the system sensitivity and angular resolution of the VLA tenfold and to cover a frequency range of 1.2-116 GHz.A number of key technical challenges have been identified for the project. These include cost-effective antenna manufacturing (in the hundreds), suitable wide-band feed and receiver designs, broad-band data transmission, and large-N correlators. Minimizing the overall operations cost is also a fundamental design requirement.The designs of the antenna electronics, reference distribution system, and data transmission system are anticipated to be major construction and operations cost drivers for the facility. The electronics must achieve a high level of performance, while maintaining low operation and maintenance costs and a high level of reliability. Additionally, due to the uncertainty in the feasibility of wideband receivers, advancements in digitizer technology, and budget constraints, the hardware system architecture should be scalable to the number of receiver bands and the speed and resolution of available digitizers.Here, we present the projected performance requirements of the ngVLA, a proposed block diagram for the instrument’s electronics systems, parameter tradeoffs within the system specifications, and areas of technical risk where technical advances may be required for successful production and installation.
NASA Astrophysics Data System (ADS)
Sugar, Thomas G.; Hollander, Kevin W.; Hitt, Joseph K.
2011-04-01
Developing bionic ankles poses great challenges due to the large moment, power, and energy that are required at the ankle. Researchers have added springs in series with a motor to reduce the peak power and energy requirements of a robotic ankle. We developed a "robotic tendon" that reduces the peak power by altering the required motor speed. By changing the required speed, the spring acts as a "load variable transmission." If a simple motor/gearbox solution is used, one walking step would require 38.8J and a peak motor power of 257 W. Using an optimized robotic tendon, the energy required is 21.2 J and the peak motor power is reduced to 96.6 W. We show that adding a passive spring in parallel with the robotic tendon reduces peak loads but the power and energy increase. Adding a passive spring in series with the robotic tendon reduces the energy requirements. We have built a prosthetic ankle SPARKy, Spring Ankle with Regenerative Kinetics, that allows a user to walk forwards, backwards, ascend and descend stairs, walk up and down slopes as well as jog.
Verdant: automated annotation, alignment and phylogenetic analysis of whole chloroplast genomes.
McKain, Michael R; Hartsock, Ryan H; Wohl, Molly M; Kellogg, Elizabeth A
2017-01-01
Chloroplast genomes are now produced in the hundreds for angiosperm phylogenetics projects, but current methods for annotation, alignment and tree estimation still require some manual intervention reducing throughput and increasing analysis time for large chloroplast systematics projects. Verdant is a web-based software suite and database built to take advantage a novel annotation program, annoBTD. Using annoBTD, Verdant provides accurate annotation of chloroplast genomes without manual intervention. Subsequent alignment and tree estimation can incorporate newly annotated and publically available plastomes and can accommodate a large number of taxa. Verdant sharply reduces the time required for analysis of assembled chloroplast genomes and removes the need for pipelines and software on personal hardware. Verdant is available at: http://verdant.iplantcollaborative.org/plastidDB/ It is implemented in PHP, Perl, MySQL, Javascript, HTML and CSS with all major browsers supported. mrmckain@gmail.comSupplementary information: Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press.
NASA Technical Reports Server (NTRS)
Montgomery, Edward E., IV; Heaton, Andrew F.; Garbe, Gregory P.
2003-01-01
Solar sails are a near term, low thrust, propellantless propulsion technology suitable for orbital maneuvering, station keeping, and attitude control applications for small payloads. Furthermore, these functions can be highly integrated, reducing mass, cost and complexity. The solar sail concept is based on momentum exchange with solar flux reflected from a large, deployed thin membrane. Thrust performance increases as the square of the distance to the sun. In comparison to conventional chemical systems, there are missions where solar sails are vastly more and less economical. The less attractive applications involve large payloads, outer solar system transfers, and short trip times. However, for inclination changes and station keeping at locations requiring constant thrust, the solar sail is the only economical option for missions of more than a few weeks duration. We compare the location and energies required for these applications between solar sails, advanced electric propulsion, and conventional rockets. We address the effect on mass fraction to understand solar sail mission cost and capability. Finally, the benefit of potential applications to near term science missions is reported.
Stitching Type Large Aperture Depolarizer for Gas Monitoring Imaging Spectrometer
NASA Astrophysics Data System (ADS)
Liu, X.; Li, M.; An, N.; Zhang, T.; Cao, G.; Cheng, S.
2018-04-01
To increase the accuracy of radiation measurement for gas monitoring imaging spectrometer, it is necessary to achieve high levels of depolarization of the incoming beam. The preferred method in space instrument is to introduce the depolarizer into the optical system. It is a combination device of birefringence crystal wedges. Limited to the actual diameter of the crystal, the traditional depolarizer cannot be used in the large aperture imaging spectrometer (greater than 100 mm). In this paper, a stitching type depolarizer is presented. The design theory and numerical calculation model for dual babinet depolarizer were built. As required radiometric accuracies of the imaging spectrometer with 250 mm × 46 mm aperture, a stitching type dual babinet depolarizer was design in detail. Based on designing the optimum structural parmeters the tolerance of wedge angle refractive index, and central thickness were given. The analysis results show that the maximum residual polarization degree of output light from depolarizer is less than 2 %. The design requirements of polarization sensitivity is satisfied.
Preparing for in situ processing on upcoming leading-edge supercomputers
Kress, James; Churchill, Randy Michael; Klasky, Scott; ...
2016-10-01
High performance computing applications are producing increasingly large amounts of data and placing enormous stress on current capabilities for traditional post-hoc visualization techniques. Because of the growing compute and I/O imbalance, data reductions, including in situ visualization, are required. These reduced data are used for analysis and visualization in a variety of different ways. Many of he visualization and analysis requirements are known a priori, but when they are not, scientists are dependent on the reduced data to accurately represent the simulation in post hoc analysis. The contributions of this paper is a description of the directions we are pursuingmore » to assist a large scale fusion simulation code succeed on the next generation of supercomputers. Finally, these directions include the role of in situ processing for performing data reductions, as well as the tradeoffs between data size and data integrity within the context of complex operations in a typical scientific workflow.« less
The BaBar Data Reconstruction Control System
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ceseracciu, A
2005-04-20
The BaBar experiment is characterized by extremely high luminosity and very large volume of data produced and stored, with increasing computing requirements each year. To fulfill these requirements a Control System has been designed and developed for the offline distributed data reconstruction system. The control system described in this paper provides the performance and flexibility needed to manage a large number of small computing farms, and takes full benefit of OO design. The infrastructure is well isolated from the processing layer, it is generic and flexible, based on a light framework providing message passing and cooperative multitasking. The system ismore » distributed in a hierarchical way: the top-level system is organized in farms, farms in services, and services in subservices or code modules. It provides a powerful Finite State Machine framework to describe custom processing models in a simple regular language. This paper describes the design and evolution of this control system, currently in use at SLAC and Padova on {approx}450 CPUs organized in 9 farms.« less
The BaBar Data Reconstruction Control System
NASA Astrophysics Data System (ADS)
Ceseracciu, A.; Piemontese, M.; Tehrani, F. S.; Pulliam, T. M.; Galeazzi, F.
2005-08-01
The BaBar experiment is characterized by extremely high luminosity and very large volume of data produced and stored, with increasing computing requirements each year. To fulfill these requirements a control system has been designed and developed for the offline distributed data reconstruction system. The control system described in this paper provides the performance and flexibility needed to manage a large number of small computing farms, and takes full benefit of object oriented (OO) design. The infrastructure is well isolated from the processing layer, it is generic and flexible, based on a light framework providing message passing and cooperative multitasking. The system is distributed in a hierarchical way: the top-level system is organized in farms, farms in services, and services in subservices or code modules. It provides a powerful finite state machine framework to describe custom processing models in a simple regular language. This paper describes the design and evolution of this control system, currently in use at SLAC and Padova on /spl sim/450 CPUs organized in nine farms.
NASA Astrophysics Data System (ADS)
Williams, C. J.; Pierson, F. B.; Robichaud, P. R.; Spaeth, K. E.; Hardegree, S. P.; Clark, P. E.; Moffet, C. A.; Al-Hamdan, O. Z.; Boll, J.
2010-12-01
Landscape-scale plant community transitions and altered fire regimes across Great Basin, USA, rangelands have increased the likelihood of post-fire flooding and erosion events. These hazards are particularly concerning for western urban centers along the rangeland urban-wildland interface where natural resources, property, and human life are at risk. Extensive conversion of 4-7 million hectares of Great Basin shrub-steppe to cheatgrass-dominated (Bromus tectorum) grasslands has increased the frequency and size of wildland fires within these ecosystems. Fire frequencies have increased by more than an order of magnitude and occur on 3-10 year intervals across much of the cheatgrass-dominated landscape. Extensive tree (Pinus spp. and Juniperus spp.) encroachment into wooded shrub-steppe has increased heavy fuel loads. Ladder fuels in these ecosystems promote rapidly spreading, high-intensity and severe ground-surface-crown fires. These altered fuel structures across much of the historical Great Basin shrub-steppe have initiated an upsurge in large rangeland wildfires and have increased the spatial and temporal vulnerability of these landscapes to amplified runoff and erosion. Resource and infrastructure damages, and loss of life have been reported due to flooding following recent large-scale burning of western rangelands and dry forests. We present a decade of post-fire rangeland hydrologic research that provides a foundation for conceptual modeling of the hydrologic impacts associated with an increased role of rangeland wildfires. We highlight advancements in predictive tools to address this large-scale phenomenon and discuss vital research voids requiring attention. Our geographic emphasis is the Great Basin Region, however, these concepts likely extend elsewhere given the increased role of fire in many geographic regions and across rangeland-to-forest ecotones in the western United States.
Developing eThread pipeline using SAGA-pilot abstraction for large-scale structural bioinformatics.
Ragothaman, Anjani; Boddu, Sairam Chowdary; Kim, Nayong; Feinstein, Wei; Brylinski, Michal; Jha, Shantenu; Kim, Joohyun
2014-01-01
While most of computational annotation approaches are sequence-based, threading methods are becoming increasingly attractive because of predicted structural information that could uncover the underlying function. However, threading tools are generally compute-intensive and the number of protein sequences from even small genomes such as prokaryotes is large typically containing many thousands, prohibiting their application as a genome-wide structural systems biology tool. To leverage its utility, we have developed a pipeline for eThread--a meta-threading protein structure modeling tool, that can use computational resources efficiently and effectively. We employ a pilot-based approach that supports seamless data and task-level parallelism and manages large variation in workload and computational requirements. Our scalable pipeline is deployed on Amazon EC2 and can efficiently select resources based upon task requirements. We present runtime analysis to characterize computational complexity of eThread and EC2 infrastructure. Based on results, we suggest a pathway to an optimized solution with respect to metrics such as time-to-solution or cost-to-solution. Our eThread pipeline can scale to support a large number of sequences and is expected to be a viable solution for genome-scale structural bioinformatics and structure-based annotation, particularly, amenable for small genomes such as prokaryotes. The developed pipeline is easily extensible to other types of distributed cyberinfrastructure.
Developing eThread Pipeline Using SAGA-Pilot Abstraction for Large-Scale Structural Bioinformatics
Ragothaman, Anjani; Feinstein, Wei; Jha, Shantenu; Kim, Joohyun
2014-01-01
While most of computational annotation approaches are sequence-based, threading methods are becoming increasingly attractive because of predicted structural information that could uncover the underlying function. However, threading tools are generally compute-intensive and the number of protein sequences from even small genomes such as prokaryotes is large typically containing many thousands, prohibiting their application as a genome-wide structural systems biology tool. To leverage its utility, we have developed a pipeline for eThread—a meta-threading protein structure modeling tool, that can use computational resources efficiently and effectively. We employ a pilot-based approach that supports seamless data and task-level parallelism and manages large variation in workload and computational requirements. Our scalable pipeline is deployed on Amazon EC2 and can efficiently select resources based upon task requirements. We present runtime analysis to characterize computational complexity of eThread and EC2 infrastructure. Based on results, we suggest a pathway to an optimized solution with respect to metrics such as time-to-solution or cost-to-solution. Our eThread pipeline can scale to support a large number of sequences and is expected to be a viable solution for genome-scale structural bioinformatics and structure-based annotation, particularly, amenable for small genomes such as prokaryotes. The developed pipeline is easily extensible to other types of distributed cyberinfrastructure. PMID:24995285
Large Animal Models for Foamy Virus Vector Gene Therapy
Trobridge, Grant D.; Horn, Peter A.; Beard, Brian C.; Kiem, Hans-Peter
2012-01-01
Foamy virus (FV) vectors have shown great promise for hematopoietic stem cell (HSC) gene therapy. Their ability to efficiently deliver transgenes to multi-lineage long-term repopulating cells in large animal models suggests they will be effective for several human hematopoietic diseases. Here, we review FV vector studies in large animal models, including the use of FV vectors with the mutant O6-methylguanine-DNA methyltransferase, MGMTP140K to increase the number of genetically modified cells after transplantation. In these studies, FV vectors have mediated efficient gene transfer to polyclonal repopulating cells using short ex vivo transduction protocols designed to minimize the negative effects of ex vivo culture on stem cell engraftment. In this regard, FV vectors appear superior to gammaretroviral vectors, which require longer ex vivo culture to effect efficient transduction. FV vectors have also compared favorably with lentiviral vectors when directly compared in the dog model. FV vectors have corrected leukocyte adhesion deficiency and pyruvate kinase deficiency in the dog large animal model. FV vectors also appear safer than gammaretroviral vectors based on a reduced frequency of integrants near promoters and also near proto-oncogenes in canine repopulating cells. Together, these studies suggest that FV vectors should be highly effective for several human hematopoietic diseases, including those that will require relatively high percentages of gene-modified cells to achieve clinical benefit. PMID:23223198
Natural hazards in the Alps triggered by ski slope engineering and artificial snow production
NASA Astrophysics Data System (ADS)
de Jong, C.
2012-04-01
In the Alps there is increasing concern of man-made triggering of natural hazards in association with ski slope engineering and pressures from climate change. However literature on the topic is rare. Ski run development has been intensified in the past decade to accommodate a higher density of skiers. In order to absorb the increased flux of skiers promoted by continually increasing lift capacity, ski runs are subject to more and more enlargement, straightening and leveling. This has required large-scale re-leveling of slopes with the removal of soil and protective vegetation using heavy machinery during the summer season. Slope-ward incision on steep slopes, creation of artificial embankments by leeward deposition and development of straight ski runs perpendicular to steep slopes have resulted in both shallow and deep erosion, gullying, triggering of small landslides and even bedload transport in marginal channels. Other natural hazards have been triggered directly or indirectly due to intensification of artificial snow production. This has increased exponentially in the last decade in order to secure the skiing season under increasingly warm temperatures and erratic snowfall and decreasing snow depth and snow duration in association with climate change. The consequences are multiple. Firstly, in order to economize both costs and quantity of artificial snow production, ski runs are leveled as far as possible in order to avoid topographical irregularities, protruding vegetation or rocks. The combination of topsoil removal and prolonged duration of artificial snow cover results in a decreased vegetation cover and period as well as species alteration. Together with greatly decreased permeability of the underground, snowmelt and intensive summer precipitation trigger surface runoff, erosion and even small landslides. After more than a decade of intensive cover by artificial snow, most such steep ski runs at altitudes above 1400 m are reduced into highly erosive, vegetation-poor scree slopes in summertime. Secondly, the production of artificial snow requires increasingly large quantities of water during low flow periods and causes an exponential increase in the construction of water reservoirs and pipelines. Such reservoirs are often constructed in depressions occupied by wetlands but also on slopes, hilltops and in proglacial locations at high altitudes up to 3000m. Reservoir construction removes vegetation, soil and regolith over surface areas of up to 150 000 m2 and depths of more than 20 m. During their construction, the temporary or permanent storage of large quantities of sediment on steep slopes has lead in several cases to the production of debris flows. Each reservoir requires road construction and vehicle parking areas for heavy weight vehicle access. These are frequently subject to erosion, gullying, and small landslides. Some reservoirs are vulnerable to catastrophic drainage triggered by earthquakes, avalanches and other natural hazards typical for mountain environments since they are only sealed with plastic membranes. Thirdly, the melt of artificial snow introduced by water transfers from other catchments can cause a relatively large local surplus of water which in turn increases spring and summer flood peaks as well as sediment transport. Most steep ski runs have introduced artificial drainage canals across the ski runs to avoid concentration of surface flow and to prevent erosion. Slopes are also covered with organic soils and re-vegetated where possible. However, given the present trends of intensification of use and precipitation extremes, it is unlikely that erosion and mass movements can be prevented in the next few decades for the duration of the amortization of investments.
Climate change induced transformations of agricultural systems: insights from a global model
NASA Astrophysics Data System (ADS)
Leclère, D.; Havlík, P.; Fuss, S.; Schmid, E.; Mosnier, A.; Walsh, B.; Valin, H.; Herrero, M.; Khabarov, N.; Obersteiner, M.
2014-12-01
Climate change might impact crop yields considerably and anticipated transformations of agricultural systems are needed in the coming decades to sustain affordable food provision. However, decision-making on transformational shifts in agricultural systems is plagued by uncertainties concerning the nature and geography of climate change, its impacts, and adequate responses. Locking agricultural systems into inadequate transformations costly to adjust is a significant risk and this acts as an incentive to delay action. It is crucial to gain insight into how much transformation is required from agricultural systems, how robust such strategies are, and how we can defuse the associated challenge for decision-making. While implementing a definition related to large changes in resource use into a global impact assessment modelling framework, we find transformational adaptations to be required of agricultural systems in most regions by 2050s in order to cope with climate change. However, these transformations widely differ across climate change scenarios: uncertainties in large-scale development of irrigation span in all continents from 2030s on, and affect two-thirds of regions by 2050s. Meanwhile, significant but uncertain reduction of major agricultural areas affects the Northern Hemisphere’s temperate latitudes, while increases to non-agricultural zones could be large but uncertain in one-third of regions. To help reducing the associated challenge for decision-making, we propose a methodology exploring which, when, where and why transformations could be required and uncertain, by means of scenario analysis.
Cine-servo lens technology for 4K broadcast and cinematography
NASA Astrophysics Data System (ADS)
Nurishi, Ryuji; Wakazono, Tsuyoshi; Usui, Fumiaki
2015-09-01
Central to the rapid evolution of 4K image capture technology in the past few years, deployment of large-format cameras with Super35mm Single Sensors is increasing in TV production for diverse shows such as dramas, documentaries, wildlife, and sports. While large format image capture has been the standard in the cinema world for quite some time, the recent experiences within the broadcast industry have revealed a variety of requirement differences for large format lenses compared to those of the cinema industry. A typical requirement for a broadcast lens is a considerably higher zoom ratio in order to avoid changing lenses in the middle of a live event, which is mostly not the case for traditional cinema productions. Another example is the need for compact size, light weight, and servo operability for a single camera operator shooting in a shoulder-mount ENG style. On the other hand, there are new requirements that are common to both worlds, such as smooth and seamless change in angle of view throughout the long zoom range, which potentially offers new image expression that never existed in the past. This paper will discuss the requirements from the two industries of cinema and broadcast, while at the same time introducing the new technologies and new optical design concepts applied to our latest "CINE-SERVO" lens series which presently consists of two models, CN7x17KAS-S and CN20x50IAS-H. It will further explain how Canon has realized 4K optical performance and fast servo control while simultaneously achieving compact size, light weight and high zoom ratio, by referring to patent-pending technologies such as the optical power layout, lens construction, and glass material combinations.
Large shift in source of fine sediment in the upper Mississippi River
Belmont, P.; Gran, K.B.; Schottler, S.P.; Wilcock, P.R.; Day, S.S.; Jennings, C.; Lauer, J.W.; Viparelli, E.; Willenbring, J.K.; Engstrom, D.R.; Parker, G.
2011-01-01
Although sediment is a natural constituent of rivers, excess loading to rivers and streams is a leading cause of impairment and biodiversity loss. Remedial actions require identification of the sources and mechanisms of sediment supply. This task is complicated by the scale and complexity of large watersheds as well as changes in climate and land use that alter the drivers of sediment supply. Previous studies in Lake Pepin, a natural lake on the Mississippi River, indicate that sediment supply to the lake has increased 10-fold over the past 150 years. Herein we combine geochemical fingerprinting and a suite of geomorphic change detection techniques with a sediment mass balance for a tributary watershed to demonstrate that, although the sediment loading remains very large, the dominant source of sediment has shifted from agricultural soil erosion to accelerated erosion of stream banks and bluffs, driven by increased river discharge. Such hydrologic amplification of natural erosion processes calls for a new approach to watershed sediment modeling that explicitly accounts for channel and floodplain dynamics that amplify or dampen landscape processes. Further, this finding illustrates a new challenge in remediating nonpoint sediment pollution and indicates that management efforts must expand from soil erosion to factors contributing to increased water runoff. ?? 2011 American Chemical Society.
Fischer, Stefan; Zöttl, Markus; Groenewoud, Frank; Taborsky, Barbara
2014-01-01
In cooperative breeding systems, dominant breeders sometimes tolerate unrelated individuals even if they inflict costs on the dominants. According to the ‘pay-to-stay’ hypothesis, (i) subordinates can outweigh these costs by providing help and (ii) dominants should be able to enforce help by punishing subordinates that provide insufficient help. This requires that dominants can monitor helping and can recognize group members individually. In a field experiment, we tested whether cooperatively breeding cichlid Neolamprologus pulcher subordinates increase their help after a forced ‘idle’ period, how other group members respond to a previously idle helper, and how helper behaviour and group responses depend on group size. Previously, idle helpers increased their submissiveness and received more aggression than control helpers, suggesting that punishment occurred to enforce help. Subordinates in small groups increased their help more than those in large groups, despite receiving less aggression. When subordinates were temporarily removed, dominants in small groups were more likely to evict returning subordinates. Our results suggest that only in small groups do helpers face a latent threat of punishment by breeders as predicted by the pay-to-stay hypothesis. In large groups, cognitive constraints may prevent breeders from tracking the behaviour of a large number of helpers. PMID:24990673