Sample records for alamos nisp monte

  1. The EUCLID NISP tolerancing concept and results

    NASA Astrophysics Data System (ADS)

    Grupp, Frank; Prieto, Eric; Geis, Norbert; Bode, Andreas; Katterloher, Reinhard; Bodendorf, Christof; Penka, Daniela; Bender, Ralf

    2014-08-01

    Within ESAs 2015 - 2025 Cosmic Vision framework the EUCLID mission satellite addresses cosmological questions related to dark matter and dark energy. EUCLID is equipped with two instruments that are simultaneously observing patches of > 0.5 square degree on the sky. The VIS visual light high spacial resolution imager and the NISP near infrared spectrometer and photometer are separated by a di-chroic beam splitter. Having a large FoV (larger than the full moon disk), together with high demands on the optical performance and strong requirements on in flight stability lead to very challenging demands on alignment and post launch { post cool-down optical element position. The role of an accurate and trust-worthy tolerance analysis which is well adopted to the stepwise integration and alignment concept, as well as to the missions stability properties is therefore crucial for the missions success. With this paper we present a new iteration of the baseline tolerancing concept for EUCLID NISP. All 7 operational modes being low resolution slit-less spectroscopy and three band Y, J& H+ band photometry are being toleranced together. During the design process it was noted that the desired performance can only be reached when alignment and tolerancing methods are closely connected and optimized together. Utilizing computer generated - multi zone - holograms to align and cross reference the four lenses of the NISP optical system. We show our plan to verify these holograms and what alignment sensitivities we reach. In the main section we present the result of the tolerancing and the main contributers that drive the mechanical and thermal design of the NISO optical subsystems. This analysis presents the design status of NISP at the system PDR of the mission.

  2. EGSE customization for the Euclid NISP Instrument AIV/AIT activities

    NASA Astrophysics Data System (ADS)

    Franceschi, E.; Trifoglio, M.; Gianotti, F.; Conforti, V.; Andersen, J. J.; Stephen, J. B.; Valenziano, L.; Auricchio, N.; Bulgarelli, A.; De Rosa, A.; Fioretti, V.; Maiorano, E.; Morgante, G.; Nicastro, L.; Sortino, F.; Zoli, A.; Balestra, A.; Bonino, D.; Bonoli, C.; Bortoletto, F.; Capobianco, V.; Corcione, L.; Dal Corso, F.; Debei, S.; Di Ferdinando, D.; Dusini, S.; Farinelli, R.; Fornari, F.; Giacomini, F.; Guizzo, G. P.; Laudisio, F.; Ligori, S.; Mauri, N.; Medinaceli, E.; Patrizii, L.; Sirignano, C.; Sirri, G.; Stanco, L.; Tenti, M.; Valieri, C.; Ventura, S.

    2016-07-01

    The Near Infrared Spectro-Photometer (NISP) on board the Euclid ESA mission will be developed and tested at various levels of integration by using various test equipment. The Electrical Ground Support Equipment (EGSE) shall be required to support the assembly, integration, verification and testing (AIV/AIT) and calibration activities at instrument level before delivery to ESA, and at satellite level, when the NISP instrument is mounted on the spacecraft. In the case of the Euclid mission this EGSE will be provided by ESA to NISP team, in the HW/SW framework called "CCS Lite", with a possible first usage already during the Warm Electronics (WE) AIV/AIT activities. In this paper we discuss how we will customize that "CCS Lite" as required to support both the WE and Instrument test activities. This customization will primarily involve building the NISP Mission Information Base (the CCS MIB tables) by gathering the relevant data from the instrument sub-units and validating these inputs through specific tools. Secondarily, it will imply developing a suitable set of test sequences, by using uTOPE (an extension to the TCL scripting language, included in the CCS framework), in order to implement the foreseen test procedures. In addition and in parallel, custom interfaces shall be set up between the CCS and the NI-IWS (the NISP Instrument Workstation, which will be in use at any level starting from the WE activities), and also between the CCS and the TCC (the Telescope Control and command Computer, to be only and specifically used during the instrument level tests).

  3. Instrument workstation for the EGSE of the Near Infrared Spectro-Photometer instrument (NISP) of the EUCLID mission

    NASA Astrophysics Data System (ADS)

    Trifoglio, M.; Gianotti, F.; Conforti, V.; Franceschi, E.; Stephen, J. B.; Bulgarelli, A.; Fioretti, V.; Maiorano, E.; Nicastro, L.; Valenziano, L.; Zoli, A.; Auricchio, N.; Balestra, A.; Bonino, D.; Bonoli, C.; Bortoletto, F.; Capobianco, V.; Chiarusi, T.; Corcione, L.; Debei, S.; De Rosa, A.; Dusini, S.; Fornari, F.; Giacomini, F.; Guizzo, G. P.; Ligori, S.; Margiotta, A.; Mauri, N.; Medinaceli, E.; Morgante, G.; Patrizii, L.; Sirignano, C.; Sirri, G.; Sortino, F.; Stanco, L.; Tenti, M.

    2016-07-01

    The NISP instrument on board the Euclid ESA mission will be developed and tested at different levels of integration using various test equipment which shall be designed and procured through a collaborative and coordinated effort. The NISP Instrument Workstation (NI-IWS) will be part of the EGSE configuration that will support the NISP AIV/AIT activities from the NISP Warm Electronics level up to the launch of Euclid. One workstation is required for the NISP EQM/AVM, and a second one for the NISP FM. Each workstation will follow the respective NISP model after delivery to ESA for Payload and Satellite AIV/AIT and launch. At these levels the NI-IWS shall be configured as part of the Payload EGSE, the System EGSE, and the Launch EGSE, respectively. After launch, the NI-IWS will be also re-used in the Euclid Ground Segment in order to support the Commissioning and Performance Verification (CPV) phase, and for troubleshooting purposes during the operational phase. The NI-IWS is mainly aimed at the local storage in a suitable format of the NISP instrument data and metadata, at local retrieval, processing and display of the stored data for on-line instrument assessment, and at the remote retrieval of the stored data for off-line analysis on other computers. We describe the design of the IWS software that will create a suitable interface to the external systems in each of the various configurations envisaged at the different levels, and provide the capabilities required to monitor and verify the instrument functionalities and performance throughout all phases of the NISP lifetime.

  4. Focal plane mechanical design of the NISP/Euclid instrument

    NASA Astrophysics Data System (ADS)

    Bonnefoi, Anne; Bon, William; Niclas, Mathieu; Solheim, Bjarte G. B.; Torvanger, Oyvind; Schistad, Robert; Foulon, Benjamin; Garcia, José; Vives, Sébastien

    2016-07-01

    Currently in phase C, the Euclid mission selected by ESA in the Cosmic Vision program is dedicated to understand dark energy and dark matter. NISP (standing for Near Infrared Spectro-Photometer) is one of the two instruments of the mission. NISP will combine a photometer and a spectrometer working in the near-IR (0.9-2 microns). Its detection subsystem (called NI-DS) is based on a mosaic of 16 IR detectors cooled down to 90K which are supported by a molybdenum plate. The front-end readout electronics (working at 130K) are supported by another structure in Aluminum. The NI-DS is mounted on the rest of the instrument thanks to a panel in Silicon Carbide (SiC). Finally an optical baffle in Titanium will prevent the rogue light to reach the detectors. On top of the complexity due to the wide range of temperatures and the various materials imposed at the interfaces; the NI-DS has also to incorporate an internal adjustment capability of the position of the focal plane in tip/tilt and focus. This article will present current status of the development of the detection system of NISP.

  5. On-board data processing for the near infrared spectrograph and photometer instrument (NISP) of the EUCLID mission

    NASA Astrophysics Data System (ADS)

    Bonoli, Carlotta; Balestra, Andrea; Bortoletto, Favio; D'Alessandro, Maurizio; Farinelli, Ruben; Medinaceli, Eduardo; Stephen, John; Borsato, Enrico; Dusini, Stefano; Laudisio, Fulvio; Sirignano, Chiara; Ventura, Sandro; Auricchio, Natalia; Corcione, Leonardo; Franceschi, Enrico; Ligori, Sebastiano; Morgante, Gianluca; Patrizii, Laura; Sirri, Gabriele; Trifoglio, Massimo; Valenziano, Luca

    2016-07-01

    The Near Infrared Spectrograph and Photometer (NISP) is one of the two instruments on board the EUCLID mission now under implementation phase; VIS, the Visible Imager is the second instrument working on the same shared optical beam. The NISP focal plane is based on a detector mosaic deploying 16x, 2048x2048 pixels^2 HAWAII-II HgCdTe detectors, now in advanced delivery phase from Teledyne Imaging Scientific (TIS), and will provide NIR imaging in three bands (Y, J, H) plus slit-less spectroscopy in the range 0.9÷2.0 micron. All the NISP observational modes will be supported by different parametrization of the classic multi-accumulation IR detector readout mode covering the specific needs for spectroscopic, photometric and calibration exposures. Due to the large number of deployed detectors and to the limited satellite telemetry available to ground, a consistent part of the data processing, conventionally performed off-line, will be accomplished on board, in parallel with the flow of data acquisitions. This has led to the development of a specific on-board, HW/SW, data processing pipeline, and to the design of computationally performing control electronics, suited to cope with the time constraints of the NISP acquisition sequences during the sky survey. In this paper we present the architecture of the NISP on-board processing system, directly interfaced to the SIDECAR ASICs system managing the detector focal plane, and the implementation of the on-board pipe-line allowing all the basic operations of input frame averaging, final frame interpolation and data-volume compression before ground down-link.

  6. Space-borne survey instrument operations: lessons learned and new concepts for the Euclid NISP instrument

    NASA Astrophysics Data System (ADS)

    Valenziano, L.; Gregorio, A.; Butler, R. C.; Amiaux, J.; Bonoli, C.; Bortoletto, F.; Burigana, C.; Corcione, L.; Ealet, A.; Frailis, M.; Jahnke, K.; Ligori, S.; Maiorano, E.; Morgante, G.; Nicastro, L.; Pasian, F.; Riva, M.; Scaramella, R.; Schiavone, F.; Tavagnacco, D.; Toledo-Moreo, R.; Trifoglio, M.; Zacchei, A.; Zerbi, F. M.; Maciaszek, T.

    2012-09-01

    Euclid is the future ESA mission, mainly devoted to Cosmology. Like WMAP and Planck, it is a survey mission, to be launched in 2019 and injected in orbit far away from the Earth, for a nominal lifetime of 7 years. Euclid has two instruments on-board, the Visible Imager (VIS) and the Near- Infrared Spectro-Photometer (NISP). The NISP instrument includes cryogenic mechanisms, active thermal control, high-performance Data Processing Unit and requires periodic in-flight calibrations and instrument parameters monitoring. To fully exploit the capability of the NISP, a careful control of systematic effects is required. From previous experiments, we have built the concept of an integrated instrument development and verification approach, where the scientific, instrument and ground-segment expertise have strong interactions from the early phases of the project. In particular, we discuss the strong integration of test and calibration activities with the Ground Segment, starting from early pre-launch verification activities. We want to report here the expertise acquired by the Euclid team in previous missions, only citing the literature for detailed reference, and indicate how it is applied in the Euclid mission framework.

  7. Design of the on-board application software for the instrument control unit of Euclid-NISP

    NASA Astrophysics Data System (ADS)

    Ligori, Sebastiano; Corcione, Leonardo; Capobianco, Vito; Valenziano, Luca

    2014-08-01

    In this paper we describe the main requirements driving the development of the Application software of the ICU of NISP, the Near-Infrared Spectro-Photometer of the Euclid mission. This software will be based on a real-time operating system and will interface with all the subunits of NISP, as well as the CMDU of the spacecraft for the Telecommand and Housekeeping management. We briefly detail the services (following the PUS standard) that will be made available, and also possible commonalities in the approach with the ASW of the VIS CDPU, which could make the development effort more efficient; this approach could also make easier the maintenance of the SW during the mission. The development plan of the ASW and the next milestones foreseen are described, together with the architectural design approach and the development environment we are setting up.

  8. EUCLID/NISP GRISM qualification model AIT/AIV campaign: optical, mechanical, thermal and vibration tests

    NASA Astrophysics Data System (ADS)

    Caillat, A.; Costille, A.; Pascal, S.; Rossin, C.; Vives, S.; Foulon, B.; Sanchez, P.

    2017-09-01

    Dark matter and dark energy mysteries will be explored by the Euclid ESA M-class space mission which will be launched in 2020. Millions of galaxies will be surveyed through visible imagery and NIR imagery and spectroscopy in order to map in three dimensions the Universe at different evolution stages over the past 10 billion years. The massive NIR spectroscopic survey will be done efficiently by the NISP instrument thanks to the use of grisms (for "Grating pRISMs") developed under the responsibility of the LAM. In this paper, we present the verification philosophy applied to test and validate each grism before the delivery to the project. The test sequence covers a large set of verifications: optical tests to validate efficiency and WFE of the component, mechanical tests to validate the robustness to vibration, thermal tests to validate its behavior in cryogenic environment and a complete metrology of the assembled component. We show the test results obtained on the first grism Engineering and Qualification Model (EQM) which will be delivered to the NISP project in fall 2016.

  9. Detailed design and first tests of the application software for the instrument control unit of Euclid-NISP

    NASA Astrophysics Data System (ADS)

    Ligori, S.; Corcione, L.; Capobianco, V.; Bonino, D.; Sirri, G.; Fornari, F.; Giacomini, F.; Patrizii, L.; Valenziano, L.; Travaglini, R.; Colodro, C.; Bortoletto, F.; Bonoli, C.; Chiarusi, T.; Margiotta, A.; Mauri, N.; Pasqualini, L.; Spurio, M.; Tenti, M.; Dal Corso, F.; Dusini, S.; Laudisio, F.; Sirignano, C.; Stanco, L.; Ventura, S.; Auricchio, N.; Balestra, A.; Franceschi, E.; Morgante, G.; Trifoglio, M.; Medinaceli, E.; Guizzo, G. P.; Debei, S.; Stephen, J. B.

    2016-07-01

    In this paper we describe the detailed design of the application software (ASW) of the instrument control unit (ICU) of NISP, the Near-Infrared Spectro-Photometer of the Euclid mission. This software is based on a real-time operating system (RTEMS) and will interface with all the subunits of NISP, as well as the command and data management unit (CDMU) of the spacecraft for telecommand and housekeeping management. We briefly review the main requirements driving the design and the architecture of the software that is approaching the Critical Design Review level. The interaction with the data processing unit (DPU), which is the intelligent subunit controlling the detector system, is described in detail, as well as the concept for the implementation of the failure detection, isolation and recovery (FDIR) algorithms. The first version of the software is under development on a Breadboard model produced by AIRBUS/CRISA. We describe the results of the tests and the main performances and budgets.

  10. Monte Carlo: in the beginning and some great expectations

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Metropolis, N.

    1985-01-01

    The central theme will be on the historical setting and origins of the Monte Carlo Method. The scene was post-war Los Alamos Scientific Laboratory. There was an inevitability about the Monte Carlo Event: the ENIAC had recently enjoyed its meteoric rise (on a classified Los Alamos problem); Stan Ulam had returned to Los Alamos; John von Neumann was a frequent visitor. Techniques, algorithms, and applications developed rapidly at Los Alamos. Soon, the fascination of the Method reached wider horizons. The first paper was submitted for publication in the spring of 1949. In the summer of 1949, the first open conferencemore » was held at the University of California at Los Angeles. Of some interst perhaps is an account of Fermi's earlier, independent application in neutron moderation studies while at the University of Rome. The quantum leap expected with the advent of massively parallel processors will provide stimuli for very ambitious applications of the Monte Carlo Method in disciplines ranging from field theories to cosmology, including more realistic models in the neurosciences. A structure of multi-instruction sets for parallel processing is ideally suited for the Monte Carlo approach. One may even hope for a modest hardening of the soft sciences.« less

  11. On-ground tests of the NISP infrared spectrometer instrument for Euclid

    NASA Astrophysics Data System (ADS)

    Jomni, Cyril; Ealet, Anne; Gillard, William; Prieto, Éric; Grupp, Frank U.

    2017-09-01

    Euclid is an ESA mission dedicated to understand the acceleration of the expansion of the Universe. The mission will measure hundred of millions of galaxies in spectrophotometry and photometry in the near infrared thanks to a spectro-photometer called NISP. This instrument will be assembled and tested in Marseille. To prepare the on-ground test plan and develop the test procedure, we have used simulated PSF images, based on a Zemax optical design of the instrument. We have developed the analysis tools that will be further used to build the procedure verification. We present here the method and analysis results to adjust the focus of the instrument. We will in particular show that because of the sampling of the PSF, a dithering strategy should be adapted and will constraint the development of the test plan.

  12. SABRINA - An interactive geometry modeler for MCNP (Monte Carlo Neutron Photon)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    West, J.T.; Murphy, J.

    SABRINA is an interactive three-dimensional geometry modeler developed to produce complicated models for the Los Alamos Monte Carlo Neutron Photon program MCNP. SABRINA produces line drawings and color-shaded drawings for a wide variety of interactive graphics terminals. It is used as a geometry preprocessor in model development and as a Monte Carlo particle-track postprocessor in the visualization of complicated particle transport problem. SABRINA is written in Fortran 77 and is based on the Los Alamos Common Graphics System, CGS. 5 refs., 2 figs.

  13. A Wildfire Behavior Modeling System at Los Alamos National Laboratory for Operational Applications

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    S.W. Koch; R.G.Balice

    2004-11-01

    To support efforts to protect facilities and property at Los Alamos National Laboratory from damages caused by wildfire, we completed a multiyear project to develop a system for modeling the behavior of wildfires in the Los Alamos region. This was accomplished by parameterizing the FARSITE wildfire behavior model with locally gathered data representing topography, fuels, and weather conditions from throughout the Los Alamos region. Detailed parameterization was made possible by an extensive monitoring network of permanent plots, weather towers, and other data collection facilities. We also incorporated a database of lightning strikes that can be used individually as repeatable ignitionmore » points or can be used as a group in Monte Carlo simulation exercises and in other randomization procedures. The assembled modeling system was subjected to sensitivity analyses and was validated against documented fires, including the Cerro Grande Fire. The resulting modeling system is a valuable tool for research and management. It also complements knowledge based on professional expertise and information gathered from other modeling technologies. However, the modeling system requires frequent updates of the input data layers to produce currently valid results, to adapt to changes in environmental conditions within the Los Alamos region, and to allow for the quick production of model outputs during emergency operations.« less

  14. Oral vaccine of Lactococcus lactis harbouring pandemic H1N1 2009 haemagglutinin1 and nisP anchor fusion protein elevates anti-HA1 sIgA levels in mice.

    PubMed

    Joan, Stella Siaw Xiu; Pui-Fong, Jee; Song, Adelene Ai-Lian; Chang, Li-Yen; Yusoff, Khatijah; AbuBakar, Sazaly; Rahim, Raha Abdul

    2016-05-01

    An oral lactococcal-based vaccine which haboured the haemagglutinin1 (HA1) antigen fused to nisP anchor protein for the purpose of surface displaying the HA1 antigen was developed against H1N1 virus. Recombinant L. lactis strains expressed HA1-nisP fusion proteins when induced with nisin, as confirmed through western blotting. However, immunofluorescense did not detect any surface-displayed proteins, suggesting that the protein was either unsuccessfully translocated or improperly displayed. Despite this, oral administration of recombinant L. lactis strains to BALB/c mice revealed that significant levels of anti-HA1 sIgA antibodies were detected in mice fecal suspension samples of mice group NZ9000 (pNZ:HN) when compared to the negative control NZ9000 (pNZ8048) group. Specific anti-HA1 sIgA antibodies were locally produced and live recombinant lactococcal vaccine was able to elicit humoral response of BALB/c mice despite unsuccessful surface display of the HA1 epitope.

  15. Los Alamos Science Facilities

    Science.gov Websites

    Los Alamos National Laboratory Search Site submit About Mission Business Newsroom Publications Los Innovation in New Mexico Los Alamos Collaboration for Explosives Detection (LACED) SensorNexus Exascale Computing Project (ECP) User Facilities Center for Integrated Nanotechnologies (CINT) Los Alamos Neutron

  16. Living in Los Alamos

    Science.gov Websites

    Los Alamos National Laboratory Search Site submit About Mission Business Newsroom Publications Los Innovation in New Mexico Los Alamos Collaboration for Explosives Detection (LACED) SensorNexus Exascale Computing Project (ECP) User Facilities Center for Integrated Nanotechnologies (CINT) Los Alamos Neutron

  17. Metis: A Pure Metropolis Markov Chain Monte Carlo Bayesian Inference Library

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bates, Cameron Russell; Mckigney, Edward Allen

    The use of Bayesian inference in data analysis has become the standard for large scienti c experiments [1, 2]. The Monte Carlo Codes Group(XCP-3) at Los Alamos has developed a simple set of algorithms currently implemented in C++ and Python to easily perform at-prior Markov Chain Monte Carlo Bayesian inference with pure Metropolis sampling. These implementations are designed to be user friendly and extensible for customization based on speci c application requirements. This document describes the algorithmic choices made and presents two use cases.

  18. Stockpile Stewardship: Los Alamos

    ScienceCinema

    McMillan, Charlie; Morgan, Nathanial; Goorley, Tom; Merrill, Frank; Funk, Dave; Korzekwa, Deniece; Laintz, Ken

    2018-01-16

    "Heritage of Science" is a short video that highlights the Stockpile Stewardship program at Los Alamos National Laboratory. Stockpile Stewardship was conceived in the early 1990s as a national science-based program that could assure the safety, security, and effectiveness of the U.S. nuclear deterrent without the need for full-scale underground nuclear testing. This video was produced by Los Alamos National Laboratory for screening at the Lab's Bradbury Science Museum in Los Alamos, NM and is narrated by science correspondent Miles O'Brien.

  19. New Mexico: Los Alamos

    Atmospheric Science Data Center

    2014-05-15

    article title:  Los Alamos, New Mexico     View Larger JPEG image ... kb) Multi-angle views of the Fire in Los Alamos, New Mexico, May 9, 2000. These true-color images covering north-central New Mexico ...

  20. MCNP (Monte Carlo Neutron Photon) capabilities for nuclear well logging calculations

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Forster, R.A.; Little, R.C.; Briesmeister, J.F.

    The Los Alamos Radiation Transport Code System (LARTCS) consists of state-of-the-art Monte Carlo and discrete ordinates transport codes and data libraries. The general-purpose continuous-energy Monte Carlo code MCNP (Monte Carlo Neutron Photon), part of the LARTCS, provides a computational predictive capability for many applications of interest to the nuclear well logging community. The generalized three-dimensional geometry of MCNP is well suited for borehole-tool models. SABRINA, another component of the LARTCS, is a graphics code that can be used to interactively create a complex MCNP geometry. Users can define many source and tally characteristics with standard MCNP features. The time-dependent capabilitymore » of the code is essential when modeling pulsed sources. Problems with neutrons, photons, and electrons as either single particle or coupled particles can be calculated with MCNP. The physics of neutron and photon transport and interactions is modeled in detail using the latest available cross-section data. A rich collections of variance reduction features can greatly increase the efficiency of a calculation. MCNP is written in FORTRAN 77 and has been run on variety of computer systems from scientific workstations to supercomputers. The next production version of MCNP will include features such as continuous-energy electron transport and a multitasking option. Areas of ongoing research of interest to the well logging community include angle biasing, adaptive Monte Carlo, improved discrete ordinates capabilities, and discrete ordinates/Monte Carlo hybrid development. Los Alamos has requested approval by the Department of Energy to create a Radiation Transport Computational Facility under their User Facility Program to increase external interactions with industry, universities, and other government organizations. 21 refs.« less

  1. Science and Innovation at Los Alamos

    Science.gov Websites

    Los Alamos National Laboratory Search Site submit About Mission Business Newsroom Publications Los Innovation in New Mexico Los Alamos Collaboration for Explosives Detection (LACED) SensorNexus Exascale Computing Project (ECP) User Facilities Center for Integrated Nanotechnologies (CINT) Los Alamos Neutron

  2. Los Alamos Climatology 2016 Update

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bruggeman, David Alan

    The Los Alamos National Laboratory (LANL or the Laboratory) operates a meteorology monitoring network to support LANL emergency response, engineering designs, environmental compliance, environmental assessments, safety evaluations, weather forecasting, environmental monitoring, research programs, and environmental restoration. Weather data has been collected in Los Alamos since 1910. Bowen (1990) provided climate statistics (temperature and precipitation) for the 1961– 1990 averaging period, and included other analyses (e.g., wind and relative humidity) based on the available station locations and time periods. This report provides an update to the 1990 publication Los Alamos Climatology (Bowen 1990).

  3. Los Alamos offers Fellowships

    NASA Astrophysics Data System (ADS)

    Los Alamos National Laboratory in New Mexico is calling for applications for postdoctoral appointments and research fellowships. The positions are available in geoscience as well as other scientific disciplines.The laboratory, which is operated by the University of California for the Department of Energy, awards J. Robert Oppenheimer Research Fellowships to scientists that either have or will soon complete doctoral degrees. The appointments are for two years, are renewable for a third year, and carry a stipend of $51,865 per year. Potential applicants should send a resume or employment application and a statement of research goals to Carol M. Rich, Div. 89, Human Resources Development Division, MS P290, Los Alamos National Laboratory, Los Alamos, New Mexico 87545 by mid-November.

  4. Numerical integration of detector response functions via Monte Carlo simulations

    NASA Astrophysics Data System (ADS)

    Kelly, K. J.; O'Donnell, J. M.; Gomez, J. A.; Taddeucci, T. N.; Devlin, M.; Haight, R. C.; White, M. C.; Mosby, S. M.; Neudecker, D.; Buckner, M. Q.; Wu, C. Y.; Lee, H. Y.

    2017-09-01

    Calculations of detector response functions are complicated because they include the intricacies of signal creation from the detector itself as well as a complex interplay between the detector, the particle-emitting target, and the entire experimental environment. As such, these functions are typically only accessible through time-consuming Monte Carlo simulations. Furthermore, the output of thousands of Monte Carlo simulations can be necessary in order to extract a physics result from a single experiment. Here we describe a method to obtain a full description of the detector response function using Monte Carlo simulations. We also show that a response function calculated in this way can be used to create Monte Carlo simulation output spectra a factor of ∼ 1000 × faster than running a new Monte Carlo simulation. A detailed discussion of the proper treatment of uncertainties when using this and other similar methods is provided as well. This method is demonstrated and tested using simulated data from the Chi-Nu experiment, which measures prompt fission neutron spectra at the Los Alamos Neutron Science Center.

  5. Numerical integration of detector response functions via Monte Carlo simulations

    DOE PAGES

    Kelly, Keegan John; O'Donnell, John M.; Gomez, Jaime A.; ...

    2017-06-13

    Calculations of detector response functions are complicated because they include the intricacies of signal creation from the detector itself as well as a complex interplay between the detector, the particle-emitting target, and the entire experimental environment. As such, these functions are typically only accessible through time-consuming Monte Carlo simulations. Furthermore, the output of thousands of Monte Carlo simulations can be necessary in order to extract a physics result from a single experiment. Here we describe a method to obtain a full description of the detector response function using Monte Carlo simulations. We also show that a response function calculated inmore » this way can be used to create Monte Carlo simulation output spectra a factor of ~1000× faster than running a new Monte Carlo simulation. A detailed discussion of the proper treatment of uncertainties when using this and other similar methods is provided as well. Here, this method is demonstrated and tested using simulated data from the Chi-Nu experiment, which measures prompt fission neutron spectra at the Los Alamos Neutron Science Center.« less

  6. Numerical integration of detector response functions via Monte Carlo simulations

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kelly, Keegan John; O'Donnell, John M.; Gomez, Jaime A.

    Calculations of detector response functions are complicated because they include the intricacies of signal creation from the detector itself as well as a complex interplay between the detector, the particle-emitting target, and the entire experimental environment. As such, these functions are typically only accessible through time-consuming Monte Carlo simulations. Furthermore, the output of thousands of Monte Carlo simulations can be necessary in order to extract a physics result from a single experiment. Here we describe a method to obtain a full description of the detector response function using Monte Carlo simulations. We also show that a response function calculated inmore » this way can be used to create Monte Carlo simulation output spectra a factor of ~1000× faster than running a new Monte Carlo simulation. A detailed discussion of the proper treatment of uncertainties when using this and other similar methods is provided as well. Here, this method is demonstrated and tested using simulated data from the Chi-Nu experiment, which measures prompt fission neutron spectra at the Los Alamos Neutron Science Center.« less

  7. Los Alamos National Lab: National Security Science

    Science.gov Websites

    SKIP TO PAGE CONTENT Los Alamos National Laboratory Delivering science and technology to protect Permit for Storm Water Public Reading Room Environment Home News Los Alamos National Lab: National deposition operations for the Center for Integrated Nanotechnologies at Los Alamos. Innovation drives his

  8. Performance of the New Los Alamos UCN Source and Implications for Future Experiments

    NASA Astrophysics Data System (ADS)

    Makela, Mark; LANL UCN Team

    2017-01-01

    The Los Alamos Ultracold Neutron (UCN) source was replaced during this past summer and has been commissioned during the last few months. The new source is the result of lessons learned during the 10 year operation of the first UCN source and extensive Monte Carlo analysis. The new source is a spallation driven source based on a solid deuterium UCN moderator similar the previous one. This talk will present an overview of the new source design and the results of commissioning tests. The talk will conclude with a brief overview of the implications of source performance on the neutron lifetime and LANL nEDM experiments. This work was funded by LANL LDRD.

  9. Water Supply at Los Alamos 1998-2001

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Richard J. Koch; David B. Rogers

    2003-03-01

    For the period 1998 through 2001, the total water used at Los Alamos from all sources ranged from 1325 million gallons (Mg) in 1999 to 1515 Mg in 2000. Groundwater production ranged from 1323 Mg in 1999 to 1506 Mg in 2000 from the Guaje, Pajarito, and Otowi fields. Nonpotable surface water used from Los Alamos reservoir ranged from zero gallons in 2001 to 9.3 Mg in 2000. For years 1998 through 2001, over 99% of all water used at Los Alamos was groundwater. Water use by Los Alamos National Laboratory (LANL) between 1998 and 2001 ranged from 379 Mgmore » in 2000 to 461 Mg in 1998. The LANL water use in 2001 was 393 Mg or 27% of the total water use at Los Alamos. Water use by Los Alamos County ranged from 872 Mg in 1999 to 1137 Mg in 2000, and averaged 1006 Mg/yr. Four new replacement wells in the Guaje field (G-2A, G-3A, G-4A, and G-5A) were drilled in 1998 and began production in 1999; with existing well G-1A, the Guaje field currently has five producing wells. Five of the old Guaje wells (G-1, G-2, G-4, G-5, and G-6) were plugged and abandoned in 1999, and one well (G-3) was abandoned but remains as an observation well for the Guaje field. The long-term water level observations in production and observation (test) wells at Los Alamos are consistent with the formation of a cone of depression in response to water production. The water level decline is gradual and at most has been about 0.7 to 2 ft per year for production wells and from 0.4 to 0.9 ft/yr for observation (test) wells. The largest water level declines have been in the Guaje field where nonpumping water levels were about 91 ft lower in 2001 than in 1951. The initial water levels of the Guaje replacement wells were 32 to 57 ft lower than the initial water levels of adjacent original Guaje wells. When production wells are taken off-line for pump replacement or repair, water levels have returned to within about 25 ft of initial static levels within 6 to 12 months. Thus, the water-level trends suggest no

  10. Los Alamos radiation transport code system on desktop computing platforms

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Briesmeister, J.F.; Brinkley, F.W.; Clark, B.A.

    The Los Alamos Radiation Transport Code System (LARTCS) consists of state-of-the-art Monte Carlo and discrete ordinates transport codes and data libraries. These codes were originally developed many years ago and have undergone continual improvement. With a large initial effort and continued vigilance, the codes are easily portable from one type of hardware to another. The performance of scientific work-stations (SWS) has evolved to the point that such platforms can be used routinely to perform sophisticated radiation transport calculations. As the personal computer (PC) performance approaches that of the SWS, the hardware options for desk-top radiation transport calculations expands considerably. Themore » current status of the radiation transport codes within the LARTCS is described: MCNP, SABRINA, LAHET, ONEDANT, TWODANT, TWOHEX, and ONELD. Specifically, the authors discuss hardware systems on which the codes run and present code performance comparisons for various machines.« less

  11. Los Alamos Fires From Landsat 7

    NASA Technical Reports Server (NTRS)

    2002-01-01

    On May 9, 2000, the Landsat 7 satellite acquired an image of the area around Los Alamos, New Mexico. The Landsat 7 satellite acquired this image from 427 miles in space through its sensor called the Enhanced Thematic Mapper Plus (ETM+). Evident within the imagery is a view of the ongoing Cerro Grande fire near the town of Los Alamos and the Los Alamos National Laboratory. Combining the high-resolution (30 meters per pixel in this scene) imaging capacity of ETM+ with its multi-spectral capabilities allows scientists to penetrate the smoke plume and see the structure of the fire on the surface. Notice the high-level of detail in the infrared image (bottom), in which burn scars are clearly distinguished from the hotter smoldering and flaming parts of the fire. Within this image pair several features are clearly visible, including the Cerro Grande fire and smoke plume, the town of Los Alamos, the Los Alamos National Laboratory and associated property, and Cerro Grande peak. Combining ETM+ channels 7, 4, and 2 (one visible and two infrared channels) results in a false color image where vegetation appears as bright to dark green (bottom image). Forested areas are generally dark green while herbaceous vegetation is light green. Rangeland or more open areas appear pink to light purple. Areas with extensive pavement or urban development appear light blue or white to purple. Less densely-developed residential areas appear light green and golf courses are very bright green. The areas recently burned appear black. Dark red to bright red patches, or linear features within the burned area, are the hottest and possibly actively burning areas of the fire. The fire is spreading downslope and the front of the fire is readily detectable about 2 kilometers to the west and south of Los Alamos. Combining ETM+ channels 3, 2, and 1 provides a true-color image of the greater Los Alamos region (top image). Vegetation is generally dark to medium green. Forested areas are very dark green

  12. Internship at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Dunham, Ryan Q.

    2012-07-11

    Los Alamos National Laboratory (LANL) is located in Los Alamos, New Mexico. It provides support for our country's nuclear weapon stockpile as well as many other scientific research projects. I am an Undergraduate Student Intern in the Systems Design and Analysis group within the Nuclear Nonproliferation division of the Global Security directorate at LANL. I have been tasked with data analysis and modeling of particles in a fluidized bed system for the capture of carbon dioxide from power plant flue gas.

  13. Publications of Los Alamos Research, 1983

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sheridan, C.J.; McClary, W.J.; Rich, J.A.

    1984-10-01

    This bibliography is a compilation of unclassified publications of work done at the Los Alamos National Laboratory for 1983. Papers published in 1982 are included regardless of when they were actually written. Publications received too late for inclusion in earlier compilations have also been listed. Declassification of previously classified reports is considered to constitute publication. All classified issuances are omitted - even those papers, themselves unclassified, which were published only as part of a classified document. If a paper was published more than once, all places of publication are included. The bibliography includes Los Alamos National Laboratory reports, papers releasedmore » as non-Laboratory reports, journal articles, books, chapters of books, conference papers either published separately or as part of conference proceedings issued as books or reports, papers publishd in congressional hearings, theses, and US patents. Publications by Los Alamos authors that are not records of Laboratory-sponsored work are included when the Library becomes aware of them.« less

  14. Airport-Noise Levels and Annoyance Model (ALAMO) system's reference manual

    NASA Technical Reports Server (NTRS)

    Deloach, R.; Donaldson, J. L.; Johnson, M. J.

    1986-01-01

    The airport-noise levels and annoyance model (ALAMO) is described in terms of the constituent modules, the execution of ALAMO procedure files, necessary for system execution, and the source code documentation associated with code development at Langley Research Center. The modules constituting ALAMO are presented both in flow graph form, and through a description of the subroutines and functions that comprise them.

  15. A Sailor in the Los Alamos Navy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Judd, D. L.; Meade, Roger Allen

    As part of the War Department’s Manhattan Engineer District (MED), Los Alamos was an Army installation during World War II, complete with a base commander and a brace of MPs. But it was a unique Army installation, having more civilian then military personnel. Even more unique was the work performed by the civilian population, work that required highly educated scientists and engineers. As the breadth, scope, and complexity of the Laboratory’s work increased, more and more technically educated and trained personnel were needed. But, the manpower needs of the nation’s war economy had created a shortage of such people. Tomore » meet its manpower needs, the MED scoured the ranks of the Army for anyone who had technical training and reassigned these men to its laboratories, including Los Alamos, as part of its Special Engineer Detachment (SED). Among the SEDs assigned to Los Alamos was Val Fitch, who was awarded the Nobel Prize in Physics in 1980. Another was Al Van Vessem, who helped stack the TNT for the 100 ton test, bolted together the Trinity device, and rode shotgun with the bomb has it was driven from Los Alamos to ground zero.« less

  16. Environmental analysis of Lower Pueblo/Lower Los Alamos Canyon, Los Alamos, New Mexico

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ferenbaugh, R.W.; Buhl, T.E.; Stoker, A.K.

    1994-12-01

    The radiological survey of the former radioactive waste treatment plant site (TA-45), Acid Canyon, Pueblo Canyon, and Los Alamos Canyon found residual contamination at the site itself and in the channel and banks of Acid, Pueblo, and lower Los Alamos Canyons all the way to the Rio Grande. The largest reservoir of residual radioactivity is in lower Pueblo Canyon, which is on DOE property. However, residual radioactivity does not exceed proposed cleanup criteria in either lower Pueblo or lower Los Alamos Canyons. The three alternatives proposed are (1) to take no action, (2) to construct a sediment trap in lowermore » Pueblo Canyon to prevent further transport of residual radioactivity onto San Ildefonso Indian Pueblo land, and (3) to clean the residual radioactivity from the canyon system. Alternative 2, to cleanup the canyon system, is rejected as a viable alternative. Thousands of truckloads of sediment would have to be removed and disposed of, and this effort is unwarranted by the low levels of contamination present. Residual radioactivity levels, under either present conditions or projected future conditions, will not result in significant radiation doses to persons exposed. Modeling efforts show that future transport activity will not result in any residual radioactivity concentrations higher than those already existing. Thus, although construction of a sediment trap in lower Pueblo Canyon is a viable alternative, this effort also is unwarranted, and the no-action alternative is the preferred alternative.« less

  17. New Generation of Los Alamos Opacity Tables

    NASA Astrophysics Data System (ADS)

    Colgan, James; Kilcrease, D. P.; Magee, N. H.; Sherrill, M. E.; Abdallah, J.; Hakel, P.; Fontes, C. J.; Guzik, J. A.; Mussack, K. A.

    2016-05-01

    We present a new generation of Los Alamos OPLIB opacity tables that have been computed using the ATOMIC code. Our tables have been calculated for all 30 elements from hydrogen through zinc and are publicly available through our website. In this poster we discuss the details of the calculations that underpin the new opacity tables. We also show several recent applications of the use of our opacity tables to solar modeling and other astrophysical applications. In particular, we demonstrate that use of the new opacities improves the agreement between solar models and helioseismology, but does not fully resolve the long-standing `solar abundance' problem. The Los Alamos National Laboratory is operated by Los Alamos National Security, LLC for the National Nuclear Security Administration of the U.S. Department of Energy under Contract No. DE-AC5206NA25396.

  18. Los Alamos National Laboratory Overview

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Neu, Mary

    Mary Neu, Associate Director for Chemistry, Life and Earth Sciences at Los Alamos National Laboratory, delivers opening remarks at the "Sequencing, Finishing, Analysis in the Future" meeting in Santa Fe, NM.

  19. SEDs at Los Alamos: A Personal Memoir

    NASA Astrophysics Data System (ADS)

    Bederson, Benjamin

    2001-03-01

    I have written this personal memoir approximately 55 years after the events I describe. It is based almost exclusively on memory, since apart from the diary I kept while on Tinian, I have few documents concerning it. It covers my service in the U.S. Army's Special Engineering Detachment (SED) in Oak Ridge and Los Alamos in 1944-45, on Tinian island, the launching pad for the bombing raids on Japan, in the summer and fall of 1945, and my return to Los Alamos until my discharge in January 1946.

  20. Fifty-one years of Los Alamos Spacecraft

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fenimore, Edward E.

    2014-09-04

    From 1963 to 2014, the Los Alamos National Laboratory was involved in at least 233 spacecraft. There are probably only one or two institutions in the world that have been involved in so many spacecraft. Los Alamos space exploration started with the Vela satellites for nuclear test detection, but soon expanded to ionospheric research (mostly barium releases), radioisotope thermoelectric generators, solar physics, solar wind, magnetospheres, astrophysics, national security, planetary physics, earth resources, radio propagation in the ionosphere, and cubesats. Here, we present a list of the spacecraft, their purpose, and their launch dates for use during RocketFest

  1. Using the Internet in Middle Schools: A Model for Success. A Collaborative Effort between Los Alamos National Laboratory (LANL) and Los Alamos Middle School (LAMS).

    ERIC Educational Resources Information Center

    Addessio, Barbara K.; And Others

    Los Alamos National Laboratory (LANL) developed a model for school networking using Los Alamos Middle School as a testbed. The project was a collaborative effort between the school and the laboratory. The school secured administrative funding for hardware and software; and LANL provided the network architecture, installation, consulting, and…

  2. A survey of macromycete diversity at Los Alamos National Laboratory, Bandelier National Monument, and Los Alamos County; A preliminary report

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Jarmie, N.; Rogers, F.J.

    The authors have completed a 5-year survey (1991--1995) of macromycetes found in Los Alamos County, Los Alamos National Laboratory, and Bandelier National Monument. The authors have compiled a database of 1,048 collections, their characteristics, and identifications. The database represents 123 (98%) genera and 175 (73%) species reliably identified. Issues of habitat loss, species extinction, and ecological relationships are addressed, and comparisons with other surveys are made. With this baseline information and modeling of this baseline data, one can begin to understand more about the fungal flora of the area.

  3. Los Alamos, Toshiba probing Fukushima with cosmic rays

    ScienceCinema

    Morris, Christopher

    2018-01-16

    Los Alamos National Laboratory has announced an impending partnership with Toshiba Corporation to use a Los Alamos technique called muon tomography to safely peer inside the cores of the Fukushima Daiichi reactors and create high-resolution images of the damaged nuclear material inside without ever breaching the cores themselves. The initiative could reduce the time required to clean up the disabled complex by at least a decade and greatly reduce radiation exposure to personnel working at the plant. Muon radiography (also called cosmic-ray radiography) uses secondary particles generated when cosmic rays collide with upper regions of Earth's atmosphere to create images of the objects that the particles, called muons, penetrate. The process is analogous to an X-ray image, except muons are produced naturally and do not damage the materials they contact. Muon radiography has been used before in imaginative applications such as mapping the interior of the Great Pyramid at Giza, but Los Alamos's muon tomography technique represents a vast improvement over earlier technology.

  4. New Rad Lab for Los Alamos

    ScienceCinema

    None

    2017-12-09

    The topping out ceremony for a key construction stage in the Los Alamos National Laboratory's newest facility, the Radiological Laboratory Utility & Office Building. This is part of the National Nu...  

  5. Notes on Los Alamos

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Meade, Roger Allen

    In 1954 an unknown author drafted a report, reprinted below, describing the Laboratory and the community as they existed in late 1953. This report, perhaps intended to be crafted into a public relations document, is valuable because it gives us an autobiographical look at Los Alamos during the first half of the 1950s. It has been edited to enhance readability.

  6. Pre Incident Planning For The Los Alamos National Laboratory

    DTIC Science & Technology

    2017-12-01

    laboratory was asked to design and build the world’s first atomic bomb . The Los Alamos Fire Department (LAFD) provides emergency response services to...Project: the newly established laboratory was asked to design and build the world’s first atomic bomb . The Los Alamos Fire Department (LAFD) provides...lower priority despite its importance to the responders’ scene safety.20 In a Carolina Fire Rescue EMS Journal article, retired New York City

  7. Los Alamos Programming Models

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bergen, Benjamin Karl

    This is the PDF of a powerpoint presentation from a teleconference on Los Alamos programming models. It starts by listing their assumptions for the programming models and then details a hierarchical programming model at the System Level and Node Level. Then it details how to map this to their internal nomenclature. Finally, a list is given of what they are currently doing in this regard.

  8. A progress report on UNICOS misuse detection at Los Alamos

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Thompson, J.L.; Jackson, K.A.; Stallings, C.A.

    An effective method for detecting computer misuse is the automatic monitoring and analysis of on-line user activity. During the past year, Los Alamos enhanced its Network Anomaly Detection and Intrusion Reporter (NADIR) to include analysis of user activity on Los Alamos` UNICOS Crays. In near real-time, NADIR compares user activity to historical profiles and tests activity against expert rules. The expert rules express Los Alamos` security policy and define improper or suspicious behavior. NADIR reports suspicious behavior to security auditors and provides tools to aid in follow-up investigations. This paper describes the implementation to date of the UNICOS component ofmore » NADIR, along with the operational experiences and future plans for the system.« less

  9. Los Alamos, Toshiba probing Fukushima with cosmic rays

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Morris, Christopher

    2014-06-16

    Los Alamos National Laboratory has announced an impending partnership with Toshiba Corporation to use a Los Alamos technique called muon tomography to safely peer inside the cores of the Fukushima Daiichi reactors and create high-resolution images of the damaged nuclear material inside without ever breaching the cores themselves. The initiative could reduce the time required to clean up the disabled complex by at least a decade and greatly reduce radiation exposure to personnel working at the plant. Muon radiography (also called cosmic-ray radiography) uses secondary particles generated when cosmic rays collide with upper regions of Earth's atmosphere to create imagesmore » of the objects that the particles, called muons, penetrate. The process is analogous to an X-ray image, except muons are produced naturally and do not damage the materials they contact. Muon radiography has been used before in imaginative applications such as mapping the interior of the Great Pyramid at Giza, but Los Alamos's muon tomography technique represents a vast improvement over earlier technology.« less

  10. Biological assessment for the effluent reduction program, Los Alamos National Laboratory, Los Alamos, New Mexico

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cross, S.P.

    1996-08-01

    This report describes the biological assessment for the effluent recution program proposed to occur within the boundaries of Los Alamos National Laboratory. Potential effects on wetland plants and on threatened and endangered species are discussed, along with a detailed description of the individual outfalls resulting from the effluent reduction program.

  11. Monte Carlo calculation of dose rate conversion factors for external exposure to photon emitters in soil

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Clovas, A.; Zanthos, S.; Antonopoulos-Domis, M.

    2000-03-01

    The dose rate conversion factors {dot D}{sub CF} (absorbed dose rate in air per unit activity per unit of soil mass, nGy h{sup {minus}1} per Bq kg{sup {minus}1}) are calculated 1 m above ground for photon emitters of natural radionuclides uniformly distributed in the soil. Three Monte Carlo codes are used: (1) The MCNP code of Los Alamos; (2) The GEANT code of CERN; and (3) a Monte Carlo code developed in the Nuclear Technology Laboratory of the Aristotle University of Thessaloniki. The accuracy of the Monte Carlo results is tested by the comparison of the unscattered flux obtained bymore » the three Monte Carlo codes with an independent straightforward calculation. All codes and particularly the MCNP calculate accurately the absorbed dose rate in air due to the unscattered radiation. For the total radiation (unscattered plus scattered) the {dot D}{sub CF} values calculated from the three codes are in very good agreement between them. The comparison between these results and the results deduced previously by other authors indicates a good agreement (less than 15% of difference) for photon energies above 1,500 keV. Antithetically, the agreement is not as good (difference of 20--30%) for the low energy photons.« less

  12. Los Alamos Using Neutrons to Stop Nuclear Smugglers

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Favalli, Andrea; Swinhoe, Martyn; Roark, Kevin

    Los Alamos National Laboratory researchers have successfully demonstrated for the first time that laser-generated neutrons can be enlisted as a useful tool in the War on Terror. The international research team used the short-pulse laser at Los Alamos's TRIDENT facility to generate a neutron beam with novel characteristics that interrogated a closed container to confirm the presence and quantity of nuclear material inside. The successful experiment paves the way for creation of a table-top-sized or truck-mounted neutron generator that could be installed at strategic locations worldwide to thwart smugglers trafficking in nuclear materials.

  13. Los Alamos Using Neutrons to Stop Nuclear Smugglers

    ScienceCinema

    Favalli, Andrea; Swinhoe, Martyn; Roark, Kevin

    2018-02-14

    Los Alamos National Laboratory researchers have successfully demonstrated for the first time that laser-generated neutrons can be enlisted as a useful tool in the War on Terror. The international research team used the short-pulse laser at Los Alamos's TRIDENT facility to generate a neutron beam with novel characteristics that interrogated a closed container to confirm the presence and quantity of nuclear material inside. The successful experiment paves the way for creation of a table-top-sized or truck-mounted neutron generator that could be installed at strategic locations worldwide to thwart smugglers trafficking in nuclear materials.

  14. Results from the CACTI experiment: Air-Cerenkov and particle measurements of PeV air showers at Los Alamos

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Paling, S.; Hillas, A.M.; Berley, D.

    1997-07-01

    An array of six wide angle Cerenkov detectors was constructed amongst the scintillator and muon detectors of the CYGNUS II array at Los Alamos National Laboratory to investigate cosmic ray composition in the PeV region through measurements of the shape of Cerenkov lateral distributions. Data were collected during clear, moonless nights over three observing periods in 1995. Estimates of depths of shower maxima determined from the recorded Cerenkov lateral distributions align well with existing results at higher energies and suggest a mixed to heavy composition in the PeV region with no significant variation observed around the knee. The accuracy ofmore » composition determination is limited by uncertainties in the expected levels of depth of maximum predicted using different Monte-Carlo shower simulation models.« less

  15. Flaws found in Los Alamos safety procedures

    NASA Astrophysics Data System (ADS)

    Gwynne, Peter

    2017-12-01

    A US government panel on nuclear safety has discovered a series of safety issues at the Los Alamos National Laboratory, concluding that government oversight of the lab's emergency preparation has been ineffective.

  16. 75 FR 72829 - Los Alamos Historical Document Retrieval and Assessment (LAHDRA) Project

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-11-26

    ... Historical Document Retrieval and Assessment (LAHDRA) Project The Centers for Disease Control and Prevention... release of the Final Report of the Los Alamos Historical Document Retrieval and Assessment (LAHDRA)Project... information about historical chemical or radionuclide releases from facilities at the Los Alamos National...

  17. A physicists guide to The Los Alamos Primer

    NASA Astrophysics Data System (ADS)

    Reed, B. Cameron

    2016-11-01

    In April 1943, a group of scientists at the newly established Los Alamos Laboratory were given a series of lectures by Robert Serber on what was then known of the physics and engineering issues involved in developing fission bombs. Serber’s lectures were recorded in a 24 page report titled The Los Alamos Primer, which was subsequently declassified and published in book form. This paper describes the background to the Primer and analyzes the physics contained in its 22 sections. The motivation for this paper is to provide a firm foundation of the background and contents of the Primer for physicists interested in the Manhattan Project and nuclear weapons.

  18. Environmental surveillance at Los Alamos during 1994

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    NONE

    1996-07-01

    This report describes environmental monitoring activities at Los Alamos National Laboratory for 1994. Data were collected to assess external penetrating radiation, airborne emissions, liquid effluents, radioactivity of environmental materials and food stuffs, and environmental compliance.

  19. A Monte Carlo Simulation for Understanding Energy Measurements of Beta Particles Detected by the UCNb Experiment

    NASA Astrophysics Data System (ADS)

    Feng, Chi; UCNb Collaboration

    2011-10-01

    It is theorized that contributions to the Fierz interference term from scalar interaction beyond the Standard Model could be detectable in the spectrum of neutron beta-decay. The UCNb experiment run at the Los Alamos Neutron Science Center aims to accurately measure the neutron beta-decay energy spectrum to detect a nonzero interference term. The instrument consists of a cubic ``integrating sphere'' calorimeter attached with up to 4 photomultiplier tubes. The inside of the calorimeter is coated with white paint and a thin UV scintillating layer made of deuterated polystyrene to contain the ultracold neutrons. A Monte Carlo simulation using the Geant4 toolkit is developed in order to provide an accurate method of energy reconstruction. Offline calibration with the Kellogg Radiation Laboratory 140 keV electron gun and conversion electron sources will be used to validate the Monte Carlo simulation to give confidence in the energy reconstruction methods and to better understand systematics in the experiment data.

  20. Recent Infrasound Calibration Activity at Los Alamos

    NASA Astrophysics Data System (ADS)

    Whitaker, R. W.; Marcillo, O. E.

    2014-12-01

    Absolute infrasound sensor calibration is necessary for estimating source sizes from measured waveforms. This can be an important function in treaty monitoring. The Los Alamos infrasound calibration chamber is capable of absolute calibration. Early in 2014 the Los Alamos infrasound calibration chamber resumed operations in its new location after an unplanned move two years earlier. The chamber has two sources of calibration signals. The first is the original mechanical piston, and the second is a CLD Dynamics Model 316 electro-mechanical unit that can be digitally controlled and provide a richer set of calibration options. During 2008-2010 a number of upgrades were incorporated for improved operation and recording. In this poster we give an overview of recent chamber work on sensor calibrations, calibration with the CLD unit, some measurements with different porous hoses and work with impulse sources.

  1. Geothermal investigation of spring and well waters of the Los Alamos Region, New Mexico

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Goff, F.E.; Sayer, S.

    1980-04-01

    The chemical and isotopic characters of 20 springs and wells in the Los Alamos area were investigated for indications of geothermal potential. These waters were compared with known hot and mineral springs from adjacent Valles Caldera and San Ysidro. All waters in the Los Alamos area are composed of meteoric water. Isotopic data show that the two primary aquifers beneath the Los Alamos region have different recharge areas. Relatively high concentrations of lithium, arsenic, chlorine, boron, and fluorine in some of the Los Alamos wells suggest these waters may contain a small fraction of thermal/mineral water of deep origin. Thermalmore » water probably rises up high-angle faults associated with a graben of the Rio Grande rift now buried by the Pajarito Plateau.« less

  2. Penetrating radiation: applications at Los Alamos National Laboratory

    NASA Astrophysics Data System (ADS)

    Watson, Scott; Hunter, James; Morris, Christopher

    2013-09-01

    Los Alamos has used penetrating radiography extensively throughout its history dating back to the Manhattan Project where imaging dense, imploding objects was the subject of intense interest. This interest continues today as major facilities like DARHT1 have become the mainstay of the US Stockpile Stewardship Program2 and the cornerstone of nuclear weapons certification. Meanwhile, emerging threats to national security from cargo containers and improvised explosive devices (IEDs) have invigorated inspection efforts using muon tomography, and compact x-ray radiography. Additionally, unusual environmental threats, like those from underwater oil spills and nuclear power plant accidents, have caused renewed interest in fielding radiography in severe operating conditions. We review the history of penetrating radiography at Los Alamos and survey technologies as presently applied to these important problems.

  3. Los Alamos National Laboratory Human and Intellectual Capital for Sustaining Nuclear Deterrence

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    McAlpine, Bradley

    2015-04-01

    This paper provides an overview of the current human and intellectual capital at Los Alamos National Laboratory, through specific research into the statistics and demographics as well as numerous personal interviews at all levels of personnel. Based on this information, a series of recommendations are provided to assist Los Alamos National Laboratory in ensuring the future of the human and intellectual capital for the nuclear deterrence mission. While the current human and intellectual capital is strong it stands on the precipice and action must be taken to ensure Los Alamos National Laboratory maintains leadership in developing and sustaining national nuclearmore » capabilities. These recommendations may be applicable to other areas of the nuclear enterprise, including the Air Force, after further research and study.« less

  4. Los Alamos Before and After the Fire

    NASA Technical Reports Server (NTRS)

    2002-01-01

    On May 4, 2000, a prescribed fire was set at Bandelier National Monument, New Mexico, to clear brush and dead and dying undergrowth to prevent a larger, subsequent wildfire. Unfortunately, due to high winds and extremely dry conditions in the surrounding area, the prescribed fire quickly raged out of control and, by May 10, the blaze had spread into the nearby town of Los Alamos. In all, more than 20,000 people were evacuated from their homes and more than 200 houses were destroyed as the flames consumed about 48,000 acres in and around the Los Alamos area. The pair of images above were acquired by the Enhanced Thematic Mapper Plus (ETM+) sensor, flying aboard NASA's Landsat 7 satellite, shortly before the Los Alamos fire (top image, acquired April 14) and shortly after the fire was extinguished (lower image, June 17). The images reveal the extent of the damage caused by the fire. Combining ETM+ channels 7, 4, and 2 (one visible and two infrared channels) results in a false-color image where vegetation appears as bright to dark green. Forested areas are generally dark green while herbaceous vegetation is light green. Rangeland or more open areas appear pink to light purple. Areas with extensive pavement or urban development appear light blue or white to purple. Less densely-developed residential areas appear light green and golf courses are very bright green. In the lower image, the areas recently burned appear bright red. Landsat 7 data courtesy United States Geological Survey EROS DataCenter. Images by Robert Simmon, NASA GSFC.

  5. Critical partnerships: Los Alamos, universities, and industry

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Berger, C.L.

    1997-04-01

    Los Alamos National Laboratory, situated 35 miles northwest of Santa Fe, NM, is one of the Department of Energy`s three Defense Programs laboratories. It encompasses 43 square miles, employees approximately 10,000 people, and has a budget of approximately $1.1B in FY97. Los Alamos has a strong post-cold war mission, that of reducing the nuclear danger. But even with that key role in maintaining the nation`s security, Los Alamos views partnerships with universities and industry as critical to its future well being. Why is that? As the federal budget for R&D comes under continued scrutiny and certain reduction, we believe thatmore » the triad of science and technology contributors to the national system of R&D must rely on and leverage each others capabilities. For us this means that we will rely on these partners to help us in 5 key ways: We expect that partnerships will help us maintain and enhance our core competencies. In doing so, we will be able to attract the best scientists and engineers. To keep on the cutting edge of research and development, we have found that partnerships maintain the excellence of staff through new and exciting challenges. Additionally, we find that from our university and corporate partners we often learn and incorporate {open_quotes}best practices{close_quotes} in organizational management and operations. Finally, we believe that a strong national system of R&D will ensure and enhance our ability to generate revenues.« less

  6. Los Alamos National Laboratory Facility Review

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Nelson, Ronald Owen

    2015-06-05

    This series of slides depicts the Los Alamos Neutron Science Center (LANSCE). The Center's 800-MeV linac produces H + and H - beams as well as beams of moderated (cold to 1 MeV) and unmoderated (0.1 to 600 MeV) neutrons. Experimental facilities and their capabilities and characteristics are outlined. Among these are LENZ, SPIDER, and DANCE.

  7. Los Alamos National Laboratory Prepares for Fire Season

    ScienceCinema

    L’Esperance, Manny

    2018-01-16

    Through the establishment of a Wildland Fire Program Office, and the Interagency Fire Base located on Laboratory property, Los Alamos National Laboratory is continuing and improving a program to prepare for wildland fire.

  8. Los Alamos National Laboratory Prepares for Fire Season

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    L’Esperance, Manny

    Through the establishment of a Wildland Fire Program Office, and the Interagency Fire Base located on Laboratory property, Los Alamos National Laboratory is continuing and improving a program to prepare for wildland fire.

  9. History of Los Alamos Participation in Active Experiments in Space

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Pongratz, Morris B.

    Beginning with the Teak nuclear test in 1958, Los Alamos has a long history of participation in active experiments in space. The last pertinent nuclear tests were the five explosions as part of the Dominic series in 1962. The Partial Test Ban Treaty signed in August 1963 prohibited all test detonations of nuclear weapons except for those conducted underground. Beginning with the “Apple” thermite barium release in June 1968 Los Alamos has participated in nearly 100 non-nuclear experiments in space, the last being the NASA-sponsored “AA-2” strontium and europium doped barium thermite releases in the Arecibo beam in July ofmore » 1992. The rationale for these experiments ranged from studying basic plasma processes such as gradientdriven structuring and velocity-space instabilities to illuminating the convection of plasmas in the ionosphere and polar cap to ionospheric depletion experiments to the B.E.A.R. 1-MeV neutral particle beam test in 1989. This report reviews the objectives, techniques and diagnostics of Los Alamos participation in active experiments in space.« less

  10. Water Supply at Los Alamos during 1997

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    M. N. Maes; S. G. McLin; W. D. Purtymun

    1998-12-01

    Production of potable municipal water supplies during 1997 totaled about 1,285.9 million gallons from wells in the Guaje, Pajarito, and Otowi well fields. There was no water used from the spring gallery in Water Canyon or from Guaje Reservoir during 1997. About 2.4 million gallons of water from Los Alamos Reservoir was used to irrigate public parks and recreational lands. The total water usage in 1997 was about 1,288.3 million gallons, or about 135 gallons per day per person living in Los Alamos County. Groundwater pumpage was down about 82.2 million gallons in 1997 compared with the pumpage in 1996.more » Four new replacement wells were drilled and cased in Guaje Canyon between October 1997 and March 1998. These wells are currently being developed and aquifer tests are being performed. A special report summarizing the geological, geophysical, and well construction logs will be issued in the near future for these new wells.« less

  11. Development of the Los Alamos continuous high average-power microsecond pulser ion accelerator

    NASA Astrophysics Data System (ADS)

    Bitteker, L. J.; Wood, B. P.; Davis, H. A.; Waganaar, W. J.; Boyd, I. D.; Lovberg, R. H.

    2000-10-01

    The continuous high average-power microsecond pulser (CHAMP) ion accelerator is being constructed at Los Alamos National Laboratory. Progress on the testing of the CHAMP diode is discussed. A direct simulation Monte Carlo computer code is used to investigate the puffed gas fill of the CHAMP anode. High plenum pressures and low plenum volumes are found to be desirable for effective gas puffs. The typical gas fill time is 150-180 μs from initiation of valve operation to end of fill. Results of anode plasma production at three stages of development are discussed. Plasma properties are monitored with electric and magnetic field probes. From this data, the near coil plasma density under nominal conditions is found to be on the order of 1×1016 cm-3. Large error is associated with this calculation due to inconsistencies between tests and the limitations of the instrumentation used. The diode insulating magnetic field is observed to result in lower density plasma with a more diffuse structure than for the cases when the insulating field is not applied. The importance of these differences in plasma quality on the beam production is yet to be determined.

  12. Los Alamos high-power proton linac designs

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lawrence, G.P.

    1995-10-01

    Medium-energy high-power proton linear accelerators have been studied at Los Alamos as drivers for spallation neutron applications requiring large amounts of beam power. Reference designs for such accelerators are discussed, important design factors are reviewed, and issues and concern specific to this unprecedented power regime are discussed.

  13. Audit Report, "Fire Protection Deficiencies at Los Alamos National Laboratory"

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    None

    2009-06-01

    The Department of Energy's Los Alamos National Laboratory (Los Alamos) maintains some of the Nation's most important national security assets, including nuclear materials. Many of Los Alamos' facilities are located in close proximity to one another, are occupied by large numbers of contract and Federal employees, and support activities ranging from nuclear weapons design to science-related activities. Safeguarding against fires, regardless of origin, is essential to protecting employees, surrounding communities, and national security assets. On June 1, 2006, Los Alamos National Security, LLC (LANS), became the managing and operating contractor for Los Alamos, under contract with the Department's National Nuclearmore » Security Administration (NNSA). In preparation for assuming its management responsibilities at Los Alamos, LANS conducted walk-downs of the Laboratory's facilities to identify pre-existing deficiencies that could give rise to liability, obligation, loss or damage. The walk-downs, which identified 812 pre-existing fire protection deficiencies, were conducted by subject matter professionals, including fire protection experts. While the Los Alamos Site Office has overall responsibility for the effectiveness of the fire protection program, LANS, as the Laboratory's operating contractor, has a major, day-to-day role in minimizing fire-related risks. The issue of fire protection at Los Alamos is more than theoretical. In May 2000, the 'Cerro Grande' fire burned about 43,000 acres, including 7,700 acres of Laboratory property. Due to the risk posed by fire to the Laboratory's facilities, workforce, and surrounding communities, we initiated this audit to determine whether pre-existing fire protection deficiencies had been addressed. Our review disclosed that LANS had not resolved many of the fire protection deficiencies that had been identified in early 2006: (1) Of the 296 pre-existing deficiencies we selected for audit, 174 (59 percent) had not been

  14. Los Alamos on Radio Café: Nina Lanza

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lanza, Nina; Domandi, Mary-Charlotte

    2017-04-11

    First up in the new series is Los Alamos National Laboratory’s Nina Lanza from the Space and Remote Sensing group. Lanza is a planetary geologist who has been part of the Mars Curiosity Rover “ChemCam” team since 2012.

  15. Airport-Noise Levels and Annoyance Model (ALAMO) user's guide

    NASA Technical Reports Server (NTRS)

    Deloach, R.; Donaldson, J. L.; Johnson, M. J.

    1986-01-01

    A guide for the use of the Airport-Noise Level and Annoyance MOdel (ALAMO) at the Langley Research Center computer complex is provided. This document is divided into 5 primary sections, the introduction, the purpose of the model, and an in-depth description of the following subsystems: baseline, noise reduction simulation and track analysis. For each subsystem, the user is provided with a description of architecture, an explanation of subsystem use, sample results, and a case runner's check list. It is assumed that the user is familiar with the operations at the Langley Research Center (LaRC) computer complex, the Network Operating System (NOS 1.4) and CYBER Control Language. Incorporated within the ALAMO model is a census database system called SITE II.

  16. A New Generation of Los Alamos Opacity Tables

    DOE PAGES

    Colgan, James Patrick; Kilcrease, David Parker; Magee, Jr., Norman H.; ...

    2016-01-26

    We present a new, publicly available, set of Los Alamos OPLIB opacity tables for the elements hydrogen through zinc. Our tables are computed using the Los Alamos ATOMIC opacity and plasma modeling code, and make use of atomic structure calculations that use fine-structure detail for all the elements considered. Our equation-of-state (EOS) model, known as ChemEOS, is based on the minimization of free energy in a chemical picture and appears to be a reasonable and robust approach to determining atomic state populations over a wide range of temperatures and densities. In this paper we discuss in detail the calculations thatmore » we have performed for the 30 elements considered, and present some comparisons of our monochromatic opacities with measurements and other opacity codes. We also use our new opacity tables in solar modeling calculations and compare and contrast such modeling with previous work.« less

  17. James L. Tuck Los Alamos ball lightning pioneer

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Baker, D.A.

    1999-07-01

    James Tuck was well known for starting the Project Sherwood group at Los Alamos Scientific Laboratory in 1952. This group was formed to study and develop concepts for controlled fusion energy. In his later years after retiring from Controlled Fusion Division, he continued research at Los Alamos on the topic of ball lightning. He traveled widely giving lectures on both observations of others and his own experimental efforts. He collected anecdotal observations obtained from those in his lecture audiences during his travels and from responses from newspaper articles where he asked for specific information from ball lightning observers. He finallymore » cut off this collection of data when the number of responses became overwhelming. The author's primary publication on ball lightning was a short laboratory report. He planned on publishing a book on the subject but this was never completed before his death. Tuck focused his experimental effort on attempting to duplicate the production of plasma balls claimed to be observed in US Navy submarines when a switch was opened under overload conditions with battery power. During lunch breaks he made use of a Los Alamos N-division battery bank facility to mock up a submarine power pack and switch gear. This non-funded effort was abruptly terminated when an explosion occurred in the facility. An overview of Tuck's research and views will be given. The flavor Jim's personality as well as a ball produced with his experimental apparatus will be shown using video chips.« less

  18. 78 FR 44100 - Submission for OMB Review; Comment Request

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-07-23

    ...: Executive Order 12829, ``National Industrial Security Program (NISP)'' Section 202(a) stipulates that the... implementation of the NISP. Department of Defense Directive 5105.42, ``Defense Security Service,'' dated August 3...

  19. Upgrades and Enclosure of Building 15 at Technical Area 40: Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Plimpton, Kathryn D; Garcia, Kari L. M; Brunette, Jeremy Christopher

    The U.S. Department of Energy, National Nuclear Security Administration, Los Alamos Field Office (Field Office) proposes to upgrade and enclose Building 15 at Technical Area (TA) 40, Los Alamos National Laboratory. Building TA-40-15, a Cold War-era firing site, was determined eligible for listing in the National Register of Historic Places (Register) in DX Division’s Facility Strategic Plan: Consolidation and Revitalization at Technical Areas 6, 8, 9, 14, 15, 22, 36, 39, 40, 60, and 69 (McGehee et al. 2005). Building TA-40-15 was constructed in 1950 to support detonator testing. The firing site will be enclosed by a steel building tomore » create a new indoor facility that will allow for year-round mission capability. Enclosing TA-40-15 will adversely affect the building by altering the characteristics that make it eligible for the Register. In compliance with Section 106 of the National Historic Preservation Act of 1966, as amended, the Field Office is initiating consultation for this proposed undertaking. The Field Office is also requesting concurrence with the use of standard practices to resolve adverse effects as defined in the Programmatic Agreement among the U.S. Department of Energy, National Nuclear Security Administration, Los Alamos Field Office, the New Mexico State Historic Preservation Office and the Advisory Council on Historic Preservation Concerning Management of the Historic Properties at Los Alamos National Laboratory, Los Alamos, New Mexico.« less

  20. Los Alamos on Radio Café: Ludmil Alexandrov

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Domandi, Mary-Charlotte; Alexandrov, Ludmil

    In a creative breakthrough in cancer research, Ludmil Alexandrov, the J. Robert Oppenheimer Distinguished Postdoctoral Fellow at Los Alamos National Laboratory, combines Big Data, supercomputing and machine-learning to identify the telltale mutations of cancer. Knowing these mutational signatures can help researchers develop new methods of prevention.

  1. Induction Inserts at the Los Alamos PSR

    NASA Astrophysics Data System (ADS)

    Ng, K. Y.

    2002-12-01

    Ferrite-loaded induction tuners installed in the Los Alamos Proton Storage Ring have been successful in compensating space-charge effects. However, the resistive part of the ferrite introduces unacceptable microwave instability and severe bunch lengthening. An effective cure was found by heating the ferrite cores up to ˜ 130°C. An understanding of the instability and cure is presented.

  2. Los Alamos Team Demonstrates Bottle Scanner Technology

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Espy, Michelle; Schultz, Larry

    2014-05-06

    Los Alamos scientists are demonstrating a Nuclear Magnetic Resonance Imaging (NMR) technology that may provide a breakthrough for screening liquids at airport security. By adding low-power X-ray data to the NMR mix, scientists believe they have unlocked a new detection technology. Funded in part by the Department of Homeland Security's Science and Technology Directorate, the new technology is called MagRay.

  3. Los Alamos Team Demonstrates Bottle Scanner Technology

    ScienceCinema

    Espy, Michelle; Schultz, Larry

    2018-02-13

    Los Alamos scientists are demonstrating a Nuclear Magnetic Resonance Imaging (NMR) technology that may provide a breakthrough for screening liquids at airport security. By adding low-power X-ray data to the NMR mix, scientists believe they have unlocked a new detection technology. Funded in part by the Department of Homeland Security's Science and Technology Directorate, the new technology is called MagRay.

  4. Preliminary estimates of nucleon fluxes in a water target exposed to solar-flare protons: BRYNTRN versus Monte Carlo code

    NASA Technical Reports Server (NTRS)

    Shinn, Judy L.; Wilson, John W.; Lone, M. A.; Wong, P. Y.; Costen, Robert C.

    1994-01-01

    A baryon transport code (BRYNTRN) has previously been verified using available Monte Carlo results for a solar-flare spectrum as the reference. Excellent results were obtained, but the comparisons were limited to the available data on dose and dose equivalent for moderate penetration studies that involve minor contributions from secondary neutrons. To further verify the code, the secondary energy spectra of protons and neutrons are calculated using BRYNTRN and LAHET (Los Alamos High-Energy Transport code, which is a Monte Carlo code). These calculations are compared for three locations within a water slab exposed to the February 1956 solar-proton spectrum. Reasonable agreement was obtained when various considerations related to the calculational techniques and their limitations were taken into account. Although the Monte Carlo results are preliminary, it appears that the neutron albedo, which is not currently treated in BRYNTRN, might be a cause for the large discrepancy seen at small penetration depths. It also appears that the nonelastic neutron production cross sections in BRYNTRN may underestimate the number of neutrons produced in proton collisions with energies below 200 MeV. The notion that the poor energy resolution in BRYNTRN may cause a large truncation error in neutron elastic scattering requires further study.

  5. 75 FR 1793 - Study Team for the Los Alamos Historical Document Retrieval and Assessment (LAHDRA) Project

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-01-13

    ... DEPARTMENT OF HEALTH AND HUMAN SERVICES Centers for Disease Control and Prevention Study Team for the Los Alamos Historical Document Retrieval and Assessment (LAHDRA) Project The Centers for Disease... the following meeting. Name: Public Meeting of the Study Team for the Los Alamos Historical Document...

  6. Total electron content (TEC) variability at Los Alamos, New Mexico: A comparative study: FORTE-derived TEC analysis

    NASA Astrophysics Data System (ADS)

    Huang, Zhen; Roussel-Dupré, Robert

    2005-12-01

    Data collected from Fast On-Orbit Recording of Transient Events (FORTE) satellite-received Los Alamos Portable Pulser (LAPP) signals during 1997-2002 are used to derive the total electron content (TEC) at Los Alamos, New Mexico. The LAPP-derived TECs at Los Alamos are analyzed for diurnal, seasonal, interannual, and 27-day solar cycle variations. Several aspects in deriving TEC are analyzed, including slant to vertical TEC conversion, quartic effects on transionosperic signals, and geomagnetic storm effects on the TEC variance superimposed on the averaged TEC values.

  7. Los Alamos Novel Rocket Design Flight Tested

    ScienceCinema

    Tappan, Bryce

    2018-04-16

    Los Alamos National Laboratory scientists recently flight tested a new rocket design that includes a high-energy fuel and a motor design that also delivers a high degree of safety. Researchers will now work to scale-up the design, as well as explore miniaturization of the system, in order to exploit all potential applications that would require high-energy, high-velocity, and correspondingly high safety margins.

  8. Los Alamos Novel Rocket Design Flight Tested

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tappan, Bryce

    Los Alamos National Laboratory scientists recently flight tested a new rocket design that includes a high-energy fuel and a motor design that also delivers a high degree of safety. Researchers will now work to scale-up the design, as well as explore miniaturization of the system, in order to exploit all potential applications that would require high-energy, high-velocity, and correspondingly high safety margins.

  9. Los Alamos Science: The Human Genome Project. Number 20, 1992

    DOE R&D Accomplishments Database

    Cooper, N. G.; Shea, N. eds.

    1992-01-01

    This document provides a broad overview of the Human Genome Project, with particular emphasis on work being done at Los Alamos. It tries to emphasize the scientific aspects of the project, compared to the more speculative information presented in the popular press. There is a brief introduction to modern genetics, including a review of classic work. There is a broad overview of the Genome Project, describing what the project is, what are some of its major five-year goals, what are major technological challenges ahead of the project, and what can the field of biology, as well as society expect to see as benefits from this project. Specific results on the efforts directed at mapping chromosomes 16 and 5 are discussed. A brief introduction to DNA libraries is presented, bearing in mind that Los Alamos has housed such libraries for many years prior to the Genome Project. Information on efforts to do applied computational work related to the project are discussed, as well as experimental efforts to do rapid DNA sequencing by means of single-molecule detection using applied spectroscopic methods. The article introduces the Los Alamos staff which are working on the Genome Project, and concludes with brief discussions on ethical, legal, and social implications of this work; a brief glimpse of genetics as it may be practiced in the next century; and a glossary of relevant terms.

  10. Optical velocimetry at the Los Alamos Proton Radiography Facility

    NASA Astrophysics Data System (ADS)

    Tupa, Dale; Tainter, Amy; Neukirch, Levi; Hollander, Brian; Buttler, William; Holtkamp, David; The Los Alamos Proton Radiography Team Team

    2016-05-01

    The Los Alamos Proton Radiography Facility (pRad) employs a high-energy proton beam to image the properties and behavior of materials driven by high explosives. We will discuss features of pRad and describe some recent experiments, highlighting optical diagnostics for surface velocity measurements.

  11. Aqueous Nitrate Recovery Line at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Finstad, Casey Charles

    2016-06-15

    This powerpoint is part of the ADPSM Plutonium Engineering Lecture Series, which is an opportunity for new hires at LANL to get an overview of work done at TA55. It goes into detail about the aqueous nitrate recovery line at Los Alamos National Laboratory.

  12. Evaluation of Macroinvertebrate Communities and Habitat for Selected Stream Reaches at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    L.J. Henne; K.J. Buckley

    2005-08-12

    This is the second aquatic biological monitoring report generated by Los Alamos National Laboratory's (LANL's) Water Quality and Hydrology Group. The study has been conducted to generate impact-based assessments of habitat and water quality for LANL waterways. The monitoring program was designed to allow for the detection of spatial and temporal trends in water and habitat quality through ongoing, biannual monitoring of habitat characteristics and benthic aquatic macroinvertebrate communities at six key sites in Los Alamos, Sandia, Water, Pajarito, and Starmer's Gulch Canyons. Data were collected on aquatic habitat characteristics, channel substrate, and macroinvertebrate communities during 2001 and 2002. Aquaticmore » habitat scores were stable between 2001 and 2002 at all locations except Starmer's Gulch and Pajarito Canyon, which had lower scores in 2002 due to low flow conditions. Channel substrate changes were most evident at the upper Los Alamos and Pajarito study reaches. The macroinvertebrate Stream Condition Index (SCI) indicated moderate to severe impairment at upper Los Alamos Canyon, slight to moderate impairment at upper Sandia Canyon, and little or no impairment at lower Sandia Canyon, Starmer's Gulch, and Pajarito Canyon. Habitat, substrate, and macroinvertebrate data from the site in upper Los Alamos Canyon indicated severe impacts from the Cerro Grande Fire of 2000. Impairment in the macroinvertebrate community at upper Sandia Canyon was probably due to effluent-dominated flow at that site. The minimal impairment SCI scores for the lower Sandia site indicated that water quality improved with distance downstream from the outfall at upper Sandia Canyon.« less

  13. Los Alamos - A Short History

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Meade, Roger A.

    At 5:45 am on the morning of July 16, 1945, the world’s first atomic bomb exploded over a remote section of the southern New Mexican desert known as the Jornada del Muerto, the Journey of Death. Three weeks later, the atomic bombs known as Little Boy and Fat Man brought World War II to an end. Working literally around the clock, these first atomic bombs were designed and built in just thirty months by scientists working at a secret scientific laboratory in the mountains of New Mexico known by its codename, Project Y, better known to the world as Losmore » Alamos.« less

  14. Recent UCN source developments at Los Alamos

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Seestrom, S.J.; Anaya, J.M.; Bowles, T.J.

    The most intense sources of ultra cold neutrons (UCN) have bee built at reactors where the high average thermal neutron flux can overcome the low UCN production rate to achieve usable densities of UCN. At spallation neutron sources the average flux available is much lower than at a reactor, though the peak flux can be comparable or higher. The authors have built a UCN source that attempts to take advantage of the high peak flux available at the short pulse spallation neutron source at the Los Alamos Neutron Science Center (LANSCE) to generate a useful number of UCN. In themore » source UCN are produced by Doppler-shifted Bragg scattering of neutrons to convert 400-m/s neutrons down into the UCN regime. This source was initially tested in 1996 and various improvements were made based on the results of the 1996 running. These improvements were implemented and tested in 1997. In sections 2 and 3 they discuss the improvements that have been made and the resulting source performance. Recently an even more interesting concept was put forward by Serebrov et al. This involves combining a solid Deuterium UCN source, previously studied by Serebrov et al., with a pulsed spallation source to achieve world record UCN densities. They have initiated a program of calculations and measurements aimed at verifying the solid Deuterium UCN source concept. The approach has been to develop an analytical capability, combine with Monte Carlo calculations of neutron production, and perform benchmark experiments to verify the validity of the calculations. Based on the calculations and measurements they plan to test a modified version of the Serebrov UCN factory. They estimate that they could produce over 1,000 UCN/cc in a 15 liter volume, using 1 {micro}amp of 800 MeV protons for two seconds every 500 seconds. They will discuss the result UCN production measurements in section 4.« less

  15. Los Alamos Science: The Human Genome Project. Number 20, 1992

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cooper, N G; Shea, N

    1992-01-01

    This article provides a broad overview of the Human Genome Project, with particular emphasis on work being done at Los Alamos. It tries to emphasize the scientific aspects of the project, compared to the more speculative information presented in the popular press. There is a brief introduction to modern genetics, including a review of classic work. There is a broad overview of the Genome Project, describing what the project is, what are some of its major five-year goals, what are major technological challenges ahead of the project, and what can the field of biology, as well as society expect tomore » see as benefits from this project. Specific results on the efforts directed at mapping chromosomes 16 and 5 are discussed. A brief introduction to DNA libraries is presented, bearing in mind that Los Alamos has housed such libraries for many years prior to the Genome Project. Information on efforts to do applied computational work related to the project are discussed, as well as experimental efforts to do rapid DNA sequencing by means of single-molecule detection using applied spectroscopic methods. The article introduces the Los Alamos staff which are working on the Genome Project, and concludes with brief discussions on ethical, legal, and social implications of this work; a brief glimpse of genetics as it may be practiced in the next century; and a glossary of relevant terms.« less

  16. An independent evaluation of plutonium body burdens in populations near Los Alamos Laboratory using human autopsy data.

    PubMed

    Gaffney, Shannon H; Donovan, Ellen P; Shonka, Joseph J; Le, Matthew H; Widner, Thomas E

    2013-06-01

    In the mid-1940s, the United States began producing atomic weapon components at the Los Alamos National Laboratory (LANL). In an attempt to better understand historical exposure to nearby residents, this study evaluates plutonium activity in human tissue relative to residential location and length of time at residence. Data on plutonium activity in the lung, vertebrae, and liver of nearby residents were obtained during autopsies as a part of the Los Alamos Tissue Program. Participant residential histories and the distance from each residence to the primary plutonium processing buildings at LANL were evaluated in the analysis. Summary statistics, including Student t-tests and simple regressions, were calculated. Because the biological half-life of plutonium can vary significantly by organ, data were analyzed separately by tissue type (lung, liver, vertebrae). The ratios of plutonium activity (vertebrae:liver; liver:lung) were also analyzed in order to evaluate the importance of timing of exposure. Tissue data were available for 236 participants who lived in a total of 809 locations, of which 677 were verified postal addresses. Residents of Los Alamos were found to have higher plutonium activities in the lung than non-residents. Further, those who moved to Los Alamos before 1955 had higher lung activities than those who moved there later. These trends were not observed with the liver, vertebrae, or vertebrae:liver and liver:lung ratio data, however, and should be interpreted with caution. Although there are many limitations to this study, including the amount of available data and the analytical methods used to analyze the tissue, the overall results indicate that residence (defined as the year that the individual moved to Los Alamos) may have had a strong correlation to plutonium activity in human tissue. This study is the first to present the results of Los Alamos Autopsy Program in relation to residential status and location in Los Alamos. Copyright © 2012

  17. Carbon isotope chemostratigraphy and precise dating of middle Frasnian (lower Upper Devonian) Alamo Breccia, Nevada, USA

    USGS Publications Warehouse

    Morrow, J.R.; Sandberg, C.A.; Malkowski, K.; Joachimski, M.M.

    2009-01-01

    At Hancock Summit West, Nevada, western USA, uppermost Givetian (upper Middle Devonian) and lower and middle Frasnian (lower Upper Devonian) rocks of the lower Guilmette Formation include, in stratigraphic sequence, carbonate-platform facies of the conodont falsiovalis, transitans, and punctata Zones; the type Alamo Breccia Member of the middle punctata Zone; and slope facies of the punctata and hassi Zones. The catastrophically deposited Alamo Breccia and related phenomena record the ~ 382??Ma Alamo event, produced by a km-scale bolide impact into a marine setting seaward of an extensive carbonate platform fringing western North America. Re-evaluation of conodonts from the lower Guilmette Formation and Alamo Breccia Member, together with regional sedimentologic and conodont biofacies comparisons, now firmly locates the onset of the Johnson et al. (1985) transgressive-regressive (T-R) cycle IIc, which occurred after the start of the punctata Zone, within a parautochthonous megablock low in the Alamo Breccia. Whole-rock carbon isotope analyses through the lower Guilmette Formation and Alamo Breccia Member reveal two positive ??13Ccarb excursions: (1) a small, 3??? excursion, which is possibly correlative with the falsiovalis Event previously identified from sections in Western Europe and Australia, occurs below the breccia in the Upper falsiovalis Zone to early part of the transitans Zone; and (2) a large, multi-part excursion, dominated by a 6??? positive shift, begins above the start of the punctata Zone and onset of T-R cycle IIc and continues above the Alamo Breccia, ending near the punctata- hassi zonal boundary. This large excursion correlates with the punctata Event, a major positive ??13C excursion previously recognized in eastern Laurussia and northern Gondwana. Consistent with previous studies, at Hancock Summit West the punctata Event is apparently not associated with any regional extinctions or ecosystem reorganizations. In the study area, onset of the

  18. CICE, The Los Alamos Sea Ice Model

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hunke, Elizabeth; Lipscomb, William; Jones, Philip

    The Los Alamos sea ice model (CICE) is the result of an effort to develop a computationally efficient sea ice component for a fully coupled atmosphere–land–ocean–ice global climate model. It was originally designed to be compatible with the Parallel Ocean Program (POP), an ocean circulation model developed at Los Alamos National Laboratory for use on massively parallel computers. CICE has several interacting components: a vertical thermodynamic model that computes local growth rates of snow and ice due to vertical conductive, radiative and turbulent fluxes, along with snowfall; an elastic-viscous-plastic model of ice dynamics, which predicts the velocity field of themore » ice pack based on a model of the material strength of the ice; an incremental remapping transport model that describes horizontal advection of the areal concentration, ice and snow volume and other state variables; and a ridging parameterization that transfers ice among thickness categories based on energetic balances and rates of strain. It also includes a biogeochemical model that describes evolution of the ice ecosystem. The CICE sea ice model is used for climate research as one component of complex global earth system models that include atmosphere, land, ocean and biogeochemistry components. It is also used for operational sea ice forecasting in the polar regions and in numerical weather prediction models.« less

  19. The Los Alamos Supernova Light Curve Project: Current Projects and Future Directions

    NASA Astrophysics Data System (ADS)

    Wiggins, Brandon Kerry; Los Alamos Supernovae Research Group

    2015-01-01

    The Los Alamos Supernova Light Curve Project models supernovae in the ancient and modern universe to determine the luminosities of observability of certain supernovae events and to explore the physics of supernovae in the local universe. The project utilizes RAGE, Los Alamos' radiation hydrodynamics code to evolve the explosions of progenitors prepared in well-established stellar evolution codes. RAGE allows us to capture events such as shock breakout and collisions of ejecta with shells of material which cannot be modeled well in other codes. RAGE's dumps are then ported to LANL's SPECTRUM code which uses LANL's OPLIB opacities database to calculate light curves and spectra. In this paper, we summarize our recent work in modeling supernovae.

  20. Critical Infrastructure Protection- Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bofman, Ryan K.

    Los Alamos National Laboratory (LANL) has been a key facet of Critical National Infrastructure since the nuclear bombing of Hiroshima exposed the nature of the Laboratory’s work in 1945. Common knowledge of the nature of sensitive information contained here presents a necessity to protect this critical infrastructure as a matter of national security. This protection occurs in multiple forms beginning with physical security, followed by cybersecurity, safeguarding of classified information, and concluded by the missions of the National Nuclear Security Administration.

  1. Amphibians and Reptiles of Los Alamos County

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Teralene S. Foxx; Timothy K. Haarmann; David C. Keller

    Recent studies have shown that amphibians and reptiles are good indicators of environmental health. They live in terrestrial and aquatic environments and are often the first animals to be affected by environmental change. This publication provides baseline information about amphibians and reptiles that are present on the Pajarito Plateau. Ten years of data collection and observations by researchers at Los Alamos National Laboratory, the University of New Mexico, the New Mexico Department of Game and Fish, and hobbyists are represented.

  2. Portable MRI developed at Los Alamos

    ScienceCinema

    Espy, Michelle

    2018-02-14

    Scientists at Los Alamos National Laboratory are developing an ultra-low-field Magnetic Resonance Imaging (MRI) system that could be low-power and lightweight enough for forward deployment on the battlefield and to field hospitals in the World's poorest regions. "MRI technology is a powerful medical diagnostic tool," said Michelle Espy, the Battlefield MRI (bMRI) project leader, "ideally suited for imaging soft-tissue injury, particularly to the brain." But hospital-based MRI devices are big and expensive, and require considerable infrastructure, such as large quantities of cryogens like liquid nitrogen and helium, and they typically use a large amount of energy. "Standard MRI machines just can't go everywhere," said Espy. "Soldiers wounded in battle usually have to be flown to a large hospital and people in emerging nations just don't have access to MRI at all. We've been in contact with doctors who routinely work in the Third World and report that MRI would be extremely valuable in treating pediatric encephalopathy, and other serious diseases in children." So the Los Alamos team started thinking about a way to make an MRI device that could be relatively easy to transport, set up, and use in an unconventional setting. Conventional MRI machines use very large magnetic fields that align the protons in water molecules to then create magnetic resonance signals, which are detected by the machine and turned into images. The large magnetic fields create exceptionally detailed images, but they are difficult and expensive to make. Espy and her team wanted to see if images of sufficient quality could be made with ultra-low-magnetic fields, similar in strength to the Earth's magnetic field. To achieve images at such low fields they use exquisitely sensitive detectors called Superconducting Quantum Interference Devices, or SQUIDs. SQUIDs are among the most sensitive magnetic field detectors available, so interference with the signal is the primary stumbling block. "SQUIDs are

  3. Portable MRI developed at Los Alamos

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Espy, Michelle

    Scientists at Los Alamos National Laboratory are developing an ultra-low-field Magnetic Resonance Imaging (MRI) system that could be low-power and lightweight enough for forward deployment on the battlefield and to field hospitals in the World's poorest regions. "MRI technology is a powerful medical diagnostic tool," said Michelle Espy, the Battlefield MRI (bMRI) project leader, "ideally suited for imaging soft-tissue injury, particularly to the brain." But hospital-based MRI devices are big and expensive, and require considerable infrastructure, such as large quantities of cryogens like liquid nitrogen and helium, and they typically use a large amount of energy. "Standard MRI machines justmore » can't go everywhere," said Espy. "Soldiers wounded in battle usually have to be flown to a large hospital and people in emerging nations just don't have access to MRI at all. We've been in contact with doctors who routinely work in the Third World and report that MRI would be extremely valuable in treating pediatric encephalopathy, and other serious diseases in children." So the Los Alamos team started thinking about a way to make an MRI device that could be relatively easy to transport, set up, and use in an unconventional setting. Conventional MRI machines use very large magnetic fields that align the protons in water molecules to then create magnetic resonance signals, which are detected by the machine and turned into images. The large magnetic fields create exceptionally detailed images, but they are difficult and expensive to make. Espy and her team wanted to see if images of sufficient quality could be made with ultra-low-magnetic fields, similar in strength to the Earth's magnetic field. To achieve images at such low fields they use exquisitely sensitive detectors called Superconducting Quantum Interference Devices, or SQUIDs. SQUIDs are among the most sensitive magnetic field detectors available, so interference with the signal is the primary stumbling block

  4. Geohydrology and simulation of ground-water flow near Los Alamos, north-central New Mexico

    USGS Publications Warehouse

    Frenzel, P.F.

    1995-01-01

    An existing model was modified in recognition of new geohydrologic interpretations and adjusted to simulate hydrographs in well fields in the Los Alamos area. Hydraulic-head drawdowns at the Buckman well field resulting from two projected ground-water-withdrawal alternatives were estimated with the modified model. The Chaquehui formation (informal usage) is the main new feature of recent hydrologic interpretations for the Los Alamos area. The Chaquehui occupies a 'channel' that was eroded or faulted into the Tesuque Formation, and the Chaquehui is more permeable than the Tesuque. The Chaquehui is a major producing zone in the Pajarito Mesa well field and to a lesser extent in the Guaje well field. Model modification included splitting the four layers of the McAda-Wasiolek model (McAda, D.P., and Wasiolek, Maryann, 1988, Simulation of the regional geohydrology of the Tesuque aquifer system near Santa Fe, New Mexico: U.S. Geological Survey Water- Resources Investigations Report 87-4056, 71 p.) into eight layers to better simulate vertical ground-water movement. Other model modifications were limited as much as possible to the area of interest near Los Alamos and consisted mainly of adjusting hydraulic-conductivity values representing the Tesuque Formation, Chaquehui formation (informal usage), and Puye Formation, and adjusting simulated recharge along the Pajarito Fault Zone west of Los Alamos. Adjustments were based mainly on simulation of fluctuations in measured hydraulic heads near Los Alamos. Two possible alternative plans for replacing Guaje well field production were suggested by Los Alamos National Laboratory. In the first plan (Guaje alternative), the Guaje field would be renewed with four new wells replacing the existing production wells in the Guaje field. In the second plan (Pajarito-Otowi alternative), the Guaje well field would be retired and its former production would be made up by additional withdrawals from the Pajarito Mesa and Otowi well fields. A

  5. The Los Alamos universe: Using multimedia to promote laboratory capabilities

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kindel, J.

    2000-03-01

    This project consists of a multimedia presentation that explains the technological capabilities of Los Alamos National Laboratory. It takes the form of a human-computer interface built around the metaphor of the universe. The project is intended promote Laboratory capabilities to a wide audience. Multimedia is simply a means of communicating information through a diverse set of tools--be they text, sound, animation, video, etc. Likewise, Los Alamos National Laboratory is a collection of diverse technologies, projects, and people. Given the ample material available at the Laboratory, there are tangible benefits to be gained by communicating across media. This paper consists ofmore » three parts. The first section provides some basic information about the Laboratory, its mission, and its needs. The second section introduces this multimedia presentation and the metaphor it is based on along with some basic concepts of color and user interaction used in the building of this project. The final section covers construction of the project, pitfalls, and future improvements.« less

  6. Analysis of Radiation Effects in Silicon using Kinetic Monte Carlo Methods

    DOE PAGES

    Hehr, Brian Douglas

    2014-11-25

    The transient degradation of semiconductor device performance under irradiation has long been an issue of concern. Neutron irradiation can instigate the formation of quasi-stable defect structures, thereby introducing new energy levels into the bandgap that alter carrier lifetimes and give rise to such phenomena as gain degradation in bipolar junction transistors. Normally, the initial defect formation phase is followed by a recovery phase in which defect-defect or defect-dopant interactions modify the characteristics of the damaged structure. A kinetic Monte Carlo (KMC) code has been developed to model both thermal and carrier injection annealing of initial defect structures in semiconductor materials.more » The code is employed to investigate annealing in electron-irradiated, p-type silicon as well as the recovery of base current in silicon transistors bombarded with neutrons at the Los Alamos Neutron Science Center (LANSCE) “Blue Room” facility. Our results reveal that KMC calculations agree well with these experiments once adjustments are made, within the appropriate uncertainty bounds, to some of the sensitive defect parameters.« less

  7. Tritium concentrations in bees and honey at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fresquez, P.R.; Armstrong, D.R.; Salazar, J.G.

    Los Alamos National Laboratory (LANL) has maintained a network of honey bee colonies at LANL, perimeter (Los Alamos townsite and White Rock/Pajarito Acres) and regional (background) areas for over 15 years; the main objective of this honey bee network was to help determine the bioavailability of certain radionuclides in the environment. Of all the radionuclides studied ({sup 3}H, {sup 57}Co, {sup 7}Be, {sup 22}Na, {sup 54}Mn, {sup 83}Rb, {sup 137}Cs, {sup 238}Pu, {sup 239}Pu, {sup 90}Sr and total U), tritium was consistently detected in bees and was most readily transferred to the honey. In fact, honey collected from hives locatedmore » at TA-21, TA-33, TA-50, TA-53, and TA-54 and from White Rock/Pajarito Acres contained significantly higher concentrations of {sup 3}H than regional background hives. Based on the average concentration of all radionuclides measured over the years, the effective dose equivalent (EDE) from consuming 5 kg (11 lb) of honey collected from Los Alamos (townsite) and White Rock/Pajarito Acres, after regional background has been subtracted, was 0.0186 ({+-}0.0507) and 0.0016 ({+-}0.0010) mrem/yr, respectively. The highest EDE, based on the mean + 2SD (95% confidence level), was 0.1200 mrem/y; this was <0.2% of the International Commission on Radiological Protection permissible dose limit of 100 mrem/yr from all pathways.« less

  8. Benchmarking the MCNP Monte Carlo code with a photon skyshine experiment

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Olsher, R.H.; Hsu, Hsiao Hua; Harvey, W.F.

    1993-07-01

    The MCNP Monte Carlo transport code is used by the Los Alamos National Laboratory Health and Safety Division for a broad spectrum of radiation shielding calculations. One such application involves the determination of skyshine dose for a variety of photon sources. To verify the accuracy of the code, it was benchmarked with the Kansas State Univ. (KSU) photon skyshine experiment of 1977. The KSU experiment for the unshielded source geometry was simulated in great detail to include the contribution of groundshine, in-silo photon scatter, and the effect of spectral degradation in the source capsule. The standard deviation of the KSUmore » experimental data was stated to be 7%, while the statistical uncertainty of the simulation was kept at or under 1%. The results of the simulation agreed closely with the experimental data, generally to within 6%. At distances of under 100 m from the silo, the modeling of the in-silo scatter was crucial to achieving close agreement with the experiment. Specifically, scatter off the top layer of the source cask accounted for [approximately]12% of the dose at 50 m. At distance >300m, using the [sup 60]Co line spectrum led to a dose overresponse as great as 19% at 700 m. It was necessary to use the actual source spectrum, which includes a Compton tail from photon collisions in the source capsule, to achieve close agreement with experimental data. These results highlight the importance of using Monte Carlo transport techniques to account for the nonideal features of even simple experiments''.« less

  9. Examination of the home destruction in Los Alamos associated with the Cerro Grande Fire - July 10, 2000

    Treesearch

    Jack D. Cohen

    2000-01-01

    I arrived at Los Alamos on May 14, 2000 to conduct an examination of the home destruction associated with the Cerro Grande Fire. My examination occurred between the afternoon of 5/14 and late afternoon on 5/16. I had contact with the southern command post incident management team, the Los Alamos Fire Department, and the Santa Fe National Forest.The...

  10. Monte Carlo simulations for generic granite repository studies

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Chu, Shaoping; Lee, Joon H; Wang, Yifeng

    In a collaborative study between Los Alamos National Laboratory (LANL) and Sandia National Laboratories (SNL) for the DOE-NE Office of Fuel Cycle Technologies Used Fuel Disposition (UFD) Campaign project, we have conducted preliminary system-level analyses to support the development of a long-term strategy for geologic disposal of high-level radioactive waste. A general modeling framework consisting of a near- and a far-field submodel for a granite GDSE was developed. A representative far-field transport model for a generic granite repository was merged with an integrated systems (GoldSim) near-field model. Integrated Monte Carlo model runs with the combined near- and farfield transport modelsmore » were performed, and the parameter sensitivities were evaluated for the combined system. In addition, a sub-set of radionuclides that are potentially important to repository performance were identified and evaluated for a series of model runs. The analyses were conducted with different waste inventory scenarios. Analyses were also conducted for different repository radionuelide release scenarios. While the results to date are for a generic granite repository, the work establishes the method to be used in the future to provide guidance on the development of strategy for long-term disposal of high-level radioactive waste in a granite repository.« less

  11. The Los Alamos suite of relativistic atomic physics codes

    DOE PAGES

    Fontes, C. J.; Zhang, H. L.; Jr, J. Abdallah; ...

    2015-05-28

    The Los Alamos SuitE of Relativistic (LASER) atomic physics codes is a robust, mature platform that has been used to model highly charged ions in a variety of ways. The suite includes capabilities for calculating data related to fundamental atomic structure, as well as the processes of photoexcitation, electron-impact excitation and ionization, photoionization and autoionization within a consistent framework. These data can be of a basic nature, such as cross sections and collision strengths, which are useful in making predictions that can be compared with experiments to test fundamental theories of highly charged ions, such as quantum electrodynamics. The suitemore » can also be used to generate detailed models of energy levels and rate coefficients, and to apply them in the collisional-radiative modeling of plasmas over a wide range of conditions. Such modeling is useful, for example, in the interpretation of spectra generated by a variety of plasmas. In this work, we provide a brief overview of the capabilities within the Los Alamos relativistic suite along with some examples of its application to the modeling of highly charged ions.« less

  12. Proton Radiography at Los Alamos

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Saunders, Alexander

    2017-02-28

    The proton radiography (pRad) facility at Los Alamos National Lab uses high energy protons to acquire multiple frame flash radiographic sequences at megahertz speeds: that is, it can make movies of the inside of explosions as they happen. The facility is primarily used to study the damage to and failure of metals subjected to the shock forces of high explosives as well as to study the detonation of the explosives themselves. Applications include improving our understanding of the underlying physical processes that drive the performance of the nuclear weapons in the United States stockpile and developing novel armor technologies inmore » collaboration with the Army Research Lab. The principle and techniques of pRad will be described, and examples of some recent results will be shown.« less

  13. Keeping the Momentum and Nuclear Forensics at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Steiner, Robert Ernest; Dion, Heather M.; Dry, Donald E.

    LANL has 70 years of experience in nuclear forensics and supports the community through a wide variety of efforts and leveraged capabilities: Expanding the understanding of nuclear forensics, providing training on nuclear forensics methods, and developing bilateral relationships to expand our understanding of nuclear forensic science. LANL remains highly supportive of several key organizations tasked with carrying forth the Nuclear Security Summit messages: IAEA, GICNT, and INTERPOL. Analytical chemistry measurements on plutonium and uranium matrices are critical to numerous programs including safeguards accountancy verification measurements. Los Alamos National Laboratory operates capable actinide analytical chemistry and material science laboratories suitable formore » nuclear material and environmental forensic characterization. Los Alamos National Laboratory uses numerous means to validate and independently verify that measurement data quality objectives are met. Numerous LANL nuclear facilities support the nuclear material handling, preparation, and analysis capabilities necessary to evaluate samples containing nearly any mass of an actinide (attogram to kilogram levels).« less

  14. Dissolved pesticides in the Alamo River and the Salton Sea, California, 1996-97

    USGS Publications Warehouse

    Crepeau, Kathryn L.; Kuivila, Kathryn; Bergamaschi, Brian A.

    2002-01-01

    Water samples were collected from the Alamo River and the Salton Sea, California, in autumn 1996 and late winter/early spring 1997 and analyzed for dissolved pesticides. The two seasons chosen for sampling were during pesticide application periods in the Imperial Valley. Pesticide concentrations were measured in filtered water samples using solid-phase extraction and analyzed by gas chromatography/mass spectrometry. Generally, the highest concentrations were measured in the Alamo River. The concentrations of carbaryl, chlorpyrifos, cycloate, dacthal, diazinon, and eptam were highest in samples collected in autumn 1996. In contrast, the concentrations of atrazine, carbofuran, and malathion were highest in samples collected in late winter/early spring 1997. The highest concentrations measured of atrazine, carbofuran, dacthal, eptam, and malathion all exceeded 1,000 nanograms per liter.

  15. Los Alamos Scientific Laboratory energy-related history, research, managerial reorganization proposals, actions taken, and results. History report, 1945--1979

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hammel, E.F.

    1997-03-01

    This report documents the development of major energy-related programs at the Los Alamos Scientific Laboratory between 1945 and 1979. Although the Laboratory`s primary mission during that era was the design and development of nuclear weapons and most of the Laboratory`s funding came from a single source, a number of factors were at work that led to the development of these other programs. Some of those factors were affected by the Laboratory`s internal management structure and organization; others were the result of increasing environmental awareness within the general population and the political consequences of that awareness; still others were related tomore » the increasing demand for energy and the increasing turmoil in the energy-rich Middle East. This report also describes the various activities in Los Alamos, in Washington, and in other areas of the world that contributed to the development of major energy-related programs at Los Alamos. The author has a unique historical perspective because of his involvement as a scientist and manager at the Los Alamos Scientific Laboratory during the time period described within the report. In addition, in numerous footnotes and references, he cites a large body of documents that include the opinions and perspectives of many others who were involved at one time or another in these programs. Finally the report includes a detailed chronology of geopolitical events that led to the development of energy-related programs at Los Alamos.« less

  16. An Analysis on the TEC Variability and Ionospheric Scintillation at Los Alamos, New Mexico Derived from FORTE-Received LAPP Signals

    NASA Astrophysics Data System (ADS)

    Huang, Z.; Roussel-Dupre, R.

    2003-12-01

    The total electron content (TEC) of ionosphere and its electron density irregularities (scintillation) have effects of degradation and disruption on radio signals passed between ground stations and orbiting man-made satellites. With the rapid increase in operational reliance on UHF/VHF satellite communication, it is desirable to obtain understandings of ionosphere TEC variability and scintillation characteristics to improve our ability of predicting satellite communication outages. In this work, data collected from FORTE satellite received LAPP (Los Alamos Portable Pulser) signals during 1998-2002 are used to derive TEC and ionospheric scintillation index at Los Alamos, New Mexico. To characterize in-situ TEC variability at Los Alamos, the FORTE-LAPP derived TECs are analyzed against diurnal, seasonal, solar activity, magnetic storm, and stratospheric warming. The results are also compared with the TEC estimates from the Los Alamos ionospheric transfer function (ITF) implemented with the global ionospheric models (IRI, PIM), and GPS -derived TEC maps. The FORTE-LAPP signals are also analyzed against two important measures of the effect of scintillation on broadband signals, the mean time delay and the time delay jitter. The results are used to examine coherence frequency bandwidth and compared with the predictions from a global scintillation model (WBMOD). The FORTE-LAPP analyzed and WBMOD predicted scintillation characteristics are used to investigate temporal and seasonal behavior of scintillation at Los Alamos.

  17. Igniting the Light Elements: The Los Alamos Thermonuclear Weapon Project, 1942-1952

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fitzpatrick, Anne C.

    1999-07-01

    The American system of nuclear weapons research and development was conceived and developed not as a result of technological determinism, but by a number of individual architects who promoted the growth of this large technologically-based complex. While some of the technological artifacts of this system, such as the fission weapons used in World War II, have been the subject of many historical studies, their technical successors--fusion (or hydrogen) devices--are representative of the largely unstudied highly secret realms of nuclear weapons science and engineering. In the postwar period a small number of Los Alamos Scientific Laboratory's staff and affiliates were responsiblemore » for theoretical work on fusion weapons, yet the program was subject to both the provisions and constraints of the US Atomic Energy Commission, of which Los Alamos was a part. The Commission leadership's struggle to establish a mission for its network of laboratories, least of all to keep them operating, affected Los Alamos's leaders' decisions as to the course of weapons design and development projects. Adapting Thomas P. Hughes's ''large technological systems'' thesis, I focus on the technical, social, political, and human problems that nuclear weapons scientists faced while pursuing the thermonuclear project, demonstrating why the early American thermonuclear bomb project was an immensely complicated scientific and technological undertaking. I concentrate mainly on Los Alamos Scientific Laboratory's Theoretical, or T, Division, and its members' attempts to complete an accurate mathematical treatment of the ''Super''--the most difficult problem in physics in the postwar period--and other fusion weapon theories. Although tackling a theoretical problem, theoreticians had to address technical and engineering issues as well. I demonstrate the relative value and importance of H-bomb research over time in the postwar era to scientific, politician, and military participants in this project

  18. Integrating Safety with Science,Technology and Innovation at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Rich, Bethany M

    2012-04-02

    The mission of Los Alamos National Laboratory (LANL) is to develop and apply science, technology and engineering solutions to ensure the safety, security, and reliability of the U.S. nuclear deterrent; reduce global threats; and solve emerging national security challenges. The most important responsibility is to direct and conduct efforts to meet the mission with an emphasis on safety, security, and quality. In this article, LANL Environmental, Safety, and Health (ESH) trainers discuss how their application and use of a kinetic learning module (learn by doing) with a unique fall arrest system is helping to address one the most common industrialmore » safety challenges: slips and falls. A unique integration of Human Performance Improvement (HPI), Behavior Based Safety (BBS) and elements of the Voluntary Protection Program (VPP) combined with an interactive simulator experience is being used to address slip and fall events at Los Alamos.« less

  19. Environmental surveillance at Los Alamos during 2005

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    None

    2006-09-30

    Environmental Surveillance at Los Alamos reports are prepared annually by the Los Alamos National Laboratory (LANL or the Laboratory) environmental organization, as required by US Department of Energy Order 5400.1, General Environmental Protection Program, and US Department of Energy Order 231.IA, Environment, Safety, and Health Reporting. These annual reports summarize environmental data that are used to determine compliance with applicable federal, state, and local environmental laws and regulations, executive orders, and departmental policies. Additional data, beyond the minimum required, are also gathered and reported as part of the Laboratory's efforts to ensure public safety and to monitor environmental quality atmore » and near the Laboratory. Chapter 1 provides an overview of the Laboratory's major environmental programs. Chapter 2 reports the Laboratory's compliance status for 2005. Chapter 3 provides a summary of the maximum radiological dose the public and biota populations could have potentially received from Laboratory operations. The environmental surveillance and monitoring data are organized by environmental media (Chapter 4, Air; Chapters 5 and 6, Water and Sediments; Chapter 7, Soils; and Chapter 8, Foodstuffs and Biota) in a format to meet the needs of a general and scientific audience. Chapter 9, new for this year, provides a summary of the status of environmental restoration work around LANL. A glossary and a list ofacronyms and abbreviations are in the back of the report. Appendix A explains the standards for environmental contaminants, Appendix B explains the units of measurements used in this report, Appendix C describes the Laboratory's technical areas and their associated programs, and Appendix D provides web links to more information.« less

  20. Surface water data at Los Alamos National Laboratory: 2009 water year

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ortiz, David; McCullough, Betsy

    2010-05-01

    The principal investigators collected and computed surface water discharge data from 73 stream-gage stations that cover most of Los Alamos National Laboratory and one at Bandelier National Monument. Also included are discharge data from three springs— two that flow into Cañon de Valle and one that flows into Water Canyon.

  1. Surface water data at Los Alamos National Laboratory: 2008 water year

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ortiz, David; Cata, Betsy; Kuyumjian, Gregory

    2009-09-01

    The principal investigators collected and computed surface water discharge data from 69 stream-gage stations that cover most of Los Alamos National Laboratory and one at Bandelier National Monument. Also included are discharge data from three springs— two that flow into Cañon de Valle and one that flows into Water Canyon.

  2. Fuels Inventories in the Los Alamos National Laboratory Region: 1997

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Balice, R.G.; Oswald, B.P.; Martin, C.

    1999-03-01

    Fifty-four sites were surveyed for fuel levels, vegetational structures, and topographic characteristics. Most of the surveyed sites were on Los Alamos National Laboratory property, however, some surveys were also conducted on U.S. Forest Service property. The overall vegetation of these sites ranged from pinon-juniper woodlands to ponderosa pine forests to mixed conifer forests, and the topographic positions included canyons, mesas, and mountains. The results of these surveys indicate that the understory fuels are the greatest in mixed conifer forests and that overstory fuels are greatest in both mixed conifer forests and ponderosa pine forests on mesas. The geographic distribution ofmore » these fuels would suggest a most credible wildfire scenario for the Los Alamos region. Three major fires have occurred since 1954 and these fires behaved in a manner that is consistent with this scenario. The most credible wildfire scenario was also supported by the results of BEHAVE modeling that used the fuels inventory data as inputs. Output from the BEHAVE model suggested that catastrophic wildfires would continue to occur during any season with sufficiently dry, windy weather.« less

  3. Los Alamos National Laboratory Economic Analysis Capability Overview

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Boero, Riccardo; Edwards, Brian Keith; Pasqualini, Donatella

    Los Alamos National Laboratory has developed two types of models to compute the economic impact of infrastructure disruptions. FastEcon is a fast running model that estimates first-­order economic impacts of large scale events such as hurricanes and floods and can be used to identify the amount of economic activity that occurs in a specific area. LANL’s Computable General Equilibrium (CGE) model estimates more comprehensive static and dynamic economic impacts of a broader array of events and captures the interactions between sectors and industries when estimating economic impacts.

  4. Performance of the upgraded ultracold neutron source at Los Alamos National Laboratory and its implication for a possible neutron electric dipole moment experiment

    DOE PAGES

    Ito, Takeyasu M.; Adamek, E. R.; Callahan, N. B.; ...

    2018-01-29

    We report the ultracold neutron (UCN) source at Los Alamos National Laboratory (LANL), which uses solid deuterium as the UCN converter and is driven by accelerator spallation neutrons, has been successfully operated for over 10 years, providing UCN to various experiments, as the first production UCN source based on the superthermal process. It has recently undergone a major upgrade. This paper describes the design and performance of the upgraded LANL UCN source. Measurements of the cold neutron spectrum and UCN density are presented and compared to Monte Carlo predictions. The source is shown to perform as modeled. The UCN densitymore » measured at the exit of the biological shield was 184(32) UCN / cm 3, a fourfold increase from the highest previously reported. Finally, the polarized UCN density stored in an external chamber was measured to be 39(7) UCN / cm 3, which is sufficient to perform an experiment to search for the nonzero neutron electric dipole moment with a one-standard-deviation sensitivity of σ(d n) = 3 × 10 -27 e cm.« less

  5. Performance of the upgraded ultracold neutron source at Los Alamos National Laboratory and its implication for a possible neutron electric dipole moment experiment

    NASA Astrophysics Data System (ADS)

    Ito, T. M.; Adamek, E. R.; Callahan, N. B.; Choi, J. H.; Clayton, S. M.; Cude-Woods, C.; Currie, S.; Ding, X.; Fellers, D. E.; Geltenbort, P.; Lamoreaux, S. K.; Liu, C.-Y.; MacDonald, S.; Makela, M.; Morris, C. L.; Pattie, R. W.; Ramsey, J. C.; Salvat, D. J.; Saunders, A.; Sharapov, E. I.; Sjue, S.; Sprow, A. P.; Tang, Z.; Weaver, H. L.; Wei, W.; Young, A. R.

    2018-01-01

    The ultracold neutron (UCN) source at Los Alamos National Laboratory (LANL), which uses solid deuterium as the UCN converter and is driven by accelerator spallation neutrons, has been successfully operated for over 10 years, providing UCN to various experiments, as the first production UCN source based on the superthermal process. It has recently undergone a major upgrade. This paper describes the design and performance of the upgraded LANL UCN source. Measurements of the cold neutron spectrum and UCN density are presented and compared to Monte Carlo predictions. The source is shown to perform as modeled. The UCN density measured at the exit of the biological shield was 184 (32 ) UCN /cm3 , a fourfold increase from the highest previously reported. The polarized UCN density stored in an external chamber was measured to be 39 (7 ) UCN /cm3 , which is sufficient to perform an experiment to search for the nonzero neutron electric dipole moment with a one-standard-deviation sensitivity of σ (dn) =3 ×10-27e cm .

  6. Performance of the upgraded ultracold neutron source at Los Alamos National Laboratory and its implication for a possible neutron electric dipole moment experiment

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ito, Takeyasu M.; Adamek, E. R.; Callahan, N. B.

    We report the ultracold neutron (UCN) source at Los Alamos National Laboratory (LANL), which uses solid deuterium as the UCN converter and is driven by accelerator spallation neutrons, has been successfully operated for over 10 years, providing UCN to various experiments, as the first production UCN source based on the superthermal process. It has recently undergone a major upgrade. This paper describes the design and performance of the upgraded LANL UCN source. Measurements of the cold neutron spectrum and UCN density are presented and compared to Monte Carlo predictions. The source is shown to perform as modeled. The UCN densitymore » measured at the exit of the biological shield was 184(32) UCN / cm 3, a fourfold increase from the highest previously reported. Finally, the polarized UCN density stored in an external chamber was measured to be 39(7) UCN / cm 3, which is sufficient to perform an experiment to search for the nonzero neutron electric dipole moment with a one-standard-deviation sensitivity of σ(d n) = 3 × 10 -27 e cm.« less

  7. Hydrologic transport of depleted uranium associated with open air dynamic range testing at Los Alamos National Laboratory, New Mexico, and Eglin Air Force Base, Florida

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Becker, N.M.; Vanta, E.B.

    Hydrologic investigations on depleted uranium fate and transport associated with dynamic testing activities were instituted in the 1980`s at Los Alamos National Laboratory and Eglin Air Force Base. At Los Alamos, extensive field watershed investigations of soil, sediment, and especially runoff water were conducted. Eglin conducted field investigations and runoff studies similar to those at Los Alamos at former and active test ranges. Laboratory experiments complemented the field investigations at both installations. Mass balance calculations were performed to quantify the mass of expended uranium which had transported away from firing sites. At Los Alamos, it is estimated that more thanmore » 90 percent of the uranium still remains in close proximity to firing sites, which has been corroborated by independent calculations. At Eglin, we estimate that 90 to 95 percent of the uranium remains at test ranges. These data demonstrate that uranium moves slowly via surface water, in both semi-arid (Los Alamos) and humid (Eglin) environments.« less

  8. Refinements in the Los Alamos model of the prompt fission neutron spectrum

    DOE PAGES

    Madland, D. G.; Kahler, A. C.

    2017-01-01

    This paper presents a number of refinements to the original Los Alamos model of the prompt fission neutron spectrum and average prompt neutron multiplicity as derived in 1982. The four refinements are due to new measurements of the spectrum and related fission observables many of which were not available in 1982. Here, they are also due to a number of detailed studies and comparisons of the model with previous and present experimental results including not only the differential spectrum, but also integal cross sections measured in the field of the differential spectrum. The four refinements are (a) separate neutron contributionsmore » in binary fission, (b) departure from statistical equilibrium at scission, (c) fission-fragment nuclear level-density models, and (d) center-of-mass anisotropy. With these refinements, for the first time, good agreement has been obtained for both differential and integral measurements using the same Los Alamos model spectrum.« less

  9. Surface Water Data at Los Alamos National Laboratory: 2002 Water Year

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    D.A. Shaull; D. Ortiz; M.R. Alexander

    2003-03-03

    The principal investigators collected and computed surface water discharge data from 34 stream-gaging stations that cover most of Los Alamos National Laboratory and one at Bandelier National Monument. Also included are discharge data from three springs--two that flow into Canon de Valle and one that flows into Water Canyon--and peak flow data from 16 stations.

  10. Surface Water Data at Los Alamos National Laboratory 2006 Water Year

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    R.P. Romero, D. Ortiz, G. Kuyumjian

    2007-08-01

    The principal investigators collected and computed surface water discharge data from 44 stream-gaging stations that cover most of Los Alamos National Laboratory and one at Bandelier National Monument. Also included are discharge data from three springs--two that flow into Canon de Valle and one that flows into Water Canyon--and peak flow data for 44 stations.

  11. Simplifying Complexity: Miriam Blake--Los Alamos National Laboratory Research Library, NM

    ERIC Educational Resources Information Center

    Library Journal, 2004

    2004-01-01

    The holy grail for many research librarians is one-stop searching: seamless access to all the library's resources on a topic, regardless of the source. Miriam Blake, Library Without Walls Project Leader at Los Alamos National laboratory (LANL), is making this vision a reality. Blake is part of a growing cadre of experts: a techie who is becoming a…

  12. Resource Management Technology: Los Alamos Technical Capabilities for Emergency Management,

    DTIC Science & Technology

    1983-07-18

    synthetic fuels from coal (analogous to the Fischer-Tropsch process), olefin polymerization, and flue - gas desulfurization . In order to successfully...world. It has been a major research effort here for decades. Also, in the area of desulfurization of flue gases, Los Alamos scientists have been...Tectonic and Geochemical Controls on Copper-Molybdenum Porphyry Mineralization in the Southwestern United States (M. J. Aldrich and A. W. Laughlin) 1.0.6

  13. Los Alamos Canyon Ice Rink Parking Flood Plain Assessment

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hathcock, Charles Dean; Keller, David Charles

    2015-02-10

    The project location is in Los Alamos Canyon east of the ice rink facility at the intersection of West and Omega roads (Figure 1). Forty eight parking spaces will be constructed on the north and south side of Omega Road, and a lighted walking path will be constructed to the ice rink. Some trees will be removed during this action. A guardrail of approximately 400 feet will be constructed along the north side of West Road to prevent unsafe parking in that area.

  14. Environmental surveillance at Los Alamos during 2008

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fuehne, David; Gallagher, Pat; Hjeresen, Denny

    2009-09-30

    Environmental Surveillance at Los Alamos reports are prepared annually by the Los Alamos National Laboratory (the Laboratory) Environmental Programs Directorate, as required by US Department of Energy Order 450.1, General Environmental Protection Program, and US Department of Energy Order 231.1A, Environment, Safety, and Health Reporting. These annual reports summarize environmental data that are used to determine compliance with applicable federal, state, and local environmental laws and regulations, executive orders, and departmental policies. Additional data, beyond the minimum required, are also gathered and reported as part of the Laboratory’s efforts to ensure public safety and to monitor environmental quality at andmore » near the Laboratory. Chapter 1 provides an overview of the Laboratory’s major environmental programs and explains the risks and the actions taken to reduce risks at the Laboratory from environmental legacies and waste management operations. Chapter 2 reports the Laboratory’s compliance status for 2007. Chapter 3 provides a summary of the maximum radiological dose the public and biota populations could have potentially received from Laboratory operations and discusses chemical exposures. The environmental surveillance and monitoring data are organized by environmental media (Chapter 4, air; Chapters 5 and 6, water and sediments; Chapter 7, soils; and Chapter 8, foodstuffs and biota) in a format to meet the needs of a general and scientific audience. Chapter 9 provides a summary of the status of environmental restoration work around LANL. A glossary and a list of acronyms and abbreviations are in the back of the report. Appendix A explains the standards for environmental contaminants, Appendix B explains the units of measurements used in this report, Appendix C describes the Laboratory’s technical areas and their associated programs, and Appendix D provides web links to more information.« less

  15. An organizational survey of the Los Alamos Site

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Shurberg, D.A.; Haber, S.B.

    An Organizational Survey (OS) was administered at the Los Alamos Site that queried employees on the subjects of organizational culture, various aspects of communications, employee commitment, work group cohesion, coordination of work, environmental, safety, and health concern, hazardous nature of work, safety and overall job satisfaction. The purpose of the OS is to measure in a quantitative and objective way the notion of culture;'' that is, the values, attitudes, and beliefs of the individuals working within the organization. In addition, through the OS, a broad sample of individuals can be reached that would probably not be interviewed or observed duringmore » the course of a typical assessment. The OS also provides a descriptive profile of the organization at one point in time that can then be compared to a profile taken at a different point in time to assess changes in the culture of the organization. While comparisons among groups are made, it is not the purpose of this report to make evaluative statements of which profile may be positive or negative. However, using the data presented in this report in conjunction with other evaluative activities, may provide useful insight into the organization. The OS administration at the Los Alamos Site was the ninth to occur at a Department of Energy (DOE) facility. All data from the OS is presented in group summaries, by organization, department or directorate within organization, supervisory level both overall and within organization, and staff classification within organization. Statistically significant differences between groups are identified and discussed. 9 refs., 94 figs., 11 tabs.« less

  16. An organizational survey of the Los Alamos Site

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Shurberg, D.A.; Haber, S.B.

    An Organizational Survey (OS) was administered at the Los Alamos Site that queried employees on the subjects of organizational culture, various aspects of communications, employee commitment, work group cohesion, coordination of work, environmental, safety, and health concern, hazardous nature of work, safety and overall job satisfaction. The purpose of the OS is to measure in a quantitative and objective way the notion of ``culture;`` that is, the values, attitudes, and beliefs of the individuals working within the organization. In addition, through the OS, a broad sample of individuals can be reached that would probably not be interviewed or observed duringmore » the course of a typical assessment. The OS also provides a descriptive profile of the organization at one point in time that can then be compared to a profile taken at a different point in time to assess changes in the culture of the organization. While comparisons among groups are made, it is not the purpose of this report to make evaluative statements of which profile may be positive or negative. However, using the data presented in this report in conjunction with other evaluative activities, may provide useful insight into the organization. The OS administration at the Los Alamos Site was the ninth to occur at a Department of Energy (DOE) facility. All data from the OS is presented in group summaries, by organization, department or directorate within organization, supervisory level both overall and within organization, and staff classification within organization. Statistically significant differences between groups are identified and discussed. 9 refs., 94 figs., 11 tabs.« less

  17. New geochronologic and stratigraphic evidence confirms the paleocene age of the dinosaur-bearing ojo alamo sandstone and animas formation in the San Juan Basin, New Mexico and Colorado

    USGS Publications Warehouse

    Fassett, J.E.

    2009-01-01

    Dinosaur fossils are present in the Paleocene Ojo Alamo Sandstone and Animas Formation in the San Juan Basin, New Mexico, and Colorado. Evidence for the Paleo-cene age of the Ojo Alamo Sandstone includes palynologic and paleomagnetic data. Palynologic data indicate that the entire Ojo Alamo Sandstone, including the lower dinosaur-bearing part, is Paleocene in age. All of the palynomorph-productive rock samples collected from the Ojo Alamo Sandstone at multiple localities lacked Creta-ceous index palynomorphs (except for rare, reworked specimens) and produced Paleocene index palynomorphs. Paleocene palynomorphs have been identified strati-graphically below dinosaur fossils at two separate localities in the Ojo Alamo Sand-stone in the central and southern parts of the basin. The Animas Formation in the Colorado part of the basin also contains dinosaur fossils, and its Paleocene age has been established based on fossil leaves and palynology. Magnetostratigraphy provides independent evidence for the Paleocene age of the Ojo Alamo Sandstone and its dinosaur-bearing beds. Normal-polarity magnetochron C29n (early Paleocene) has been identified in the Ojo Alamo Sandstone at six localities in the southern part of the San Juan Basin. An assemblage of 34 skeletal elements from a single hadrosaur, found in the Ojo Alamo Sandstone in the southern San Juan Basin, provided conclusive evidence that this assemblage could not have been reworked from underlying Cretaceous strata. In addition, geochemical studies of 15 vertebrate bones from the Paleocene Ojo Alamo Sandstone and 15 bone samples from the underlying Kirtland Formation of Late Creta-ceous (Campanian) age show that each sample suite contained distinctly different abundances of uranium and rare-earth elements, indicating that the bones were miner-alized in place soon after burial, and that none of the Paleocene dinosaur bones ana-lyzed had been reworked. ?? U.S. Geological Survey, Public Domain April 2009.

  18. Tiger Team Assessment of the Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Not Available

    1991-11-01

    The purpose of the safety and health assessment was to determine the effectiveness of representative safety and health programs at the Los Alamos National Laboratory (LANL). Within the safety and health programs at LANL, performance was assessed in the following technical areas: Organization and Administration, Quality Verification, Operations, Maintenance, Training and Certification, Auxiliary Systems, Emergency Preparedness, Technical Support, Packaging and Transportation, Nuclear Criticality Safety, Security/Safety Interface, Experimental Activities, Site/Facility Safety Review, Radiological Protection, Personnel Protection, Worker Safety and Health (OSHA) Compliance, Fire Protection, Aviation Safety, Explosives Safety, Natural Phenomena, and Medical Services.

  19. Growth promotion and colonization of switchgrass (Panicum virgatum) cv. Alamo by bacterial endophyte Burkholderia phytofirmans strain PsJN

    PubMed Central

    2012-01-01

    Background Switchgrass is one of the most promising bioenergy crop candidates for the US. It gives relatively high biomass yield and can grow on marginal lands. However, its yields vary from year to year and from location to location. Thus it is imperative to develop a low input and sustainable switchgrass feedstock production system. One of the most feasible ways to increase biomass yields is to harness benefits of microbial endophytes. Results We demonstrate that one of the most studied plant growth promoting bacterial endophytes, Burkholderia phytofirmans strain PsJN, is able to colonize and significantly promote growth of switchgrass cv. Alamo under in vitro, growth chamber, and greenhouse conditions. In several in vitro experiments, the average fresh weight of PsJN-inoculated plants was approximately 50% higher than non-inoculated plants. When one-month-old seedlings were grown in a growth chamber for 30 days, the PsJN-inoculated Alamo plants had significantly higher shoot and root biomass compared to controls. Biomass yield (dry weight) averaged from five experiments was 54.1% higher in the inoculated treatment compared to non-inoculated control. Similar results were obtained in greenhouse experiments with transplants grown in 4-gallon pots for two months. The inoculated plants exhibited more early tillers and persistent growth vigor with 48.6% higher biomass than controls. We also found that PsJN could significantly promote growth of switchgrass cv. Alamo under sub-optimal conditions. However, PsJN-mediated growth promotion in switchgrass is genotype specific. Conclusions Our results show B. phytofirmans strain PsJN significantly promotes growth of switchgrass cv. Alamo under different conditions, especially in the early growth stages leading to enhanced production of tillers. This phenomenon may benefit switchgrass establishment in the first year. Moreover, PsJN significantly stimulated growth of switchgrass cv. Alamo under sub-optimal conditions

  20. Radonuclide concentrations in bees and honey in the vicinity of Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fresquez, P.R.; Armstrong, D.R.

    Honeybees are effective monitors of environmental pollution; they forage for P len and nectar over a large area ({congruent}7 km{sup 2}), accumulate contaminants from air, water, plants, and soil, and return to a fixed location (the hive) for sampling. Los Alamos National Laboratory (LANL), in fact, has maintained a network of honeybee colonies within and around LANL for 16 years (1979 to 1994); the objectives for maintaining this honeybee network were to (1) determine the bioavailability of radionuclides in the environment, and (2) the committed effective dose equivalent (CEDE) to people who may consume honey from these beehives (Los Alamosmore » and White Rock/Pajarito Acres lownsites). Of all the radionuclides studied over the years, tritium (314) was consistently picked up by the bees and was most readily transferred to the honey. Tritium in honey collected from hives located within LANL, for example, ranged in concentration from 0.07 Bq mL{sup -1} (1.9 pCi mL{sup -1}) to 27.75 Bq mL{sup -1} (749.9 pCi mL{sup -1}) (LANL Neutron Science Center); the average concentration of {sup 3}H in honey Collected from hives located around the LANL area (perimeter) ranged in concentration from 0.34 Bq mL{sup -1} (9.3 pCi mL{sup -1}) (White Rock/Pajarito Acres townsite) to 3.67 Bq mL{sup -1} (99.3 pCi mL{sup -1}) (Los Alamos townsite). Overall, the CEDE-based on the average concentration of all radionuclides measured over the years-from consuming 5 kg (11 lbs) of honey collected from hives located within the townsites of Los Alamos and White Rock/Pajarito Acres, after regional (background) as been subtracted, was 0.074 {mu}Sv y{sup -1} (0.0074 mrem y{sup -1}) and 0.024 pSv y{sup -1} (0.0024 mrem y{sup -1}), respectively. The highest CEDE, based on the mean + 2 standard deviations (95% confidence level), was 0.334 fiSv y{sup -1} (0.0334 mrem y{sup -1}) (Los Alamos townsitc).« less

  1. Environmental Survey preliminary report, Los Alamos National Laboratory, Los Alamos, New Mexico

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Not Available

    1988-01-01

    This report presents the preliminary findings from the first phase of the Environmental Survey of the United States Department of Energy's (DOE) Los Alamos National Laboratory (LANL), conducted March 29, 1987 through April 17, 1987. The Survey is being conducted by an interdisciplinary team of environmental specialists, led and managed by the Office of Environment, Safety and Health's Office of Environmental Audit. Individual team components are outside experts being supplied by a private contractor. The objective of the Survey is to identify environmental problems and areas of environmental risk associated with the LANL. The Survey covers all environmental media andmore » all areas of environmental regulation. It is being performed in accordance with the DOE Environmental Survey Manual. The on-site phase of the Survey involves the review of existing site environmental data, observations of the operations carried on at the LANL, and interviews with site personnel. The Survey team developed Sampling and Analysis Plan to assist in further assessing certain of the environmental problems identified during its on-site activities. The Sampling and Analysis Plan will be executed by the Idaho National Engineering Laboratory. When completed, the results will be incorporated into the LANL Environmental Survey Interim Report. The Interim Report will reflect the final determinations of the Survey for the LANL. 65 refs., 68 figs., 73 tabs.« less

  2. Stormwater Pollution Prevention Plan for the TA-60-02 Salvage Warehouse, Los Alamos National Laboratory, Revision 3, January 2018

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Burgin, Jillian Elizabeth

    This Storm Water Pollution Prevention Plan (SWPPP) was developed in accordance with the provisions of the Clean Water Act (33 U.S.C. §§1251 et seq., as amended), and the Multi-Sector General Permit for Storm Water Discharges Associated with Industrial Activity (U.S. EPA, June 2015) issued by the U.S. Environmental Protection Agency (EPA) for the National Pollutant Discharge Elimination System (NPDES) and using the industry specific permit requirements for Sector P-Land Transportation and Warehousing as a guide. The applicable stormwater discharge permit is EPA General Permit Registration Number NMR053915 (Los Alamos National Security (LANS) (U.S. EPA, June 2015). Contents of the Junemore » 4, 2015 Multi-sector General Permit can be viewed at: https://www.epa.gov/sites/production/files/2015- 10/documents/msgp2015_finalpermit.pdf This SWPPP applies to discharges of stormwater from the operational areas of the TA-60-02 Salvage and Warehouse facility at Los Alamos National Laboratory. Los Alamos National Laboratory (also referred to as LANL or the “Laboratory”) is owned by the Department of Energy (DOE), and is operated by Los Alamos National Security, LLC (LANS). Throughout this document, the term “facility” refers to the TA-60-02 Salvage/ Warehouse and associated areas. The current permit expires at midnight on June 4, 2020. A copy of the facility NOI and LANS Delegation of Authority Letter are located in Appendix C of this SWPPP.« less

  3. Environmental surveillance at Los Alamos during 2009

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fuehne, David; Poff, Ben; Hjeresen, Denny

    2010-09-30

    Environmental Surveillance at Los Alamos reports are prepared annually by the Los Alamos National Laboratory (the Laboratory) environmental organization, as required by US Department of Energy Order 5400.1, General Environmental Protection Program, and US Department of Energy Order 231.1A, Environment, Safety, and Health Reporting. These annual reports summarize environmental data that are used to determine compliance with applicable federal, state, and local environmental laws and regulations, executive orders, and departmental policies. Additional data, beyond the minimum required, are also gathered and reported as part of the Laboratory’s efforts to ensure public safety and to monitor environmental quality at and nearmore » the Laboratory. Chapter 1 provides an overview of the Laboratory’s major environmental programs and explains the risks and the actions taken to reduce risks at the Laboratory from environmental legacies and waste management operations. Chapter 2 reports the Laboratory’s compliance status for 2009. Chapter 3 provides a summary of the maximum radiological dose the public and biota populations could have potentially received from Laboratory operations and discusses chemical exposures. The environmental surveillance and monitoring data are organized by environmental media (air in Chapter 4; water and sediments in Chapters 5 and 6; soils in Chapter 7; and foodstuffs and biota in Chapter 8) in a format to meet the needs of a general and scientific audience. Chapter 9 provides a summary of the status of environmental restoration work around LANL. The new Chapter 10 describes the Laboratory’s environmental stewardship efforts and provides an overview of the health of the Rio Grande. A glossary and a list of acronyms and abbreviations are in the back of the report. Appendix A explains the standards for environmental contaminants, Appendix B explains the units of measurements used in this report, Appendix C describes the Laboratory

  4. Environmental surveillance at Los Alamos during 1992

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kohen, K.; Stoker, A.; Stone, G.

    1994-07-01

    This report describes the environmental surveillance program at Los Alamos National Laboratory during 1992. The Laboratory routinely monitors for radiation and for radioactive and nonradioactive materials at (or on) Laboratory sites as well as in the surrounding region. LANL uses the monitoring results to determine compliance with appropriate standards and to identify potentially undesirable trends. Data were collected in 1992 to assess external penetrating radiation; quantities of airborne emissions and liquid effluents; concentrations of chemicals and radionuclides in ambient air, surface waters and groundwaters, municipal water supply, soils and sediments, and foodstuffs; and environmental compliance. Using comparisons with standards, regulations,more » and background levels, this report concludes that environmental effects from Laboratory operations are small and do not pose a demonstrable threat to the public, laboratory employees, or the environment.« less

  5. Calculated criticality for sup 235 U/graphite systems using the VIM Monte Carlo code

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Collins, P.J.; Grasseschi, G.L.; Olsen, D.N.

    1992-01-01

    Calculations for highly enriched uranium and graphite systems gained renewed interest recently for the new production modular high-temperature gas-cooled reactor (MHTGR). Experiments to validate the physics calculations for these systems are being prepared for the Transient Reactor Test Facility (TREAT) reactor at Argonne National Laboratory (ANL-West) and in the Compact Nuclear Power Source facility at Los Alamos National Laboratory. The continuous-energy Monte Carlo code VIM, or equivalently the MCNP code, can utilize fully detailed models of the MHTGR and serve as benchmarks for the approximate multigroup methods necessary in full reactor calculations. Validation of these codes and their associated nuclearmore » data did not exist for highly enriched {sup 235}U/graphite systems. Experimental data, used in development of more approximate methods, dates back to the 1960s. The authors have selected two independent sets of experiments for calculation with the VIM code. The carbon-to-uranium (C/U) ratios encompass the range of 2,000, representative of the new production MHTGR, to the ratio of 10,000 in the fuel of TREAT. Calculations used the ENDF/B-V data.« less

  6. A Handbook for Derivative Classifiers at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sinkula, Barbara Jean

    The Los Alamos Classification Office (within the SAFE-IP group) prepared this handbook as a resource for the Laboratory’s derivative classifiers (DCs). It contains information about United States Government (USG) classification policy, principles, and authorities as they relate to the LANL Classification Program in general, and to the LANL DC program specifically. At a working level, DCs review Laboratory documents and material that are subject to classification review requirements, while the Classification Office provides the training and resources for DCs to perform that vital function.

  7. Extraordinary Diversity of Immune Response Proteins among Sea Urchins: Nickel-Isolated Sp185/333 Proteins Show Broad Variations in Size and Charge

    PubMed Central

    Sherman, Lauren S.; Schrankel, Catherine S.; Brown, Kristy J.; Smith, L. Courtney

    2015-01-01

    Effective protection against pathogens requires the host to produce a wide range of immune effector proteins. The Sp185/333 gene family, which is expressed by the California purple sea urchin Strongylocentrotus purpuratus in response to bacterial infection, encodes a highly diverse repertoire of anti-pathogen proteins. A subset of these proteins can be isolated by affinity to metal ions based on multiple histidines, resulting in one to four bands of unique molecular weight on standard Western blots, which vary depending on the individual sea urchin. Two dimensional gel electrophoresis (2DE) of nickel-isolated protein samples followed by Western blot was employed to detect nickel-isolated Sp185/333 (Ni-Sp185/333) proteins and to evaluate protein diversity in animals before and after immune challenge with marine bacteria. Ni-Sp185/333 proteins of the same molecular weight on standard Western blots appear as a broad complex of variants that differ in pI on 2DE Western blots. The Ni-Sp185/333 protein repertoire is variable among animals, and shows a variety of changes among individual sea urchins in response to immune challenges with both the same and different species of bacteria. The extraordinary diversity of the Ni-Sp185/333 proteins may provide significant anti-pathogen capabilities for sea urchins that survive solely on innate immunity. PMID:26406912

  8. Destructive analysis capabilities for plutonium and uranium characterization at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tandon, Lav; Kuhn, Kevin J; Drake, Lawrence R

    Los Alamos National Laboratory's (LANL) Actinide Analytical Chemistry (AAC) group has been in existence since the Manhattan Project. It maintains a complete set of analytical capabilities for performing complete characterization (elemental assay, isotopic, metallic and non metallic trace impurities) of uranium and plutonium samples in different forms. For a majority of the customers there are strong quality assurance (QA) and quality control (QC) objectives including highest accuracy and precision with well defined uncertainties associated with the analytical results. Los Alamos participates in various international and national programs such as the Plutonium Metal Exchange Program, New Brunswick Laboratory's (NBL' s) Safeguardsmore » Measurement Evaluation Program (SME) and several other inter-laboratory round robin exercises to monitor and evaluate the data quality generated by AAC. These programs also provide independent verification of analytical measurement capabilities, and allow any technical problems with analytical measurements to be identified and corrected. This presentation will focus on key analytical capabilities for destructive analysis in AAC and also comparative data between LANL and peer groups for Pu assay and isotopic analysis.« less

  9. Investigation of Radiation Protection Methodologies for Radiation Therapy Shielding Using Monte Carlo Simulation and Measurement

    NASA Astrophysics Data System (ADS)

    Tanny, Sean

    The advent of high-energy linear accelerators for dedicated medical use in the 1950's by Henry Kaplan and the Stanford University physics department began a revolution in radiation oncology. Today, linear accelerators are the standard of care for modern radiation therapy and can generate high-energy beams that can produce tens of Gy per minute at isocenter. This creates a need for a large amount of shielding material to properly protect members of the public and hospital staff. Standardized vault designs and guidance on shielding properties of various materials are provided by the National Council on Radiation Protection (NCRP) Report 151. However, physicists are seeking ways to minimize the footprint and volume of shielding material needed which leads to the use of non-standard vault configurations and less-studied materials, such as high-density concrete. The University of Toledo Dana Cancer Center has utilized both of these methods to minimize the cost and spatial footprint of the requisite radiation shielding. To ensure a safe work environment, computer simulations were performed to verify the attenuation properties and shielding workloads produced by a variety of situations where standard recommendations and guidance documents were insufficient. This project studies two areas of concern that are not addressed by NCRP 151, the radiation shielding workload for the vault door with a non-standard design, and the attenuation properties of high-density concrete for both photon and neutron radiation. Simulations have been performed using a Monte-Carlo code produced by the Los Alamos National Lab (LANL), Monte Carlo Neutrons, Photons 5 (MCNP5). Measurements have been performed using a shielding test port designed into the maze of the Varian Edge treatment vault.

  10. A history of the working group to address Los Alamos community health concerns - A case study of community involvement and risk communication

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Harry Otway; Jon Johnson

    2000-01-01

    In May 1991, at a Department of Energy (DOE) public hearing at Los Alamos, New Mexico, a local artist claimed there had been a recent brain tumor cluster in a small Los Alamos neighborhood. He suggested the cause was radiation from past operations of Los Alamos National Laboratory. Data from the Laboratory's extensive environmental monitoring program gave no reason to believe this charge to be true but also could not prove it false. These allegations, reported in the local and regional media, alarmed the community and revealed an unsuspected lack of trust in the Laboratory. Having no immediate and definitivemore » response, the Laboratory offered to collaborate with the community to address this concern. The Los Alamos community accepted this offer and a joint Community-Laboratory Working Group met for the first time 29 days later. The working group set as its primary goal the search for possible carcinogens in the local environment. Meanwhile, the DOE announced its intention to fund the New Mexico Department of Health to perform a separate and independent epidemiological study of all Los Alamos cancer rates. In early 1994, after commissioning 17 environmental studies and meeting 34 times, the working group decided that the public health concerns had been resolved to the satisfaction of the community and voted to disband. This paper tells the story of the artist and the working group, and how the media covered their story. It summarizes the environmental studies directed by the working group and briefly reviews the main findings of the epidemiology study. An epilogue records the present-day recollections of some of the key players in this environmental drama.« less

  11. Bradbury science museum: your window to Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Deck, Linda Theresa

    The Bradbury Science Museum is the public's window to Los Alamos National Laboratory and supports the Community Program Office's mission to develop community support to accomplish LANL's national security and science mission. It does this by stimulating interest in and increasing basic knowledge of science and technology in northern New Mexico audiences, and increasing public understanding and appreciation of how LANL science and technology solve our global problems. In performing these prime functions, the Museum also preserves the history of scientific accomplishment at the Lab by collecting and preserving artifacts of scientific and historical importance.

  12. Environmental surveillance at Los Alamos during 1987

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Not Available

    1988-05-01

    This report describes the environmental surveillance program conducted by Los Alamos National Laboratory during 1987. Routine monitoring for radiation and radioactive or chemical materials is conducted on the Laboratory site as well as in the surrounding region. Monitoring results are used to determine compliance with appropriate standards and to permit early identification of potentially undesirable trends. Results and interpretation of data for 1987 cover: external penetrating radiation; quantities of airborne emissions and liquid effluents; concentrations of chemicals and radionuclides in ambient air, surface and ground waters, municipal water supply, soils and sediments, and foodstuffs; and environmental compliance. Comparisons with appropriatemore » standards, regulations, and background levels provide the basis for concluding that environmental effects from Laboratory operations are insignificant and do not pose a threat to the public, Laboratory employees, or the environment. 113 refs., 33 figs., 120 tabs.« less

  13. Environmental surveillance at Los Alamos during 1995

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    NONE

    1996-10-01

    This report describes the environmental surveillance program at Los Alamos National Laboratory (LANL or the Laboratory) during 1995. The Laboratory routinely monitors for radiation and for radioactive and nonradioactive materials at (or on) Laboratory sites as well as in the surrounding region. LANL uses the monitoring result to determine compliance with appropriate standards and to identify potentially undesirable trends. Data were collected in 1995 to assess external penetrating radiation; quantities of airborne emissions and liquid effluents; concentrations of chemicals and radionuclides in ambient air, surface waters and groundwaters, municipal water supply, soils and sediments, and foodstuffs; and environmental compliance. Usingmore » comparisons with standards, regulations, and background levels, this report concludes that environmental effects from Laboratory operations are small and do not pose a demonstrable threat to the public, Laboratory employees, or the environment.« less

  14. Los Alamos Plutonium Facility Waste Management System

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Smith, K.; Montoya, A.; Wieneke, R.

    1997-02-01

    This paper describes the new computer-based transuranic (TRU) Waste Management System (WMS) being implemented at the Plutonium Facility at Los Alamos National Laboratory (LANL). The Waste Management System is a distributed computer processing system stored in a Sybase database and accessed by a graphical user interface (GUI) written in Omnis7. It resides on the local area network at the Plutonium Facility and is accessible by authorized TRU waste originators, count room personnel, radiation protection technicians (RPTs), quality assurance personnel, and waste management personnel for data input and verification. Future goals include bringing outside groups like the LANL Waste Management Facilitymore » on-line to participate in this streamlined system. The WMS is changing the TRU paper trail into a computer trail, saving time and eliminating errors and inconsistencies in the process.« less

  15. The Fermiac or Fermi's Trolley

    NASA Astrophysics Data System (ADS)

    Coccetti, F.

    2016-03-01

    The Fermiac, known also as Fermi's trolley or Monte Carlo trolley, is an analog computer used to determine the change in time of the neutron population in a nuclear device, via the Monte Carlo method. It was invented by Enrico Fermi and constructed by Percy King at Los Alamos in 1947, and used for about two years. A replica of the Fermiac was built at INFN mechanical workshops of Bologna in 2015, on behalf of the Museo Storico della Fisica e Centro Studi e Ricerche "Enrico Fermi", thanks to the original drawings made available by Los Alamos National Laboratory (LANL). This reproduction of the Fermiac was put in use, and a simulation was developed.

  16. Fixed-node quantum Monte Carlo

    NASA Astrophysics Data System (ADS)

    Anderson, James B.

    Quantum Monte Carlo methods cannot at present provide exact solutions of the Schrödinger equation for systems with more than a few electrons. But, quantum Monte Carlo calculations can provide very low energy, highly accurate solutions for many systems ranging up to several hundred electrons. These systems include atoms such as Be and Fe, molecules such as H2O, CH4, and HF, and condensed materials such as solid N2 and solid silicon. The quantum Monte Carlo predictions of their energies and structures may not be `exact', but they are the best available. Most of the Monte Carlo calculations for these systems have been carried out using approximately correct fixed nodal hypersurfaces and they have come to be known as `fixed-node quantum Monte Carlo' calculations. In this paper we review these `fixed node' calculations and the accuracies they yield.

  17. 2013 Los Alamos National Laboratory Hazardous Waste Minimization Report

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Salzman, Sonja L.; English, Charles J.

    2015-08-24

    Waste minimization and pollution prevention are inherent goals within the operating procedures of Los Alamos National Security, LLC (LANS). The US Department of Energy (DOE) and LANS are required to submit an annual hazardous waste minimization report to the New Mexico Environment Department (NMED) in accordance with the Los Alamos National Laboratory (LANL or the Laboratory) Hazardous Waste Facility Permit. The report was prepared pursuant to the requirements of Section 2.9 of the LANL Hazardous Waste Facility Permit. This report describes the hazardous waste minimization program (a component of the overall Waste Minimization/Pollution Prevention [WMin/PP] Program) administered by the Environmentalmore » Stewardship Group (ENV-ES). This report also supports the waste minimization and pollution prevention goals of the Environmental Programs Directorate (EP) organizations that are responsible for implementing remediation activities and describes its programs to incorporate waste reduction practices into remediation activities and procedures. LANS was very successful in fiscal year (FY) 2013 (October 1-September 30) in WMin/PP efforts. Staff funded four projects specifically related to reduction of waste with hazardous constituents, and LANS won four national awards for pollution prevention efforts from the National Nuclear Security Administration (NNSA). In FY13, there was no hazardous, mixedtransuranic (MTRU), or mixed low-level (MLLW) remediation waste generated at the Laboratory. More hazardous waste, MTRU waste, and MLLW was generated in FY13 than in FY12, and the majority of the increase was related to MTRU processing or lab cleanouts. These accomplishments and analysis of the waste streams are discussed in much more detail within this report.« less

  18. Efficacy of Alamo for prophylactic and therapeutic treatment of oak wilt in red oaks, 2004

    Treesearch

    K. Ward; J. Juzwik; S. Bernick

    2004-01-01

    An experiment (prophylactic study) to determine the efficacy of Alamo in preventing spread of C. fagacearum through grafted roots of oak wilt-affected and of apparently healthy red oaks was initiated in eight locations in east-central and southeastern Minnesota in Jul 2002.

  19. Environmental assessment for effluent reduction, Los Alamos National Laboratory, Los Alamos, New Mexico

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    NONE

    1996-09-11

    The Department of Energy (DOE) proposes to eliminate industrial effluent from 27 outfalls at Los Alamos National Laboratory (LANL). The Proposed Action includes both simple and extensive plumbing modifications, which would result in the elimination of industrial effluent being released to the environment through 27 outfalls. The industrial effluent currently going to about half of the 27 outfalls under consideration would be rerouted to LANL`s sanitary sewer system. Industrial effluent from other outfalls would be eliminated by replacing once-through cooling water systems with recirculation systems, or, in a few instances, operational changes would result in no generation of industrial effluent.more » After the industrial effluents have been discontinued, the affected outfalls would be removed from the NPDES Permit. The pipes from the source building or structure to the discharge point for the outfalls may be plugged, or excavated and removed. Other outfalls would remain intact and would continue to discharge stormwater. The No Action alternative, which would maintain the status quo for LANL`s outfalls, was also analyzed. An alternative in which industrial effluent would be treated at the source facilities was considered but dismissed from further analysis because it would not reasonably meet the DOE`s purpose for action, and its potential environmental effects were bounded by the analysis of the Proposed Action and the No Action alternatives.« less

  20. Ecological baseline studies in Los Alamos and Guaje Canyons County of Los Alamos, New Mexico. A two-year study

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Foxx, T.S.

    1995-11-01

    During the summers of 1993 and 1994, the Biological Resource Evaluations Team (BRET) of the Environmental Protection Group (ESH-8) conducted baseline studies within two canyon systems, Los Alamos and Guaje Canyons. Biological data was collected within each canyon to provide background and baseline information for Ecological Risk models. Baseline studies included establishment of permanent vegetation plots within each canyon along the elevational gradient. Then, in association with the various vegetation types, surveys were conducted for ground dwelling insects, birds, and small mammals. The stream channels associated with the permanent vegetation plots were characterized and aquatic macroinvertebrates collected within the streammore » monthly throughout a six-month period. The Geographic Position System (GPS) in combination with ARC INFO was used to map the study areas. Considerable data was collected during these surveys and are summarized in individual chapters.« less

  1. Los Alamos Science, Number 25 -- 1997: Celebrating the Neutrino

    DOE R&D Accomplishments Database

    Cooper, N. G. ed.

    1997-01-01

    This issue is devoted to the neutrino and its remaining mysteries. It is divided into the following areas: (1) The Reines-Cowan experiment -- detecting the poltergeist; (2) The oscillating neutrino -- an introduction to neutrino masses and mixing; (3) A brief history of neutrino experiments at LAMPF; (4) A thousand eyes -- the story of LSND (Los Alamos neutrino oscillation experiment); (5) The evidence for oscillations; (6) The nature of neutrinos in muon decay and physics beyond the Standard Model; (7) Exorcising ghosts -- in pursuit of the missing solar neutrinos; (8) MSW -- a possible solution to the solar neutrino problem; (8) Neutrinos and supernovae; and (9) Dark matter and massive neutrinos.

  2. Los Alamos Guns Take Aim at Material's Mysteries

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Byers, Mark; Moore, David; Dimarino, Steve

    Los Alamos National Laboratory scientists and technicians conduct thousands of experiments a year, delving into the fundamental nature of everything from supernovas to subatomic particles. One set of instruments used to better understand the fundamental nature of various materials are 10 scientific gun systems that fire various projectiles at high-tech targets to create enormous velocities, pressures, and temperatures - and using laser, x-ray, and other diagnostics - explore the very nature of metals and other materials. The hundreds of gun-based experiments conducted every year at the Laboratory require a highly-skilled staff of scientists and technicians, and has given rise tomore » a special organization called the "gun working group" to foster open communications, cooperation, problem-solving, and a healthy safety culture.« less

  3. Smoking patterns among Los Alamos National Laboratory employees

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mahoney, M.C.; Wilkinson, G.S.

    Smoking patterns among 5507 employees at Los Alamos National Laboratory were investigated for those who underwent physical examinations by occupational physicians from 1978 to 1983. More male than female employees smoked, although differences in smoking rates between the sexes were not as large as differences observed for national smoking rates. Employees over 40 were more likely to smoke than younger employees, males consumed more cigarettes than did females, and Anglo employees smoked more cigarettes than did Hispanic employees. Highly educated employees smoked less than did less-educated workers, and staff members exhibited the lowest rates of smoking. Smoking cessation programs formore » Laboratory employees should be directed toward those subpopulations with the highest rates of smoking. 31 refs., 8 figs., 1 tab.« less

  4. Los Alamos Guns Take Aim at Material's Mysteries

    ScienceCinema

    Byers, Mark; Moore, David; Dimarino, Steve

    2018-05-30

    Los Alamos National Laboratory scientists and technicians conduct thousands of experiments a year, delving into the fundamental nature of everything from supernovas to subatomic particles. One set of instruments used to better understand the fundamental nature of various materials are 10 scientific gun systems that fire various projectiles at high-tech targets to create enormous velocities, pressures, and temperatures - and using laser, x-ray, and other diagnostics - explore the very nature of metals and other materials. The hundreds of gun-based experiments conducted every year at the Laboratory require a highly-skilled staff of scientists and technicians, and has given rise to a special organization called the "gun working group" to foster open communications, cooperation, problem-solving, and a healthy safety culture.

  5. 77 FR 3257 - Transfer of Land Tracts Located at Los Alamos National Laboratory, New Mexico

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-01-23

    ... DEPARTMENT OF ENERGY Transfer of Land Tracts Located at Los Alamos National Laboratory, New Mexico AGENCY: National Nuclear Security Administration, U.S. Department of Energy. ACTION: Amended Record of Decision. SUMMARY: The U.S. Department of Energy's National Nuclear Security Administration (DOE/NNSA) is...

  6. Leading Change: A Case Study of Alamo Academies--An Industry-Driven Workforce Partnership Program

    ERIC Educational Resources Information Center

    Hu, Xiaodan; Bowman, Gene

    2016-01-01

    In this study, the authors focus on the initiation and development of the Alamo Academies, aiming to illustrate an exemplary industry-driven model that addresses workforce development in local community. After a brief introduction of the context, the authors summarized major factors that contribute to the success of the collaboration model,…

  7. Stormwater Pollution Prevention Plan for the TA-03-38 Metals Fabrication Shop, Los Alamos National Laboratory, Revision 3, January 2018

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Burgin, Jillian Elizabeth

    This Storm Water Pollution Prevention Plan (SWPPP) was developed in accordance with the provisions of the Clean Water Act (33 U.S.C. §§1251 et seq., as amended), and the Multi-Sector General Permit for Storm Water Discharges Associated with Industrial Activity (U.S. EPA, June 2015) issued by the U.S. Environmental Protection Agency (EPA) for the National Pollutant Discharge Elimination System (NPDES) and using the industry specific permit requirements for Sector AA-Fabricated Metal Products as a guide. This SWPPP applies to discharges of stormwater from the operational areas of the TA-03-38 Metals Fabrication Shop at Los Alamos National Laboratory. Los Alamos National Laboratorymore » (also referred to as LANL or the “Laboratory”) is owned by the Department of Energy (DOE), and is operated by Los Alamos National Security, LLC (LANS). Throughout this document, the term “facility” refers to the TA-03-38 Metals Fabrication Shop and associated areas. The current permit expires at midnight on June 4, 2020.« less

  8. Stormwater Pollution Prevention Plan for the TA-60-01 Heavy Equipment Shop, Los Alamos National Laboratory, Revision 3, January 2018

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Burgin, Jillian Elizabeth

    This Storm Water Pollution Prevention Plan (SWPPP) was developed in accordance with the provisions of the Clean Water Act (33 U.S.C. §§1251 et seq., as amended), and the Multi-Sector General Permit for Storm Water Discharges Associated with Industrial Activity (U.S. EPA, June 2015) issued by the U.S. Environmental Protection Agency (EPA) for the National Pollutant Discharge Elimination System (NPDES) and using the industry specific permit requirements for Sector P-Land Transportation and Warehousing as a guide. This SWPPP applies to discharges of stormwater from the operational areas of the TA-60-01 Heavy Equipment Shop at Los Alamos National Laboratory. Los Alamos Nationalmore » Laboratory (also referred to as LANL or the “Laboratory”) is owned by the Department of Energy (DOE), and is operated by Los Alamos National Security, LLC (LANS). Throughout this document, the term “facility” refers to the TA-60-01 Heavy Equipment Shop and associated areas. The current permit expires at midnight on June 4, 2020.« less

  9. Environmental Surveillance at Los Alamos during 2007

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    None

    Environmental Surveillance at Los Alamos reports are prepared annually by the Los Alamos National Laboratory (the Laboratory) Environmental Directorate, as required by US Department of Energy Order 450.1, General Environmental Protection Program, and US Department of Energy Order 231.1A, Environment, Safety, and Health Reporting. These annual reports summarize environmental data that are used to determine compliance with applicable federal, state, and local environmental laws and regulations, executive orders, and departmental policies. Additional data, beyond the minimum required, are also gathered and reported as part of the Laboratory’s efforts to ensure public safety and to monitor environmental quality at and nearmore » the Laboratory. Chapter 1 provides an overview of the Laboratory’s major environmental programs and explains the risks and the actions taken to reduce risks at the Laboratory from environmental legacies and waste management operations. Chapter 2 reports the Laboratory’s compliance status for 2007. Chapter 3 provides a summary of the maximum radiological dose the public and biota populations could have potentially received from Laboratory operations and discusses chemical exposures. The environmental surveillance and monitoring data are organized by environmental media (Chapter 4, air; Chapters 5 and 6, water and sediments; Chapter 7, soils; and Chapter 8, foodstuffs and biota) in a format to meet the needs of a general and scientific audience. Chapter 9 provides a summary of the status of environmental restoration work around LANL. A glossary and a list of acronyms and abbreviations are in the back of the report. Appendix A explains the standards for environmental contaminants, Appendix B explains the units of measurements used in this report, Appendix C describes the laboratory’s technical areas and their associated programs, and Appendix D provides web links to more information. In printed copies of this report or Executive Summary

  10. The "El Alamo" project (1990-1997): two consecutive hospital-based studies of breast cancer outcomes in Spain.

    PubMed

    Martín, M; Mahillo, E; Llombart-Cussac, A; Lluch, A; Munarriz, B; Pastor, M; Alba, E; Ruiz, A; Antón, A; Bermejo, B

    2006-07-01

    The "Alamo" project is a retrospective analysis of 14,854 patients diagnosed of breast cancer between 1990 and 1997 in 50 Spanish hospitals. Alamo I (AI) consisted of 4,532 patients diagnosed with breast cancer between 1990 and 1993. Data were collected in 2000. Alamo II (AII) consisted of 10,322 patients diagnosed between 1994 and 1997. Data were collected in 2003. At presentation, there were (AI vs. AII) 17.6% vs. 24.3% at stage I; 55.5% vs. 53.1% at stage II; 18.7% vs. 15% at stage III; 7.2% vs. 5.9 at stage IV. Median age was 57 (AI) vs. 58 years (AII) and 65.9% vs. 67.2% (AI vs. AII) were post-menopausal. Firstline treatment for disease stages I, II and III was surgery in 91% of patients in both studies. Breast conserving surgery rate increased from 20.2% (AI) to 32.7% (AII). Adjuvant systemic treatments were administered to 87.6% (AI) and 92.8% (AII) of patients. Recurrence rate diminished from 36.6% (AI) to 22.5% (AII) and the 9-year survival rate increased from 63.2% (95% CI: 61.5-64.9) to 70.1% (95% CI: 68.5-71.8). Breast cancer outcomes in Spain have improved from 1990-1993 to 1994-1997, likely because of breast cancer screening program implementation and new therapies.

  11. Erosion and Deposition Monitoring Using High-Density Aerial Lidar and Geomorphic Change Detection Software Analysis at Los Alamos National Laboratory, Los Alamos New Mexico, LA-UR-17-26743

    NASA Astrophysics Data System (ADS)

    Walker, T.; Kostrubala, T. L.; Muggleton, S. R.; Veenis, S.; Reid, K. D.; White, A. B.

    2017-12-01

    The Los Alamos National Laboratory storm water program installed sediment transport mitigation structures to reduce the migration of contaminants within the Los Alamos and Pueblo (LA/P) watershed in Los Alamos, NM. The goals of these structures are to minimize storm water runoff and erosion, enhance deposition, and reduce mobility of contaminated sediments. Previous geomorphological monitoring used GPS surveyed cross-sections on a reach scale to interpolate annual geomorphic change in sediment volumes. While monitoring has confirmed the LA/P watershed structures are performing as designed, the cross-section method proved difficult to estimate uncertainty and the coverage area was limited. A new method, using the Geomorphic Change Detection (GCD) plugin for ESRI ArcGIS developed by Wheaton et al. (2010), with high-density aerial lidar data, has been used to provide high confidence uncertainty estimates and greater areal coverage. Following the 2014 monsoon season, airborne lidar data has been collected annually and the resulting DEMs processed using the GCD method. Additionally, a more accurate characterization of low-amplitude geomorphic changes, typical of low-flow/low-rainfall monsoon years, has been documented by applying a spatially variable error to volume change calculations using the GCD based fuzzy inference system (FIS). The FIS method allows for the calculation of uncertainty based on data set quality and density e.g. point cloud density, ground slope, and degree of surface roughness. At the 95% confidence level, propagated uncertainty estimates of the 2015 and 2016 lidar DEM comparisons yielded detectable changes greater than 0.3 m - 0.46 m. Geomorphic processes identified and verified in the field are typified by low-amplitude, within-channel aggradation and incision and out of channel bank collapse that over the course of a monsoon season result in localized and dectetable change. While the resulting reach scale volume change from 2015 - 2016 was often

  12. Alamos: An International Collaboration to Provide a Space Based Environmental Monitoring Solution for the Deep Space Network

    NASA Astrophysics Data System (ADS)

    Kennedy, S. O.; Dunn, A.; Lecomte, J.; Buchheim, K.; Johansson, E.; Berger, T.

    2018-02-01

    This abstract proposes the advantages of an externally mounted instrument in support of the human physiology, space biology, and human health and performance key science area. Alamos provides Space-Based Environmental Monitoring capabilities.

  13. Pinon Pine Tree Study, Los Alamos National Laboratory: Source document

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    P. R. Fresquez; J. D. Huchton; M. A. Mullen

    One of the dominant tree species growing within and around Los Alamos National Laboratory (LANL), Los Alamos, NM, lands is the pinon pine (Pinus edulis) tree. Pinon pine is used for firewood, fence posts, and building materials and is a source of nuts for food--the seeds are consumed by a wide variety of animals and are also gathered by people in the area and eaten raw or roasted. This study investigated the (1) concentration of {sup 3}H, {sup 137}Cs, {sup 90}Sr, {sup tot}U, {sup 238}Pu, {sup 239,240}Pu, and {sup 241}Am in soils (0- to 12-in. [31 cm] depth underneath themore » tree), pinon pine shoots (PPS), and pinon pine nuts (PPN) collected from LANL lands and regional background (BG) locations, (2) concentrations of radionuclides in PPN collected in 1977 to present data, (3) committed effective dose equivalent (CEDE) from the ingestion of nuts, and (4) soil to PPS to PPN concentration ratios (CRs). Most radionuclides, with the exception of {sup 3}H in soils, were not significantly higher (p < 0.10) in soils, PPS, and PPN collected from LANL as compared to BG locations, and concentrations of most radionuclides in PPN from LANL have decreased over time. The maximum net CEDE (the CEDE plus two sigma minus BG) at the most conservative ingestion rate (10 lb [4.5 kg]) was 0.0018 mrem (0.018 {micro}Sv). Soil-to-nut CRs for most radionuclides were within the range of default values in the literature for common fruits and vegetables.« less

  14. Feral Cattle in the White Rock Canyon Reserve at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hathcock, Charles D.; Hansen, Leslie A.

    2014-03-27

    At the request of the Los Alamos Field Office (the Field Office), Los Alamos National Security (LANS) biologists placed remote-triggered wildlife cameras in and around the mouth of Ancho Canyon in the White Rock Canyon Reserve (the Reserve) to monitor use by feral cattle. The cameras were placed in October 2012 and retrieved in January 2013. Two cameras were placed upstream in Ancho Canyon away from the Rio Grande along the perennial flows from Ancho Springs, two cameras were placed at the north side of the mouth to Ancho Canyon along the Rio Grande, and two cameras were placed atmore » the south side of the mouth to Ancho Canyon along the Rio Grande. The cameras recorded three different individual feral cows using this area as well as a variety of local native wildlife. This report details our results and issues associated with feral cattle in the Reserve. Feral cattle pose significant risks to human safety, impact cultural and biological resources, and affect the environmental integrity of the Reserve. Regional stakeholders have communicated to the Field Office that they support feral cattle removal.« less

  15. Encoded physics knowledge in checking codes for nuclear cross section libraries at Los Alamos

    NASA Astrophysics Data System (ADS)

    Parsons, D. Kent

    2017-09-01

    Checking procedures for processed nuclear data at Los Alamos are described. Both continuous energy and multi-group nuclear data are verified by locally developed checking codes which use basic physics knowledge and common-sense rules. A list of nuclear data problems which have been identified with help of these checking codes is also given.

  16. Los Alamos National Security, LLC Request for Information from industrial entities that desire to commercialize Laboratory-developed Extremely Low Resource Optical Identifier (ELROI) tech

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Erickson, Michael Charles

    Los Alamos National Security, LLC (LANS) is the manager and operator of the Los Alamos National Laboratory for the U.S. Department of Energy National Nuclear Security Administration under contract DE-AC52-06NA25396. LANS is a mission-centric Federally Funded Research and Development Center focused on solving the most critical national security challenges through science and engineering for both government and private customers.

  17. Critical assembly: A technical history of Los Alamos during the Oppenheimer years, 1943--1945

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hoddeson, L.; Henriksen, P.W.; Meade, R.A.

    1993-11-01

    This volume treats the technical research that led to the first atomic bombs. The authors explore how the ``critical assembly`` of scientists, engineers, and military Personnel at Los Alamos collaborated during World War II, blending their traditions to create a new approach to large-scale research. The research was characterized by strong mission orientation, multidisciplinary teamwork, expansion of the scientists` traditional methodology with engineering techniques, and a trail-and-error methodology responding to wartime deadlines. The book opens with an introduction laying out major themes. After a synopsis of the prehistory of the bomb project, from the discovery of nuclear fission to themore » start of the Manhattan Engineer District, and an overview of the early materials program, the book examines the establishment of the Los Alamos Laboratory, the implosion and gun assembly programs, nuclear physics research, chemistry and metallurgy, explosives, uranium and plutonium development, confirmation of spontaneous fission in pile-produced plutonium, the thermonuclear bomb, critical assemblies, the Trinity test, and delivery of the combat weapons.« less

  18. Supplement Analysis for the Site-Wide Environmental Impact Statement for Continued Operation of Los Alamos National Laboratory -- Recovery and Storage of Strontium-90 Fueled Radioisotope Thermal Electric Generators at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    N /A

    2004-01-22

    This Supplement Analysis (SA) has been prepared to determine if the Site-Wide Environmental Impact Statement for Continued Operations of Los Alamos National Laboratory (SWEIS) (DOE/EIS-0238) adequately addresses the environmental effects of recovery and storage for disposal of six strontium-90 (Sr-90) fueled radioisotope thermal electric generators (RTGs) at the Los Alamos National Laboratory (LANL) Technical Area (TA)-54, Area G, or if the SWEIS needs to be supplemented. DOE's National Nuclear Security Administration (NNSA) proposed to recover and store six Sr-90 RTGs from the commercial sector as part of its Offsite-Source Recovery Project (OSRP). The OSRP focuses on the proactive recovery andmore » storage of unwanted radioactive sealed sources exceeding the US Nuclear Regulatory Commission (NRC) limits for Class C low-level waste (also known as Greater than Class C waste, or GTCC). In response to the events of September 11, 2001, NRC conducted a risk-based evaluation of potential vulnerabilities to terrorist threats involving NRC-licensed nuclear facilities and materials. NRC's evaluation concluded that possession of unwanted radioactive sealed sources with no disposal outlet presents a potential vulnerability (NRC 2002). In a November 25, 2003 letter to the manager of the NNSA's Los Alamos Site Office, the NRC Office of Nuclear Security and Incident Response identified recovery of several Sr-90 RTGs as the highest priority and requested that DOE take whatever actions necessary to recovery these sources as soon as possible. This SA specifically compares key impact assessment parameters of this proposal to the offsite source recovery program evaluated in the SWEIS and a subsequent SA that evaluated a change to the approach of a portion of the recovery program. It also provides an explanation of any differences between the Proposed Action and activities described in the previous SWEIS and SA analyses.« less

  19. Post-Cold War Science and Technology at Los Alamos

    NASA Astrophysics Data System (ADS)

    Browne, John C.

    2002-04-01

    Los Alamos National Laboratory serves the nation through the development and application of leading-edge science and technology in support of national security. Our mission supports national security by: ensuring the safety, security, and reliability of the U.S. nuclear stockpile; reducing the threat of weapons of mass destruction in support of counter terrorism and homeland defense; and solving national energy, environment, infrastructure, and health security problems. We require crosscutting fundamental and advanced science and technology research to accomplish our mission. The Stockpile Stewardship Program develops and applies, advanced experimental science, computational simulation, and technology to ensure the safety and reliability of U.S. nuclear weapons in the absence of nuclear testing. This effort in itself is a grand challenge. However, the terrorist attack of September 11, 2001, reminded us of the importance of robust and vibrant research and development capabilities to meet new and evolving threats to our national security. Today through rapid prototyping we are applying new, innovative, science and technology for homeland defense, to address the threats of nuclear, chemical, and biological weapons globally. Synergistically, with the capabilities that we require for our core mission, we contribute in many other areas of scientific endeavor. For example, our Laboratory has been part of the NASA effort on mapping water on the moon and NSF/DOE projects studying high-energy astrophysical phenomena, understanding fundamental scaling phenomena of life, exploring high-temperature superconductors, investigating quantum information systems, applying neutrons to condensed-matter and nuclear physics research, developing large-scale modeling and simulations to understand complex phenomena, and exploring nanoscience that bridges the atomic to macroscopic scales. In this presentation, I will highlight some of these post-cold war science and technology advances

  20. 2016 Los Alamos National Laboratory Hazardous Waste Minimization Report

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Salzman, Sonja L.; English, Charles Joe

    Waste minimization and pollution prevention are goals within the operating procedures of Los Alamos National Security, LLC (LANS). The US Department of Energy (DOE), inclusive of the National Nuclear Security Administration (NNSA) and the Office of Environmental Management, and LANS are required to submit an annual hazardous waste minimization report to the New Mexico Environment Department (NMED) in accordance with the Los Alamos National Laboratory (LANL or the Laboratory) Hazardous Waste Facility Permit. The report was prepared pursuant to the requirements of Section 2.9 of the LANL Hazardous Waste Facility Permit. This report describes the hazardous waste minimization program, whichmore » is a component of the overall Pollution Prevention (P2) Program, administered by the Environmental Stewardship Group (EPC-ES). This report also supports the waste minimization and P2 goals of the Associate Directorate of Environmental Management (ADEM) organizations that are responsible for implementing remediation activities and describes its programs to incorporate waste reduction practices into remediation activities and procedures. This report includes data for all waste shipped offsite from LANL during fiscal year (FY) 2016 (October 1, 2015 – September 30, 2016). LANS was active during FY2016 in waste minimization and P2 efforts. Multiple projects were funded that specifically related to reduction of hazardous waste. In FY2016, there was no hazardous, mixed-transuranic (MTRU), or mixed low-level (MLLW) remediation waste shipped offsite from the Laboratory. More non-remediation hazardous waste and MLLW was shipped offsite from the Laboratory in FY2016 compared to FY2015. Non-remediation MTRU waste was not shipped offsite during FY2016. These accomplishments and analysis of the waste streams are discussed in much more detail within this report.« less

  1. Los Alamos Explosives Performance Key to Stockpile Stewardship

    ScienceCinema

    Dattelbaum, Dana

    2018-02-14

    As the U.S. Nuclear Deterrent ages, one essential factor in making sure that the weapons will continue to perform as designed is understanding the fundamental properties of the high explosives that are part of a nuclear weapons system. As nuclear weapons go through life extension programs, some changes may be advantageous, particularly through the addition of what are known as "insensitive" high explosives that are much less likely to accidentally detonate than the already very safe "conventional" high explosives that are used in most weapons. At Los Alamos National Laboratory explosives research includes a wide variety of both large- and small-scale experiments that include small contained detonations, gas and powder gun firings, larger outdoor detonations, large-scale hydrodynamic tests, and at the Nevada Nuclear Security Site, underground sub-critical experiments.

  2. A review of the Los Alamos effort in the development of nuclear rocket propulsion

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Durham, F.P.; Kirk, W.L.; Bohl, R.J.

    1991-01-01

    This paper reviews the achievements of the Los Alamos nuclear rocket propulsion program and describes some specific reactor design and testing problems encountered during the development program along with the progress made in solving these problems. The relevance of these problems to a renewed nuclear thermal rocket development program for the Space Exploration Initiative (SEI) is discussed. 11 figs.

  3. Los Alamos Neutron Science Center

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kippen, Karen Elizabeth

    For more than 30 years the Los Alamos Neutron Science Center (LANSCE) has provided the scientific underpinnings in nuclear physics and material science needed to ensure the safety and surety of the nuclear stockpile into the future. In addition to national security research, the LANSCE User Facility has a vibrant research program in fundamental science, providing the scientific community with intense sources of neutrons and protons to perform experiments supporting civilian research and the production of medical and research isotopes. Five major experimental facilities operate simultaneously. These facilities contribute to the stockpile stewardship program, produce radionuclides for medical testing, andmore » provide a venue for industrial users to irradiate and test electronics. In addition, they perform fundamental research in nuclear physics, nuclear astrophysics, materials science, and many other areas. The LANSCE User Program plays a key role in training the next generation of top scientists and in attracting the best graduate students, postdoctoral researchers, and early-career scientists. The U.S. Department of Energy (DOE), National Nuclear Security Administration (NNSA) —the principal sponsor of LANSCE—works with the Office of Science and the Office of Nuclear Energy, which have synergistic long-term needs for the linear accelerator and the neutron science that is the heart of LANSCE.« less

  4. Los Alamos Science, Number 25 -- 1997: Celebrating the neutrino

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cooper, N.G.

    1997-12-31

    This issue is devoted to the neutrino and its remaining mysteries. It is divided into the following areas: (1) The Reines-Cowan experiment -- detecting the poltergeist; (2) The oscillating neutrino -- an introduction to neutrino masses and mixing; (3) A brief history of neutrino experiments at LAMPF; (4) A thousand eyes -- the story of LSND (Los Alamos neutrino oscillation experiment); (5) The evidence for oscillations; (6) The nature of neutrinos in muon decay and physics beyond the Standard Model; (7) Exorcising ghosts -- in pursuit of the missing solar neutrinos; (8) MSW -- a possible solution to the solarmore » neutrino problem; (8) Neutrinos and supernovae; and (9) Dark matter and massive neutrinos.« less

  5. (U) Introduction to Monte Carlo Methods

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hungerford, Aimee L.

    2017-03-20

    Monte Carlo methods are very valuable for representing solutions to particle transport problems. Here we describe a “cook book” approach to handling the terms in a transport equation using Monte Carlo methods. Focus is on the mechanics of a numerical Monte Carlo code, rather than the mathematical foundations of the method.

  6. Los Alamos National Laboratory Meteorology Monitoring Program: 2016 Data Completeness/ Quality Report

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bruggeman, David Alan

    This report summarizes data completeness by tower and by instrument for 2016 and compares that data with the Los Alamos National Laboratory (LANL) and American National Standards Institute (ANSI) 2015 standards. This report is designed to make data users aware of data completeness and any data quality issues. LANL meteorology monitoring goals include 95% completeness for all measurements. The ANSI 2015 standard requires 90% completeness for all measurements. This report documents instrument/tower issues as they impact data completeness.

  7. Quantum Gibbs ensemble Monte Carlo

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fantoni, Riccardo, E-mail: rfantoni@ts.infn.it; Moroni, Saverio, E-mail: moroni@democritos.it

    We present a path integral Monte Carlo method which is the full quantum analogue of the Gibbs ensemble Monte Carlo method of Panagiotopoulos to study the gas-liquid coexistence line of a classical fluid. Unlike previous extensions of Gibbs ensemble Monte Carlo to include quantum effects, our scheme is viable even for systems with strong quantum delocalization in the degenerate regime of temperature. This is demonstrated by an illustrative application to the gas-superfluid transition of {sup 4}He in two dimensions.

  8. Saving Water at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Erickson, Andy

    Los Alamos National Laboratory decreased its water usage by 26 percent in 2014, with about one-third of the reduction attributable to using reclaimed water to cool a supercomputing center. The Laboratory's goal during 2014 was to use only re-purposed water to support the mission at the Strategic Computing Complex. Using reclaimed water from the Sanitary Effluent Reclamation Facility, or SERF, substantially decreased water usage and supported the overall mission. SERF collects industrial wastewater and treats it for reuse. The reclamation facility contributed more than 27 million gallons of re-purposed water to the Laboratory's computing center, a secured supercomputing facility thatmore » supports the Laboratory’s national security mission and is one of the institution’s larger water users. In addition to the strategic water reuse program at SERF, the Laboratory reduced water use in 2014 by focusing conservation efforts on areas that use the most water, upgrading to water-conserving fixtures, and repairing leaks identified in a biennial survey.« less

  9. Saving Water at Los Alamos National Laboratory

    ScienceCinema

    Erickson, Andy

    2018-01-16

    Los Alamos National Laboratory decreased its water usage by 26 percent in 2014, with about one-third of the reduction attributable to using reclaimed water to cool a supercomputing center. The Laboratory's goal during 2014 was to use only re-purposed water to support the mission at the Strategic Computing Complex. Using reclaimed water from the Sanitary Effluent Reclamation Facility, or SERF, substantially decreased water usage and supported the overall mission. SERF collects industrial wastewater and treats it for reuse. The reclamation facility contributed more than 27 million gallons of re-purposed water to the Laboratory's computing center, a secured supercomputing facility that supports the Laboratory’s national security mission and is one of the institution’s larger water users. In addition to the strategic water reuse program at SERF, the Laboratory reduced water use in 2014 by focusing conservation efforts on areas that use the most water, upgrading to water-conserving fixtures, and repairing leaks identified in a biennial survey.

  10. IMPACTS OF DRILLING ADDITIVES ON DATA OBTAINED FROM HYDROGEOLOGIC CHARACTERIZATION WELLS AT LOS ALAMOS NATIONAL LABORATORY

    EPA Science Inventory

    Personnel at the EPA Ground Water and Ecosystems Restoration Division (GWERD) were requested by EPA Region 6 to evaluate the impacts of well drilling practices at the Los Alamos National Laboratory (LANL). The focus of this review involved analysis of the impacts of bentonite- a...

  11. A Tracer Test at the Los Alamos Canyon Weir

    NASA Astrophysics Data System (ADS)

    Levitt, D. G.; Stone, W. J.; Newell, D. L.; Wykoff, D. S.

    2002-12-01

    A low-head weir was constructed in the Los Alamos Canyon to reduce the transport of contaminant-bearing sediment caused by fire-enhanced runoff off Los Alamos National Laboratory (LANL) property towards the Rio Grande following the May 2000 Cerro Grande fire at Los Alamos, New Mexico. Fractured basalt was exposed in the channel by grading during construction of the weir, and water temporarily ponds behind the weir following periods of runoff. In order to monitor any downward transport of contaminants into fractured basalt, and potentially downward to the regional ground water, three boreholes (one vertical, one at 43 degrees, and one at 34 degrees from horizontal) were installed for environmental monitoring. The boreholes penetrate to depths ranging from approximately 9 to 82 m below the weir floor. The two angled boreholes are fitted with flexible FLUTe liners with resistance sensors to measure relative moisture content and absorbent sampling pads for contaminant and environmental tracer sampling within the vadose zone. The two angled boreholes are also monitored for relative changes in moisture content by neutron logging. The vertical borehole penetrates three perched water zones and is equipped with four screens and sampling ports. In April 2002, a tracer test was initiated with the application of a 0.2 M (16,000 ppm) solution of potassium bromide (KBr) onto the weir floor. The tracer experiment was intended to provide data on travel times through the complex hydrogeologic media of fractured basalt. A precipitation and runoff event in June 2002 resulted in approximately 0.61 m of standing water behind the weir. If the KBr and flood waters were well mixed, the concentration of KBr in the flood waters was approximately 24 ppm. Bromide was detected in the absorbent membrane in the 43 degree hole at concentrations up to 2 ppm. Resistance sensors in the 43 degree borehole detected moisture increases within 3 days at a depth of 27 m, indicating an average wetting

  12. Los Alamos Explosives Performance Key to Stockpile Stewardship

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Dattelbaum, Dana

    2014-11-03

    As the U.S. Nuclear Deterrent ages, one essential factor in making sure that the weapons will continue to perform as designed is understanding the fundamental properties of the high explosives that are part of a nuclear weapons system. As nuclear weapons go through life extension programs, some changes may be advantageous, particularly through the addition of what are known as "insensitive" high explosives that are much less likely to accidentally detonate than the already very safe "conventional" high explosives that are used in most weapons. At Los Alamos National Laboratory explosives research includes a wide variety of both large- andmore » small-scale experiments that include small contained detonations, gas and powder gun firings, larger outdoor detonations, large-scale hydrodynamic tests, and at the Nevada Nuclear Security Site, underground sub-critical experiments.« less

  13. Pesticide concentrations in water and in suspended and bottom sediments in the New and Alamo rivers, Salton Sea Watershed, California, April 2003

    USGS Publications Warehouse

    LeBlanc, Lawrence A.; Orlando, James L.; Kuivila, Kathryn

    2004-01-01

    This report contains pesticide concentration data for water, and suspended and bed sediment samples collected in April 2003 from twelve sites along the New and Alamo Rivers in the Salton Sea watershed, in southeastern California. The study was done in collaboration with the California State Regional Water Quality Control Board, Colorado River Region, to assess inputs of current-use pesticides associated with water and sediment into the New and Alamo Rivers. Five sites along the New River and seven sites along the Alamo River, downstream of major agricultural drains, were selected and covered the lengths of the rivers from the international boundary to approximately 1.5 km from the river mouths. Sampling from bridges occurred at seven of the twelve sites. At these sites, streamflow measurements were taken. These same sites were also characterized for cross-stream homogeneity by measuring dissolved oxygen, pH, specific conductance, temperature, and suspended solids concentration at several vertical (depths) and horizontal (cross-stream) points across the river. Large volume water samples (200?300 L) were collected for isolation of suspended sediments by flow-through centrifugation. Water from the outflow of the flow-through centrifuge was sampled for the determination of aqueous pesticide concentrations. In addition, bottom sediments were sampled at each site. Current-use pesticides and legacy organochlorine compounds (p,p'-DDT, p,p'-DDE and p,p'-DDD) were extracted from sediments and measured via gas chromatography/mass spectrometry (GC/MS). Organic carbon and percentage of fines were also determined for suspended and bottom sediments. Cross-stream transects of dissolved constituents and suspended sediments showed that the rivers were fairly homogeneous at the sites sampled. Streamflow was higher at the outlet sites, with the Alamo River having higher flow (1,240 cfs) than the New River (798 cfs). Twelve current-use pesticides, one legacy organochlorine compound (p

  14. Los Alamos Discovers Super Efficient Solar Using Perovskite Crystals

    ScienceCinema

    Mohite, Aditya; Nie, Wanyi

    2018-05-11

    State-of-the-art photovoltaics using high-purity, large-area, wafer-scale single-crystalline semiconductors grown by sophisticated, high temperature crystal-growth processes offer promising routes for developing low-cost, solar-based clean global energy solutions for the future. Solar cells composed of the recently discovered material organic-inorganic perovskites offer the efficiency of silicon, yet suffer from a variety of deficiencies limiting the commercial viability of perovskite photovoltaic technology. In research to appear in Science, Los Alamos National Laboratory researchers reveal a new solution-based hot-casting technique that eliminates these limitations, one that allows for the growth of high-quality, large-area, millimeter-scale perovskite crystals and demonstrates that highly efficient and reproducible solar cells with reduced trap assisted recombination can be realized.

  15. 75 FR 24957 - Decision to Evaluate a Petition to Designate a Class of Employees From the Los Alamos National...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-05-06

    ... Laboratory. Location: Los Alamos, New Mexico. Job Titles and/or Job Duties: All employees of the Department.... Hinnefeld, Interim Director, Division of Compensation Analysis and Support, National Institute for...

  16. Electronic Spectra from Molecular Dynamics: A Simple Approach.

    DTIC Science & Technology

    1983-10-01

    82.30.Cr. 33.20K. S2.40.1s The authors provided phototypeset copy for this paper using REFER TlL EON, TOFF On UNIX I ELECTRONIC SPECTRA FROM MOLECULAR...Alamos National Laboratory Los Alamos, NM 87545 I. INTRODUCTION In this paper we show how molecular dynamics can be used in a simple manner to com...could equally use Monte Carlo or explicit integration over coordinates to compute equilibrium electronic absorption bands. How- ever, molecular

  17. Los Alamos Laser Eye Investigation.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Odom, C. R.

    2005-01-01

    A student working in a laser laboratory at Los Alamos National Laboratory sustained a serious retinal injury to her left eye when she attempted to view suspended particles in a partially evacuated target chamber. The principle investigator was using the white light from the flash lamp of a Class 4 Nd:YAG laser to illuminate the particles. Since the Q-switch was thought to be disabled at the time of the accident, the principal investigator assumed it would be safe to view the particles without wearing laser eye protection. The Laboratory Director appointed a team to investigate the accident and to reportmore » back to him the events and conditions leading up to the accident, equipment malfunctions, safety management causal factors, supervisory and management action/inaction, adequacy of institutional processes and procedures, emergency and notification response, effectiveness of corrective actions and lessons learned from previous similar events, and recommendations for human and institutional safety improvements. The team interviewed personnel, reviewed documents, and characterized systems and conditions in the laser laboratory during an intense six week investigation. The team determined that the direct and primary failures leading to this accident were, respectively, the principle investigator's unsafe work practices and the institution's inadequate monitoring of worker performance. This paper describes the details of the investigation, the human and institutional failures, and the recommendations for improving the laser safety program.« less

  18. Structural Geology of the Northwestern Portion of Los Alamos National Laboratory, Rio Grande Rift, New Mexico: Implications for Seismic Surface Rupture Potential from TA-3 to TA-55

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Jamie N. Gardner: Alexis Lavine; Giday WoldeGabriel; Donathon Krier

    1999-03-01

    Los Alamos National Laboratory lies at the western boundary of the Rio Grande rift, a major tectonic feature of the North American Continent. Three major faults locally constitute the modem rift boundary, and each of these is potentially seismogenic. In this study we have gathered structural geologic data for the northwestern portion of Los Alamos National Laboratory through high-precision geologic mapping, conventional geologic mapping, stratigraphic studies, drilling, petrologic studies, and stereographic aerial photograph analyses. Our study area encompasses TA-55 and TA-3, where potential for seismic surface rupture is of interest, and is bounded on the north and south by themore » townsite of Los Alamos and Twomile Canyon, respectively. The study area includes parts of two of the potentially active rift boundary faults--the Pajarito and Rendija Canyon faults-that form a large graben that we name the Diamond Drive graben. The graben embraces the western part of the townsite of Los Alamos, and its southern end is in the TA-3 area where it is defined by east-southeast-trending cross faults. The cross faults are small, but they accommodate interactions between the two major fault zones and gentle tilting of structural blocks to the north into the graben. North of Los Alamos townsite, the Rendija Canyon fault is a large normal fault with about 120 feet of down-to-the-west displacement over the last 1.22 million years. South from Los Alamos townsite, the Rendija Canyon fault splays to the southwest into a broad zone of deformation. The zone of deformation is about 2,000 feet wide where it crosses Los Alamos Canyon and cuts through the Los Alamos County Landfill. Farther southwest, the fault zone is about 3,000 feet wide at the southeastern corner of TA-3 in upper Mortandad Canyon and about 5,000 feet wide in Twomile Canyon. Net down-to-the-west displacement across the entire fault zone over the last 1.22 million years decreases to the south as the fault zone

  19. A Monte Carlo Analysis of Weight Data from UF 6 Cylinder Feed and Withdrawal Stations

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Garner, James R; Whitaker, J Michael

    2015-01-01

    As the number of nuclear facilities handling uranium hexafluoride (UF 6) cylinders (e.g., UF 6 production, enrichment, and fuel fabrication) increase in number and throughput, more automated safeguards measures will likely be needed to enable the International Atomic Energy Agency (IAEA) to achieve its safeguards objectives in a fiscally constrained environment. Monitoring the process data from the load cells built into the cylinder feed and withdrawal (F/W) stations (i.e., cylinder weight data) can significantly increase the IAEA’s ability to efficiently achieve the fundamental safeguards task of confirming operations as declared (i.e., no undeclared activities). Researchers at the Oak Ridge Nationalmore » Laboratory, Los Alamos National Laboratory, the Joint Research Center (in Ispra, Italy), and University of Glasgow are investigating how this weight data can be used for IAEA safeguards purposes while fully protecting the operator’s proprietary and sensitive information related to operations. A key question that must be resolved is, what is the necessary frequency of recording data from the process F/W stations to achieve safeguards objectives? This paper summarizes Monte Carlo simulations of typical feed, product, and tails withdrawal cycles and evaluates longer sampling frequencies to determine the expected errors caused by low-frequency sampling and its impact on material balance calculations.« less

  20. SABRINA: an interactive three-dimensional geometry-mnodeling program for MCNP

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    West, J.T. III

    SABRINA is a fully interactive three-dimensional geometry-modeling program for MCNP, a Los Alamos Monte Carlo code for neutron and photon transport. In SABRINA, a user constructs either body geometry or surface geometry models and debugs spatial descriptions for the resulting objects. This enhanced capability significantly reduces effort in constructing and debugging complicated three-dimensional geometry models for Monte Carlo analysis. 2 refs., 33 figs.

  1. The Los Alamos Neutron Science Center Spallation Neutron Sources

    NASA Astrophysics Data System (ADS)

    Nowicki, Suzanne F.; Wender, Stephen A.; Mocko, Michael

    The Los Alamos Neutron Science Center (LANSCE) provides the scientific community with intense sources of neutrons, which can be used to perform experiments supporting civilian and national security research. These measurements include nuclear physics experiments for the defense program, basic science, and the radiation effect programs. This paper focuses on the radiation effects program, which involves mostly accelerated testing of semiconductor parts. When cosmic rays strike the earth's atmosphere, they cause nuclear reactions with elements in the air and produce a wide range of energetic particles. Because neutrons are uncharged, they can reach aircraft altitudes and sea level. These neutrons are thought to be the most important threat to semiconductor devices and integrated circuits. The best way to determine the failure rate due to these neutrons is to measure the failure rate in a neutron source that has the same spectrum as those produced by cosmic rays. Los Alamos has a high-energy and a low-energy neutron source for semiconductor testing. Both are driven by the 800-MeV proton beam from the LANSCE accelerator. The high-energy neutron source at the Weapons Neutron Research (WNR) facility uses a bare target that is designed to produce fast neutrons with energies from 100 keV to almost 800 MeV. The measured neutron energy distribution from WNR is very similar to that of the cosmic-ray-induced neutrons in the atmosphere. However, the flux provided at the WNR facility is typically 5×107 times more intense than the flux of the cosmic-ray-induced neutrons. This intense neutron flux allows testing at greatly accelerated rates. An irradiation test of less than an hour is equivalent to many years of neutron exposure due to cosmic-ray neutrons. The low-energy neutron source is located at the Lujan Neutron Scattering Center. It is based on a moderated source that provides useful neutrons from subthermal energies to ∼100 keV. The characteristics of these sources, and

  2. The Los Alamos Neutron Science Center Spallation Neutron Sources

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Nowicki, Suzanne F.; Wender, Stephen A.; Mocko, Michael

    The Los Alamos Neutron Science Center (LANSCE) provides the scientific community with intense sources of neutrons, which can be used to perform experiments supporting civilian and national security research. These measurements include nuclear physics experiments for the defense program, basic science, and the radiation effect programs. This paper focuses on the radiation effects program, which involves mostly accelerated testing of semiconductor parts. When cosmic rays strike the earth's atmosphere, they cause nuclear reactions with elements in the air and produce a wide range of energetic particles. Because neutrons are uncharged, they can reach aircraft altitudes and sea level. These neutronsmore » are thought to be the most important threat to semiconductor devices and integrated circuits. The best way to determine the failure rate due to these neutrons is to measure the failure rate in a neutron source that has the same spectrum as those produced by cosmic rays. Los Alamos has a high-energy and a low-energy neutron source for semiconductor testing. Both are driven by the 800-MeV proton beam from the LANSCE accelerator. The high-energy neutron source at the Weapons Neutron Research (WNR) facility uses a bare target that is designed to produce fast neutrons with energies from 100 keV to almost 800 MeV. The measured neutron energy distribution from WNR is very similar to that of the cosmic-ray-induced neutrons in the atmosphere. However, the flux provided at the WNR facility is typically 5×107 times more intense than the flux of the cosmic-ray-induced neutrons. This intense neutron flux allows testing at greatly accelerated rates. An irradiation test of less than an hour is equivalent to many years of neutron exposure due to cosmic-ray neutrons. The low-energy neutron source is located at the Lujan Neutron Scattering Center. It is based on a moderated source that provides useful neutrons from subthermal energies to ~100 keV. The characteristics of these sources

  3. The Los Alamos Neutron Science Center Spallation Neutron Sources

    DOE PAGES

    Nowicki, Suzanne F.; Wender, Stephen A.; Mocko, Michael

    2017-10-26

    The Los Alamos Neutron Science Center (LANSCE) provides the scientific community with intense sources of neutrons, which can be used to perform experiments supporting civilian and national security research. These measurements include nuclear physics experiments for the defense program, basic science, and the radiation effect programs. This paper focuses on the radiation effects program, which involves mostly accelerated testing of semiconductor parts. When cosmic rays strike the earth's atmosphere, they cause nuclear reactions with elements in the air and produce a wide range of energetic particles. Because neutrons are uncharged, they can reach aircraft altitudes and sea level. These neutronsmore » are thought to be the most important threat to semiconductor devices and integrated circuits. The best way to determine the failure rate due to these neutrons is to measure the failure rate in a neutron source that has the same spectrum as those produced by cosmic rays. Los Alamos has a high-energy and a low-energy neutron source for semiconductor testing. Both are driven by the 800-MeV proton beam from the LANSCE accelerator. The high-energy neutron source at the Weapons Neutron Research (WNR) facility uses a bare target that is designed to produce fast neutrons with energies from 100 keV to almost 800 MeV. The measured neutron energy distribution from WNR is very similar to that of the cosmic-ray-induced neutrons in the atmosphere. However, the flux provided at the WNR facility is typically 5×107 times more intense than the flux of the cosmic-ray-induced neutrons. This intense neutron flux allows testing at greatly accelerated rates. An irradiation test of less than an hour is equivalent to many years of neutron exposure due to cosmic-ray neutrons. The low-energy neutron source is located at the Lujan Neutron Scattering Center. It is based on a moderated source that provides useful neutrons from subthermal energies to ~100 keV. The characteristics of these sources

  4. Biological Assessment of the Continued Operation of Los Alamos National Laboratory on Federally Listed Threatened and Endangered Species

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hansen, Leslie A.

    2006-09-19

    This biological assessment considers the effects of continuing to operate Los Alamos National Laboratory on Federally listed threatened or endangered species, based on current and future operations identified in the 2006 Site-wide Environmental Impact Statement for the Continued Operation of Los Alamos National Laboratory (SWEIS; DOE In Prep.). We reviewed 40 projects analyzed in the SWEIS as well as two aspects on ongoing operations to determine if these actions had the potential to affect Federally listed species. Eighteen projects that had not already received U.S. Fish and Wildlife Service (USFWS) consultation and concurrence, as well as the two aspects ofmore » ongoing operations, ecological risk from legacy contaminants and the Outfall Reduction Project, were determined to have the potential to affect threatened or endangered species. Cumulative impacts were also analyzed.« less

  5. Monte Carlo Simulation for Perusal and Practice.

    ERIC Educational Resources Information Center

    Brooks, Gordon P.; Barcikowski, Robert S.; Robey, Randall R.

    The meaningful investigation of many problems in statistics can be solved through Monte Carlo methods. Monte Carlo studies can help solve problems that are mathematically intractable through the analysis of random samples from populations whose characteristics are known to the researcher. Using Monte Carlo simulation, the values of a statistic are…

  6. The Los ALamos Neutron Science Center Hydrogen Moderator System

    NASA Astrophysics Data System (ADS)

    Jarmer, J. J.; Knudson, J. N.

    2006-04-01

    At the Los Alamos Neutron Science Center (LANSCE), spallation neutrons are produced by an 800-MeV proton beam interacting with tungsten targets. Gun-barrel-type penetrations through the heavy concrete and steel shielding that surround the targets collimate neutrons to form neutron beams used for scattering experiments. Two liquid hydrogen moderators of one-liter volume each are positioned adjacent to the neutron-production targets. Some of the neutrons that pass through a moderator interact with or scatter from protons in the hydrogen. The neutron-proton interaction reduces the energy or moderates neutrons to lower energies. Lower energy "moderated" neutrons are the most useful for some neutron scattering experiments. We provide a description of the LANSCE hydrogen-moderator system and its cryogenic performance with proton beams of up to 125 micro-amp average current.

  7. Space Science at Los Alamos National Laboratory

    NASA Astrophysics Data System (ADS)

    Smith, Karl

    2017-09-01

    The Space Science and Applications group (ISR-1) in the Intelligence and Space Research (ISR) division at the Los Alamos National Laboratory lead a number of space science missions for civilian and defense-related programs. In support of these missions the group develops sensors capable of detecting nuclear emissions and measuring radiations in space including γ-ray, X-ray, charged-particle, and neutron detection. The group is involved in a number of stages of the lifetime of these sensors including mission concept and design, simulation and modeling, calibration, and data analysis. These missions support monitoring of the atmosphere and near-Earth space environment for nuclear detonations as well as monitoring of the local space environment including space-weather type events. Expertise in this area has been established over a long history of involvement with cutting-edge projects continuing back to the first space based monitoring mission Project Vela. The group's interests cut across a large range of topics including non-proliferation, space situational awareness, nuclear physics, material science, space physics, astrophysics, and planetary physics.

  8. Threatened and Endangered Species Habitat Management Plan for Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hathcock, Charles Dean; Keller, David Charles; Thompson, Brent E.

    Los Alamos National Laboratory’s (LANL) Threatened and Endangered Species Habitat Management Plan (HMP) fulfills a commitment made to the U.S. Department of Energy (DOE) in the “Final Environmental Impact Statement for the Dual-Axis Radiographic Hydrodynamic Test Facility Mitigation Action Plan” (DOE 1996). The HMP received concurrence from the U.S. Fish and Wildlife Service (USFWS) in 1999 (USFWS consultation numbers 2-22-98-I-336 and 2-22-95-I-108). This 2017 update retains the management guidelines from the 1999 HMP for listed species, and updates some descriptive information.

  9. New facility for ion beam materials characterization and modification at Los Alamos

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tesmer, J.R.; Maggiore, C.J.; Parkin, D.M.

    1988-01-01

    The Ion Beam Materials Laboratory (IBML) is a new Los Alamos laboratory devoted to the characterization and modification of the near surfaces of materials. The primary instruments of the IBML are a tandem electrostatic accelerator, a National Electrostatics Corp. Model 9SDH, coupled with a Varian CF-3000 ion implanter. The unique organizational structure of the IBML as well as the operational characteristics of the 9SDH (after approximately 3000 h of operation) and the laboratories' research capabilities will be discussed. Examples of current research results will also be presented. 5 refs., 2 figs.

  10. An experimental topographic amplification study at Los Alamos National Laboratory using ambient vibrations

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Stolte, Andrew C.; Cox, Brady R.; Lee, Richard C.

    An experimental study aimed at investigating potential topographic amplification of seismic waves was conducted on a 50-m-tall and 185-m-wide soft-rock ridge located at Los Alamos National Laboratory near Los Alamos, New Mexico. Ten portable broadband seismograph stations were placed in arrays across the ridge and left to record ambient vibration data for ~9 hours. Clear evidence of topographic amplification was observed by comparing spectral ratios calculated from ambient noise recordings at the toe, slope, and crest of the instrumented ridge. The inferred resonance frequency of the ridge obtained from the experimental recordings was found to agree well with several simplemore » estimates of the theoretical resonance frequency based on its geometry and stiffness. Results support the feasibility of quantifying the frequency range of topographic amplification solely using ambient vibrations, rather than strong or weak ground motions. Additionally, comparisons have been made between a number of widely used experimental methods for quantifying topographic effects, such as the standard spectral ratio, median reference method, and horizontal-to-vertical spectral ratio. As a result, differences in the amplification and frequency range of topographic effects indicated by these methods highlight the importance of choosing a reference condition that is appropriate for the site-specific conditions and goals associated with an experimental topographic amplification study.« less

  11. An experimental topographic amplification study at Los Alamos National Laboratory using ambient vibrations

    DOE PAGES

    Stolte, Andrew C.; Cox, Brady R.; Lee, Richard C.

    2017-03-14

    An experimental study aimed at investigating potential topographic amplification of seismic waves was conducted on a 50-m-tall and 185-m-wide soft-rock ridge located at Los Alamos National Laboratory near Los Alamos, New Mexico. Ten portable broadband seismograph stations were placed in arrays across the ridge and left to record ambient vibration data for ~9 hours. Clear evidence of topographic amplification was observed by comparing spectral ratios calculated from ambient noise recordings at the toe, slope, and crest of the instrumented ridge. The inferred resonance frequency of the ridge obtained from the experimental recordings was found to agree well with several simplemore » estimates of the theoretical resonance frequency based on its geometry and stiffness. Results support the feasibility of quantifying the frequency range of topographic amplification solely using ambient vibrations, rather than strong or weak ground motions. Additionally, comparisons have been made between a number of widely used experimental methods for quantifying topographic effects, such as the standard spectral ratio, median reference method, and horizontal-to-vertical spectral ratio. As a result, differences in the amplification and frequency range of topographic effects indicated by these methods highlight the importance of choosing a reference condition that is appropriate for the site-specific conditions and goals associated with an experimental topographic amplification study.« less

  12. 76 FR 62330 - Radio Broadcasting Services; Alamo, GA; Alton, MO; Boscobel, WI; Buffalo, OK; Cove, AR; Clayton...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-10-07

    ... Broadcasting Services; Alamo, GA; Alton, MO; Boscobel, WI; Buffalo, OK; Cove, AR; Clayton, LA; Daisy, AR; Ennis... competitive bidding process, and are considered unsold permits that were included in Auction 91. Interested... competitive bidding process. DATES: Comments must be filed on or before October 31, 2011, and reply comments...

  13. Los Alamos County Fire Department LAFD: TA-55 PF-4 Facility Familiarization Tour, OJT 55260

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Rutherford, Victor Stephen

    Los Alamos National Laboratory (LANL) will conduct familiarization tours for Los Alamos County Fire Department (LAFD) personnel at the Plutonium Facility (PF-4) at Technical Area (TA)-55. These familiarization tours are official LANL business; the purpose of these tours is to orient the firefighters to the facility so that they can respond efficiently and quickly to a variety of emergency situations. This orientation includes the ingress and egress of the area and buildings, layout and organization of the facility, evacuation procedures and assembly points, and areas of concern within the various buildings at the facility. LAFD firefighters have the skills andmore » abilities to perform firefighting operations and other emergency response tasks that cannot be provided by other LANL personnel who have the required clearance level. This handout provides details of the information, along with maps and diagrams, to be presented during the familiarization tours. The handout will be distributed to the trainees at the time of the tour. A corresponding checklist will also be used as guidance during the familiarization tours to ensure that all required information is presented to LAFD personnel.« less

  14. Statistical analyses of the background distribution of groundwater solutes, Los Alamos National Laboratory, New Mexico.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Longmire, Patrick A.; Goff, Fraser; Counce, D. A.

    2004-01-01

    Background or baseline water chemistry data and information are required to distingu ish between contaminated and non-contaminated waters for environmental investigations conducted at Los Alamos National Laboratory (referred to as the Laboratory). The term 'background' refers to natural waters discharged by springs or penetrated by wells that have not been contaminated by LANL or other municipal or industrial activities, and that are representative of groundwater discharging from their respective aquifer material. These investigations are conducted as part of the Environmental Restoration (ER) Project, Groundwater Protection Program (GWPP), Laboratory Surveillance Program, the Hydrogeologic Workplan, and the Site-Wide Environmental Impact Statement (SWEIS).more » This poster provides a comprehensive, validated database of inorganic, organic, stable isotope, and radionuclide analyses of up to 136 groundwater samples collected from 15 baseline springs and wells located in and around Los Alamos National Laboratory, New Mexico. The region considered in this investigation extends from the western edge of the Jemez Mountains eastward to the Rio Grande and from Frijoles Canyon northward to Garcia Canyon. Figure 1 shows the fifteen stations sampled for this investigation. The sampling stations and associated aquifer types are summarized in Table 1.« less

  15. Self-learning Monte Carlo method

    DOE PAGES

    Liu, Junwei; Qi, Yang; Meng, Zi Yang; ...

    2017-01-04

    Monte Carlo simulation is an unbiased numerical tool for studying classical and quantum many-body systems. One of its bottlenecks is the lack of a general and efficient update algorithm for large size systems close to the phase transition, for which local updates perform badly. In this Rapid Communication, we propose a general-purpose Monte Carlo method, dubbed self-learning Monte Carlo (SLMC), in which an efficient update algorithm is first learned from the training data generated in trial simulations and then used to speed up the actual simulation. Lastly, we demonstrate the efficiency of SLMC in a spin model at the phasemore » transition point, achieving a 10–20 times speedup.« less

  16. Climate Change and the Los Alamos National Laboratory. The Adaptation Challenge

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fowler, Kimberly M.; Hjeresen, Dennis; Silverman, Josh

    2015-02-01

    The Los Alamos National Laboratory (LANL) has been adapting to climate change related impacts that have been occurring on decadal time scales. The region where LANL is located has been subject to a cascade of climate related impacts: drought, devastating wildfires, and historic flooding events. Instead of buckling under the pressure, LANL and the surrounding communities have integrated climate change mitigation strategies into their daily operations and long-term plans by increasing coordination and communication between the Federal, State, and local agencies in the region, identifying and aggressively managing forested areas in need of near-term attention, addressing flood control and retentionmore » issues, and more.« less

  17. Bayesian statistics and Monte Carlo methods

    NASA Astrophysics Data System (ADS)

    Koch, K. R.

    2018-03-01

    The Bayesian approach allows an intuitive way to derive the methods of statistics. Probability is defined as a measure of the plausibility of statements or propositions. Three rules are sufficient to obtain the laws of probability. If the statements refer to the numerical values of variables, the so-called random variables, univariate and multivariate distributions follow. They lead to the point estimation by which unknown quantities, i.e. unknown parameters, are computed from measurements. The unknown parameters are random variables, they are fixed quantities in traditional statistics which is not founded on Bayes' theorem. Bayesian statistics therefore recommends itself for Monte Carlo methods, which generate random variates from given distributions. Monte Carlo methods, of course, can also be applied in traditional statistics. The unknown parameters, are introduced as functions of the measurements, and the Monte Carlo methods give the covariance matrix and the expectation of these functions. A confidence region is derived where the unknown parameters are situated with a given probability. Following a method of traditional statistics, hypotheses are tested by determining whether a value for an unknown parameter lies inside or outside the confidence region. The error propagation of a random vector by the Monte Carlo methods is presented as an application. If the random vector results from a nonlinearly transformed vector, its covariance matrix and its expectation follow from the Monte Carlo estimate. This saves a considerable amount of derivatives to be computed, and errors of the linearization are avoided. The Monte Carlo method is therefore efficient. If the functions of the measurements are given by a sum of two or more random vectors with different multivariate distributions, the resulting distribution is generally not known. TheMonte Carlo methods are then needed to obtain the covariance matrix and the expectation of the sum.

  18. Neutrino oscillation parameter sampling with MonteCUBES

    NASA Astrophysics Data System (ADS)

    Blennow, Mattias; Fernandez-Martinez, Enrique

    2010-01-01

    We present MonteCUBES ("Monte Carlo Utility Based Experiment Simulator"), a software package designed to sample the neutrino oscillation parameter space through Markov Chain Monte Carlo algorithms. MonteCUBES makes use of the GLoBES software so that the existing experiment definitions for GLoBES, describing long baseline and reactor experiments, can be used with MonteCUBES. MonteCUBES consists of two main parts: The first is a C library, written as a plug-in for GLoBES, implementing the Markov Chain Monte Carlo algorithm to sample the parameter space. The second part is a user-friendly graphical Matlab interface to easily read, analyze, plot and export the results of the parameter space sampling. Program summaryProgram title: MonteCUBES (Monte Carlo Utility Based Experiment Simulator) Catalogue identifier: AEFJ_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEFJ_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: GNU General Public Licence No. of lines in distributed program, including test data, etc.: 69 634 No. of bytes in distributed program, including test data, etc.: 3 980 776 Distribution format: tar.gz Programming language: C Computer: MonteCUBES builds and installs on 32 bit and 64 bit Linux systems where GLoBES is installed Operating system: 32 bit and 64 bit Linux RAM: Typically a few MBs Classification: 11.1 External routines: GLoBES [1,2] and routines/libraries used by GLoBES Subprograms used:Cat Id ADZI_v1_0, Title GLoBES, Reference CPC 177 (2007) 439 Nature of problem: Since neutrino masses do not appear in the standard model of particle physics, many models of neutrino masses also induce other types of new physics, which could affect the outcome of neutrino oscillation experiments. In general, these new physics imply high-dimensional parameter spaces that are difficult to explore using classical methods such as multi-dimensional projections and minimizations, such as those

  19. Los Alamos nEDM Experiment and Demonstration of Ramsey's Method on Stored UCNs at the LANL UCN Source

    NASA Astrophysics Data System (ADS)

    Clayton, Steven; Chupp, Tim; Cude-Woods, Christopher; Currie, Scott; Ito, Takeyasu; Liu, Chen-Yu; Long, Joshua; MacDonald, Stephen; Makela, Mark; O'Shaughnessy, Christopher; Plaster, Brad; Ramsey, John; Saunders, Andy; LANL nEDM Collaboration

    2017-09-01

    The Los Alamos National Laboratory ultracold neutron (UCN) source was recently upgraded for a factor of 5 improvement in stored density, providing the statistical precision needed for a room temperature neutron electric dipole moment measurement with sensitivity 3 ×10-27 e . cm, a factor 10 better than the limit set by the Sussex-RAL-ILL experiment. Here, we show results of a demonstration of Ramsey's separated oscillatory fields method on stored UCNs at the LANL UCN source and in a geometry relevant for a nEDM measurement. We argue a world-leading nEDM experiment could be performed at LANL with existing technology and a short lead time, providing a physics result with sensitivity intermediate between the current limit set by Sussex-RAL-ILL, and the anticipated limit from the complex, cryogenic nEDM experiment planned for the next decade at the ORNL Spallation Neutron Source (SNS-nEDM). This work was supported by the Los Alamos LDRD Program, Project 20140015DR.

  20. Availability of environmental radioactivity to honey bee colonies at Los Alamos

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hakonson, T.E.; Bostick, K.V.

    Data are presented on the availability of tritium, cesium 137, and plutonium to honey bee colonies foraging in the environment surrounding the Los Alamos Scientific Laboratory. Sources of these radionuclides in the laboratory environs include liquid and atmospheric effluents and buried solid waste. Honey bee colonies were placed in three canyon liquid waste disposal areas and were sampled frequently, along with honey, surface water, and surrounding vegetation, to qualitatively determine the availability of these radionuclides to bees (Apis mellifera) and to identify potential food chain sources of the elements. Tritium concentrations in bee and honey samples from the canyons increasedmore » rapidly from initial values of <1 pCi/ml moisture to as much as 9.2 nCi/ml in 75 days after placement of the hives in the canyons. Seasonal patterns in foraging activities as influenced by weather and food availability were apparent in the data. It appears that several sources of tritium were utilized by the colonies, including surface water in the canyons and vegetation receiving tritium from atmospheric effluents and buried solid waste. Concentrations of cesium 137 and plutonium were generally low or undetectable in bees throughout the study. However, levels of both nuclides increased by factors of 10 to 20 in bees from two of the canyon study areas during a 3-month period in 1973. It was speculated that the liquid effluents in the two canyons were the source of the increased concentrations in bee samples, since this water was the only significant source of /sup 137/Cs in the environs. The existence of at least three radionuclide sources in the Los Alamos Scientific Laboratory (LASL) environs complicates the interpretation of the data. However, it is apparent that honey bees can acquire /sup 3/H, /sup 137/Cs, and Pu from multiple sources in the environs.« less

  1. Self-Learning Monte Carlo Method

    NASA Astrophysics Data System (ADS)

    Liu, Junwei; Qi, Yang; Meng, Zi Yang; Fu, Liang

    Monte Carlo simulation is an unbiased numerical tool for studying classical and quantum many-body systems. One of its bottlenecks is the lack of general and efficient update algorithm for large size systems close to phase transition or with strong frustrations, for which local updates perform badly. In this work, we propose a new general-purpose Monte Carlo method, dubbed self-learning Monte Carlo (SLMC), in which an efficient update algorithm is first learned from the training data generated in trial simulations and then used to speed up the actual simulation. We demonstrate the efficiency of SLMC in a spin model at the phase transition point, achieving a 10-20 times speedup. This work is supported by the DOE Office of Basic Energy Sciences, Division of Materials Sciences and Engineering under Award DE-SC0010526.

  2. 2015 Los Alamos Space Weather Summer School Research Reports

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cowee, Misa; Chen, Yuxi; Desai, Ravindra

    The fifth Los Alamos Space Weather Summer School was held June 1st - July 24th, 2015, at Los Alamos National Laboratory (LANL). With renewed support from the Institute of Geophysics, Planetary Physics, and Signatures (IGPPS) and additional support from the National Aeronautics and Space Administration (NASA) and the Department of Energy (DOE) Office of Science, we hosted a new class of five students from various U.S. and foreign research institutions. The summer school curriculum includes a series of structured lectures as well as mentored research and practicum opportunities. Lecture topics including general and specialized topics in the field of spacemore » weather were given by a number of researchers affiliated with LANL. Students were given the opportunity to engage in research projects through a mentored practicum experience. Each student works with one or more LANL-affiliated mentors to execute a collaborative research project, typically linked with a larger ongoing research effort at LANL and/or the student’s PhD thesis research. This model provides a valuable learning experience for the student while developing the opportunity for future collaboration. This report includes a summary of the research efforts fostered and facilitated by the Space Weather Summer School. These reports should be viewed as work-in-progress as the short session typically only offers sufficient time for preliminary results. At the close of the summer school session, students present a summary of their research efforts. Titles of the papers included in this report are as follows: Full particle-in-cell (PIC) simulation of whistler wave generation, Hybrid simulations of the right-hand ion cyclotron anisotropy instability in a sub-Alfvénic plasma flow, A statistical ensemble for solar wind measurements, Observations and models of substorm injection dispersion patterns, Heavy ion effects on Kelvin-Helmholtz instability: hybrid study, Simulating plasmaspheric electron densities with

  3. Evaluation of aircraft crash hazard at Los Alamos National Laboratory facilities

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Selvage, R.D.

    This report selects a method for use in calculating the frequency of an aircraft crash occurring at selected facilities at the Los Alamos National Laboratory (the Laboratory). The Solomon method was chosen to determine these probabilities. Each variable in the Solomon method is defined and a value for each variable is selected for fourteen facilities at the Laboratory. These values and calculated probabilities are to be used in all safety analysis reports and hazards analyses for the facilities addressed in this report. This report also gives detailed directions to perform aircraft-crash frequency calculations for other facilities. This will ensure thatmore » future aircraft-crash frequency calculations are consistent with calculations in this report.« less

  4. Recent advances and future prospects for Monte Carlo

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Brown, Forrest B

    2010-01-01

    The history of Monte Carlo methods is closely linked to that of computers: The first known Monte Carlo program was written in 1947 for the ENIAC; a pre-release of the first Fortran compiler was used for Monte Carlo In 1957; Monte Carlo codes were adapted to vector computers in the 1980s, clusters and parallel computers in the 1990s, and teraflop systems in the 2000s. Recent advances include hierarchical parallelism, combining threaded calculations on multicore processors with message-passing among different nodes. With the advances In computmg, Monte Carlo codes have evolved with new capabilities and new ways of use. Production codesmore » such as MCNP, MVP, MONK, TRIPOLI and SCALE are now 20-30 years old (or more) and are very rich in advanced featUres. The former 'method of last resort' has now become the first choice for many applications. Calculations are now routinely performed on office computers, not just on supercomputers. Current research and development efforts are investigating the use of Monte Carlo methods on FPGAs. GPUs, and many-core processors. Other far-reaching research is exploring ways to adapt Monte Carlo methods to future exaflop systems that may have 1M or more concurrent computational processes.« less

  5. Environmental surveillance and compliance at Los Alamos during 1996

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    NONE

    1997-09-01

    This report presents environmental data that characterize environmental performance and addresses compliance with environmental standards and requirements at Los Alamos National Laboratory (LANL or the Laboratory) during 1996. The Laboratory routinely monitors for radiation and for radioactive nonradioactive materials at Laboratory sites as well as in the surrounding region. LANL uses the monitoring results to determine compliance with appropriate standards and to identify potentially undesirable trends. Data were collected in 1996 to assess external penetrating radiation; quantities of airborne emissions; and concentrations of chemicals and radionuclides in ambient air, surface waters and groundwaters, the municipal water supply, soils and sediments,more » and foodstuffs. Using comparisons with standards and regulations, this report concludes that environmental effects from Laboratory operations are small and do not pose a demonstrable threat to the public, Laboratory employees, or the environment. Laboratory operations were in compliance with all major environmental regulations.« less

  6. Monte Carlo Transport for Electron Thermal Transport

    NASA Astrophysics Data System (ADS)

    Chenhall, Jeffrey; Cao, Duc; Moses, Gregory

    2015-11-01

    The iSNB (implicit Schurtz Nicolai Busquet multigroup electron thermal transport method of Cao et al. is adapted into a Monte Carlo transport method in order to better model the effects of non-local behavior. The end goal is a hybrid transport-diffusion method that combines Monte Carlo Transport with a discrete diffusion Monte Carlo (DDMC). The hybrid method will combine the efficiency of a diffusion method in short mean free path regions with the accuracy of a transport method in long mean free path regions. The Monte Carlo nature of the approach allows the algorithm to be massively parallelized. Work to date on the method will be presented. This work was supported by Sandia National Laboratory - Albuquerque and the University of Rochester Laboratory for Laser Energetics.

  7. Los Alamos National Laboratory Science Education Programs. Progress report, October 1, 1994--December 31, 1994

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gill, D.H.

    During the 1994 summer institute NTEP teachers worked in coordination with LANL and the Los Alamos Middle School and Mountain Elementary School to gain experience in communicating on-line, to gain further information from the Internet and in using electronic Bulletin Board Systems (BBSs) to exchange ideas with other teachers. To build on their telecommunications skills, NTEP teachers participated in the International Telecommunications In Education Conference (Tel*ED `94) at the Albuquerque Convention Center on November 11 & 12, 1994. They attended the multimedia keynote address, various workshops highlighting many aspects of educational telecommunications skills, and the Telecomm Rodeo sponsored by Losmore » Alamos National Laboratory. The Rodeo featured many presentations by Laboratory personnel and educational institutions on ways in which telecommunications technologies can be use din the classroom. Many were of the `hands-on` type, so that teachers were able to try out methods and equipment and evaluate their usefulness in their own schools and classrooms. Some of the presentations featured were the Geonet educational BBS system, the Supercomputing Challenge, and the Sunrise Project, all sponsored by LANL; the `CU-seeMe` live video software, various simulation software packages, networking help, and many other interesting and useful exhibits.« less

  8. Beam Loss Measurements at the Los Alamos Proton Storage Ring

    NASA Astrophysics Data System (ADS)

    Spickermann, Thomas

    2005-06-01

    During normal operation the Los Alamos Proton Storage Ring (PSR) accumulates up to 4ṡ1013 protons over 625μs with a repetition rate of 20 Hz, corresponding to a current of 125μA to the Lujan Neutron Science Center. Beam losses in the ring as well as in the extraction beam line and the subsequent activation of material are a limiting factor at these currents. Careful tuning of injection, ring and extraction line is paramount to limiting losses to acceptable levels. Losses are typically not uniform around the ring, but occur in significantly higher levels in certain "hot spots". Here I will report on losses related to the stripper foil which are the dominant source of losses in the ring. First results of a comparison with simulations will also be presented.

  9. Neutron Capture Experiments Using the DANCE Array at Los Alamos

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Dashdorj, D.; MonAme Scientific Research Center, Ulaanbaatar; Mitchell, G. E.

    2009-03-31

    The Detector for Advanced Neutron Capture Experiments (DANCE) is designed for neutron capture measurements on very small and/or radioactive targets. The DANCE array of 160 BaF{sub 2} scintillation detectors is located at the Lujan Center at the Los Alamos Neutron Science Center (LANSCE). Accurate measurements of neutron capture data are important for many current applications as well as for basic understanding of neutron capture. The gamma rays following neutron capture reactions have been studied by the time-of-flight technique using the DANCE array. The high granularity of the array allows measurements of the gamma-ray multiplicity. The gamma-ray multiplicities and energy spectramore » for different multiplicities can be measured and analyzed for spin and parity determination of the resolved resonances.« less

  10. The Los Alamos Scientific Laboratory - An Isolated Nuclear Research Establishment

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bradbury, Norris E.; Meade, Roger Allen

    Early in his twenty-five year career as the Director of the Los Alamos Scientific Laboratory, Norris Bradbury wrote at length about the atomic bomb and the many implications the bomb might have on the world. His themes were both technical and philosophical. In 1963, after nearly twenty years of leading the nation’s first nuclear weapons laboratory, Bradbury took the opportunity to broaden his writing. In a paper delivered to the International Atomic Energy Agency’s symposium on the “Criteria in the Selection of Sites for the Construction of Reactors and Nuclear Research Centers,” Bradbury took the opportunity to talk about themore » business of nuclear research and the human component of operating a scientific laboratory. This report is the transcript of his talk.« less

  11. Common ground: An environmental ethic for Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Menlove, F.L.

    1991-01-01

    Three predominant philosophies have characterized American business ethical thinking over the past several decades. The first phase is the ethics of self-interest'' which argues that maximizing self-interest coincidentally maximizes the common good. The second phase is legality ethics.'' Proponents argue that what is important is knowing the rules and following them scrupulously. The third phase might be called stake-holder ethics.'' A central tenant is that everyone affected by a decision has a moral hold on the decision maker. This paper will discuss one recent initiative of the Los Alamos National Laboratory to move beyond rules and regulations toward an environmentalmore » ethic that integrates the values of stakeholder ethics'' into the Laboratory's historical culture and value systems. These Common Ground Principles are described. 11 refs.« less

  12. Response Matrix Monte Carlo for electron transport

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ballinger, C.T.; Nielsen, D.E. Jr.; Rathkopf, J.A.

    1990-11-01

    A Response Matrix Monte Carol (RMMC) method has been developed for solving electron transport problems. This method was born of the need to have a reliable, computationally efficient transport method for low energy electrons (below a few hundred keV) in all materials. Today, condensed history methods are used which reduce the computation time by modeling the combined effect of many collisions but fail at low energy because of the assumptions required to characterize the electron scattering. Analog Monte Carlo simulations are prohibitively expensive since electrons undergo coulombic scattering with little state change after a collision. The RMMC method attempts tomore » combine the accuracy of an analog Monte Carlo simulation with the speed of the condensed history methods. The combined effect of many collisions is modeled, like condensed history, except it is precalculated via an analog Monte Carol simulation. This avoids the scattering kernel assumptions associated with condensed history methods. Results show good agreement between the RMMC method and analog Monte Carlo. 11 refs., 7 figs., 1 tabs.« less

  13. Deterministic theory of Monte Carlo variance

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ueki, T.; Larsen, E.W.

    1996-12-31

    The theoretical estimation of variance in Monte Carlo transport simulations, particularly those using variance reduction techniques, is a substantially unsolved problem. In this paper, the authors describe a theory that predicts the variance in a variance reduction method proposed by Dwivedi. Dwivedi`s method combines the exponential transform with angular biasing. The key element of this theory is a new modified transport problem, containing the Monte Carlo weight w as an extra independent variable, which simulates Dwivedi`s Monte Carlo scheme. The (deterministic) solution of this modified transport problem yields an expression for the variance. The authors give computational results that validatemore » this theory.« less

  14. Los Alamos Shows Airport Security Technology at Work

    ScienceCinema

    Espy, Michelle; Schultz, Larry; Hunter, James

    2018-05-30

    Los Alamos scientists have advanced a Magnetic Resonance Imaging (MRI) technology that may provide a breakthrough for screening liquids at airport security. They've added low-power X-ray data to the mix, and as a result have unlocked a new detection technology. Funded in part by the Department of Homeland Security's Science and Technology Directorate, the new system is named MagRay. The goal is to quickly and accurately distinguish between liquids that visually appear identical. For example, what appears to be a bottle of white wine could potentially be nitromethane, a liquid that could be used to make an explosive. Both are clear liquids, one would be perfectly safe on a commercial aircraft, the other would be strictly prohibited. How to tell them apart quickly without error at an airport security area is the focus of Michelle Espy, Larry Schultz and their team. In this video, Espy and the MagRay team explain how the new technology works, how they've developed an easy operator interface, and what the next steps might be in transitioning this technology to the private sector.

  15. World's Largest Gold Crystal Studied at Los Alamos

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Vogel, Sven; Nakotte, Heinz

    2014-04-03

    When geologist John Rakovan needed better tools to investigate whether a dazzling 217.78-gram piece of gold was in fact the world's largest single-crystal specimen - a distinguishing factor that would not only drastically increase its market value but also provide a unique research opportunity - he traveled to Los Alamos National Laboratory's Lujan Neutron Scattering Center to peer deep inside the mineral using neutron diffractometry. Neutrons, different from other probes such as X-rays and electrons, are able to penetrate many centimeters deep into most materials. Revealing the inner structure of a crystal without destroying the sample - imperative, as thismore » one is worth an estimated $1.5 million - would allow Rakovan and Lujan Center collaborators Sven Vogel and Heinz Nakotte to prove that this exquisite nugget, which seemed almost too perfect and too big to be real, was a single crystal and hence a creation of nature. Its owner, who lives in the United States, provided the samples to Rakovan to assess the crystallinity of four specimens, all of which had been found decades ago in Venezuela.« less

  16. World's Largest Gold Crystal Studied at Los Alamos

    ScienceCinema

    Vogel, Sven; Nakotte, Heinz

    2018-02-07

    When geologist John Rakovan needed better tools to investigate whether a dazzling 217.78-gram piece of gold was in fact the world's largest single-crystal specimen - a distinguishing factor that would not only drastically increase its market value but also provide a unique research opportunity - he traveled to Los Alamos National Laboratory's Lujan Neutron Scattering Center to peer deep inside the mineral using neutron diffractometry. Neutrons, different from other probes such as X-rays and electrons, are able to penetrate many centimeters deep into most materials. Revealing the inner structure of a crystal without destroying the sample - imperative, as this one is worth an estimated $1.5 million - would allow Rakovan and Lujan Center collaborators Sven Vogel and Heinz Nakotte to prove that this exquisite nugget, which seemed almost too perfect and too big to be real, was a single crystal and hence a creation of nature. Its owner, who lives in the United States, provided the samples to Rakovan to assess the crystallinity of four specimens, all of which had been found decades ago in Venezuela.

  17. Vectorized Monte Carlo methods for reactor lattice analysis

    NASA Technical Reports Server (NTRS)

    Brown, F. B.

    1984-01-01

    Some of the new computational methods and equivalent mathematical representations of physics models used in the MCV code, a vectorized continuous-enery Monte Carlo code for use on the CYBER-205 computer are discussed. While the principal application of MCV is the neutronics analysis of repeating reactor lattices, the new methods used in MCV should be generally useful for vectorizing Monte Carlo for other applications. For background, a brief overview of the vector processing features of the CYBER-205 is included, followed by a discussion of the fundamentals of Monte Carlo vectorization. The physics models used in the MCV vectorized Monte Carlo code are then summarized. The new methods used in scattering analysis are presented along with details of several key, highly specialized computational routines. Finally, speedups relative to CDC-7600 scalar Monte Carlo are discussed.

  18. The control unit of the near infrared spectrograph of the Euclid space mission: detailed design

    NASA Astrophysics Data System (ADS)

    Toledo-Moreo, Rafael; Colodro-Conde, Carlos; Gómez-Sáenz-de-Tejada, Jaime; Pérez-Lizán, David; Díaz-García, José Javier; Tubío-Araujo, Óscar; Raichs, Cayetano; Catalán, Jordi; Rebolo-López, Rafael

    2016-07-01

    The Near Infrared Spectrograph and Photometer (NISP) is one of the instruments on board the ESA EUCLID mission. The Universidad Politécnica de Cartagena and Instituto de Astrofísica de Canarias are responsible of the Instrument Control Unit of the NISP (NI-ICU) in the Euclid Consortium. The NI-ICU hardware is developed by CRISA (Airbus Defence and Space), and its main functions are: communication with the S/C and the Data Processing Unit, control of the Filter and Grism Wheels, control of the Calibration Unit and thermal control of the instrument. This paper presents the NI-ICU status of definition and design at the end of the detailed design phase.

  19. The boot software of the control unit of the near infrared spectrograph of the Euclid space mission: technical specification

    NASA Astrophysics Data System (ADS)

    Gómez-Sáenz-de-Tejada, Jaime; Toledo-Moreo, Rafael; Colodro-Conde, Carlos; Pérez-Lizán, David; Fernández-Conde, Jesús; Sánchez-Prieto, Sebastián.

    2016-07-01

    The Near Infrared Spectrograph and Photometer (NISP) is one of the instruments on board the ESA EUCLID mission. The Boot Software (BSW) is in charge of initialization and communications after a reset occurs at hard- ware level. The Universidad Politecnica de Cartagena and Instituto de Astrofisica de Canarias are responsible of the Instrument Control Unit of the NISP (NI-ICU) in the Euclid Consortium. The NI-ICU BSW is developed by Universidad de Alcaĺa, and its main functions are: communication with the S/C for memory management, self-tests and start of a patchable Application Software (ASW). This paper presents the NI-ICU BSW status of definition and design at the end of the Technical Specification phase.

  20. Monte Carlo Techniques for Nuclear Systems - Theory Lectures

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Brown, Forrest B.

    These are lecture notes for a Monte Carlo class given at the University of New Mexico. The following topics are covered: course information; nuclear eng. review & MC; random numbers and sampling; computational geometry; collision physics; tallies and statistics; eigenvalue calculations I; eigenvalue calculations II; eigenvalue calculations III; variance reduction; parallel Monte Carlo; parameter studies; fission matrix and higher eigenmodes; doppler broadening; Monte Carlo depletion; HTGR modeling; coupled MC and T/H calculations; fission energy deposition. Solving particle transport problems with the Monte Carlo method is simple - just simulate the particle behavior. The devil is in the details, however. Thesemore » lectures provide a balanced approach to the theory and practice of Monte Carlo simulation codes. The first lectures provide an overview of Monte Carlo simulation methods, covering the transport equation, random sampling, computational geometry, collision physics, and statistics. The next lectures focus on the state-of-the-art in Monte Carlo criticality simulations, covering the theory of eigenvalue calculations, convergence analysis, dominance ratio calculations, bias in Keff and tallies, bias in uncertainties, a case study of a realistic calculation, and Wielandt acceleration techniques. The remaining lectures cover advanced topics, including HTGR modeling and stochastic geometry, temperature dependence, fission energy deposition, depletion calculations, parallel calculations, and parameter studies. This portion of the class focuses on using MCNP to perform criticality calculations for reactor physics and criticality safety applications. It is an intermediate level class, intended for those with at least some familiarity with MCNP. Class examples provide hands-on experience at running the code, plotting both geometry and results, and understanding the code output. The class includes lectures & hands-on computer use for a variety of Monte Carlo

  1. Random Numbers and Monte Carlo Methods

    NASA Astrophysics Data System (ADS)

    Scherer, Philipp O. J.

    Many-body problems often involve the calculation of integrals of very high dimension which cannot be treated by standard methods. For the calculation of thermodynamic averages Monte Carlo methods are very useful which sample the integration volume at randomly chosen points. After summarizing some basic statistics, we discuss algorithms for the generation of pseudo-random numbers with given probability distribution which are essential for all Monte Carlo methods. We show how the efficiency of Monte Carlo integration can be improved by sampling preferentially the important configurations. Finally the famous Metropolis algorithm is applied to classical many-particle systems. Computer experiments visualize the central limit theorem and apply the Metropolis method to the traveling salesman problem.

  2. Carbon Stripper Foils Used in the Los Alamos PSR

    NASA Astrophysics Data System (ADS)

    Borden, M.; Plum, M. A.; Sugai, I.

    1997-05-01

    Carbon stripper foils produced by the modified controlled ACDC arc discharge method (mCADAD) at the Institute for Nuclear Study by Dr. Isao Sugai have been tested and used for high current 800-MeV beam production in the Proton Storage Ring (PSR) since 1993. Two approximately 110 μg/cm2 foils are sandwiched together to produce an equivalent 220 μg/cm^2 foil. The combined foil is supported by 4-5 μm diameter carbon fibers attached to an aluminum frame. These foils have survived as long as five months during PSR normal beam production of near 70 μA on target average current. Typical life-times of other foils vary from seven to fourteen days with lower on-target average current. Beam loss data also indicate that Sugai's foils have slower shrinkage rates than other foils. Equipment has been assembled and used to produce foils by the mCADAD method at Los Alamos. These foils will be tested during 1997 operation.

  3. Monte Carlo simulations in X-ray imaging

    NASA Astrophysics Data System (ADS)

    Giersch, Jürgen; Durst, Jürgen

    2008-06-01

    Monte Carlo simulations have become crucial tools in many fields of X-ray imaging. They help to understand the influence of physical effects such as absorption, scattering and fluorescence of photons in different detector materials on image quality parameters. They allow studying new imaging concepts like photon counting, energy weighting or material reconstruction. Additionally, they can be applied to the fields of nuclear medicine to define virtual setups studying new geometries or image reconstruction algorithms. Furthermore, an implementation of the propagation physics of electrons and photons allows studying the behavior of (novel) X-ray generation concepts. This versatility of Monte Carlo simulations is illustrated with some examples done by the Monte Carlo simulation ROSI. An overview of the structure of ROSI is given as an example of a modern, well-proven, object-oriented, parallel computing Monte Carlo simulation for X-ray imaging.

  4. Multilevel sequential Monte Carlo samplers

    DOE PAGES

    Beskos, Alexandros; Jasra, Ajay; Law, Kody; ...

    2016-08-24

    Here, we study the approximation of expectations w.r.t. probability distributions associated to the solution of partial differential equations (PDEs); this scenario appears routinely in Bayesian inverse problems. In practice, one often has to solve the associated PDE numerically, using, for instance finite element methods and leading to a discretisation bias, with the step-size level h L. In addition, the expectation cannot be computed analytically and one often resorts to Monte Carlo methods. In the context of this problem, it is known that the introduction of the multilevel Monte Carlo (MLMC) method can reduce the amount of computational effort to estimate expectations, for a given level of error. This is achieved via a telescoping identity associated to a Monte Carlo approximation of a sequence of probability distributions with discretisation levelsmore » $${\\infty}$$ >h 0>h 1 ...>h L. In many practical problems of interest, one cannot achieve an i.i.d. sampling of the associated sequence of probability distributions. A sequential Monte Carlo (SMC) version of the MLMC method is introduced to deal with this problem. In conclusion, it is shown that under appropriate assumptions, the attractive property of a reduction of the amount of computational effort to estimate expectations, for a given level of error, can be maintained within the SMC context.« less

  5. Environmental surveillance at Los Alamos during 1991. Environmental protection group

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Dewart, J.; Kohen, K.L.

    1993-08-01

    This report describes the environmental surveillance program conducted by Los Alamos National Laboratory during 1991. Routine monitoring for radiation and for radioactive and chemical materials is conducted on the Laboratory site as well as in the surrounding region. Monitoring results are used to determine compliance with appropriate standards and to permit early identification of potentially undesirable trends. Results and interpretation of data for 1991 cover external penetrating radiation; quantities of airborne emissions and effluents; concentrations of chemicals and radionuclides in ambient air, surface waters and groundwaters, municipal water supply, soils and sediments, and foodstuffs; and environmental compliance. Comparisons with appropriatemore » standards, regulations, and background levels provide the basis for concluding that environmental effects from Laboratory operations are small and do not pose a threat to the public, Laboratory employees, or the environment.« less

  6. Recent Research with the Detector for Advanced Neutron Capture Experiments (dance) at the LOS Alamos Neutron Science Center

    NASA Astrophysics Data System (ADS)

    Ullmann, J. L.

    2014-09-01

    The DANCE detector at Los Alamos is a 160 element, nearly 4π BaF2 detector array designed to make measurements of neutron capture on rare or radioactive nuclides. It has also been used to make measurements of gamma-ray multiplicity following capture and gamma-ray output from fission. Several examples of measurements are briefly discussed.

  7. The Los Alamos Laser Acceleration of Particles Workshop and beginning of the advanced accelerator concepts field

    NASA Astrophysics Data System (ADS)

    Joshi, C.

    2012-12-01

    The first Advanced Acceleration of Particles-AAC-Workshop (actually named Laser Acceleration of Particles Workshop) was held at Los Alamos in January 1982. The workshop lasted a week and divided all the acceleration techniques into four categories: near field, far field, media, and vacuum. Basic theorems of particle acceleration were postulated (later proven) and specific experiments based on the four categories were formulated. This landmark workshop led to the formation of the advanced accelerator R&D program in the HEP office of the DOE that supports advanced accelerator research to this day. Two major new user facilities at Argonne and Brookhaven and several more directed experimental efforts were built to explore the advanced particle acceleration schemes. It is not an exaggeration to say that the intellectual breadth and excitement provided by the many groups who entered this new field provided the needed vitality to then recently formed APS Division of Beams and the new online journal Physical Review Special Topics-Accelerators and Beams. On this 30th anniversary of the AAC Workshops, it is worthwhile to look back at the legacy of the first Workshop at Los Alamos and the fine groundwork it laid for the field of advanced accelerator concepts that continues to flourish to this day.

  8. Nuclear criticality safety staff training and qualifications at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Monahan, S.P.; McLaughlin, T.P.

    1997-05-01

    Operations involving significant quantities of fissile material have been conducted at Los Alamos National Laboratory continuously since 1943. Until the advent of the Laboratory`s Nuclear Criticality Safety Committee (NCSC) in 1957, line management had sole responsibility for controlling criticality risks. From 1957 until 1961, the NCSC was the Laboratory body which promulgated policy guidance as well as some technical guidance for specific operations. In 1961 the Laboratory created the position of Nuclear Criticality Safety Office (in addition to the NCSC). In 1980, Laboratory management moved the Criticality Safety Officer (and one other LACEF staff member who, by that time, wasmore » also working nearly full-time on criticality safety issues) into the Health Division office. Later that same year the Criticality Safety Group, H-6 (at that time) was created within H-Division, and staffed by these two individuals. The training and education of these individuals in the art of criticality safety was almost entirely self-regulated, depending heavily on technical interactions between each other, as well as NCSC, LACEF, operations, other facility, and broader criticality safety community personnel. Although the Los Alamos criticality safety group has grown both in size and formality of operations since 1980, the basic philosophy that a criticality specialist must be developed through mentoring and self motivation remains the same. Formally, this philosophy has been captured in an internal policy, document ``Conduct of Business in the Nuclear Criticality Safety Group.`` There are no short cuts or substitutes in the development of a criticality safety specialist. A person must have a self-motivated personality, excellent communications skills, a thorough understanding of the principals of neutron physics, a safety-conscious and helpful attitude, a good perspective of real risk, as well as a detailed understanding of process operations and credible upsets.« less

  9. Capabilities for high explosive pulsed power research at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Goforth, James H; Oona, Henn; Tasker, Douglas G

    2008-01-01

    Research on topics requiring high magnetic fields and high currents have been pursued using high explosive pulsed power (HEPP) techniques since the 1950s at Los Alamos National Laboratory. We have developed many sophisticated HEPr systems through the years, and most of them depend on technology available from the nuclear weapons program. Through the 1980s and 1990s, our budgets would sustain parallel efforts in zpinch research using both HEPr and capacitor banks. In recent years, many changes have occurred that are driven by concerns such as safety, security, and environment, as well as reduced budgets and downsizing of the National Nuclearmore » Security Administration (NNSA) complex due to the end of the cold war era. In this paper, we review the teclmiques developed to date, and adaptations that are driven by changes in budgets and our changing complex. One new Ranchero-based solid liner z-pinch experimental design is also presented. Explosives that are cast to shape instead of being machined, and initiation systems that depend on arrays of slapper detonators are important new tools. Some materials that are seen as hazardous to the environment are avoided in designs. The process continues to allow a wide range of research however, and there are few, if any, experiments that we have done in the past that could not be perform today. The HErr firing facility at Los Alamos continues to have a 2000 lb. high explosive limit, and our 2.4 MJ capacitor bank remains a mainstay of the effort. Modem diagnostic and data analysis capabilities allow fewer personnel to achieve better results, and in the broad sense we continue to have a robust capability.« less

  10. Summarizing Monte Carlo Results in Methodological Research.

    ERIC Educational Resources Information Center

    Harwell, Michael R.

    Monte Carlo studies of statistical tests are prominently featured in the methodological research literature. Unfortunately, the information from these studies does not appear to have significantly influenced methodological practice in educational and psychological research. One reason is that Monte Carlo studies lack an overarching theory to guide…

  11. Advanced Computational Methods for Monte Carlo Calculations

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Brown, Forrest B.

    This course is intended for graduate students who already have a basic understanding of Monte Carlo methods. It focuses on advanced topics that may be needed for thesis research, for developing new state-of-the-art methods, or for working with modern production Monte Carlo codes.

  12. Quantum speedup of Monte Carlo methods.

    PubMed

    Montanaro, Ashley

    2015-09-08

    Monte Carlo methods use random sampling to estimate numerical quantities which are hard to compute deterministically. One important example is the use in statistical physics of rapidly mixing Markov chains to approximately compute partition functions. In this work, we describe a quantum algorithm which can accelerate Monte Carlo methods in a very general setting. The algorithm estimates the expected output value of an arbitrary randomized or quantum subroutine with bounded variance, achieving a near-quadratic speedup over the best possible classical algorithm. Combining the algorithm with the use of quantum walks gives a quantum speedup of the fastest known classical algorithms with rigorous performance bounds for computing partition functions, which use multiple-stage Markov chain Monte Carlo techniques. The quantum algorithm can also be used to estimate the total variation distance between probability distributions efficiently.

  13. Proton Upset Monte Carlo Simulation

    NASA Technical Reports Server (NTRS)

    O'Neill, Patrick M.; Kouba, Coy K.; Foster, Charles C.

    2009-01-01

    The Proton Upset Monte Carlo Simulation (PROPSET) program calculates the frequency of on-orbit upsets in computer chips (for given orbits such as Low Earth Orbit, Lunar Orbit, and the like) from proton bombardment based on the results of heavy ion testing alone. The software simulates the bombardment of modern microelectronic components (computer chips) with high-energy (.200 MeV) protons. The nuclear interaction of the proton with the silicon of the chip is modeled and nuclear fragments from this interaction are tracked using Monte Carlo techniques to produce statistically accurate predictions.

  14. Overview of the 1997 Dirac High-Magnetic Series at LOS Alamos

    NASA Astrophysics Data System (ADS)

    Clark, D. A.; Campbell, L. J.; Forman, K. C.; Fowler, C. M.; Goettee, J. D.; Mielke, C. H.; Rickel, D. G.; Marshall, B. R.

    2004-11-01

    During the summer of 1997, a series of high magnetic field experiments was conducted at Los Alamos National Laboratory. Four experiments utilizing Russian built MC-1 generators, which can reach fields as high as 10 Megagauss, and four smaller strip generator experiments at fields near 1.5 Megagauss were conducted. Experiments mounted on the devices included magnetoresistance of high temperature superconductors and semiconductors, optical reflectivity (conductivity) of semiconductors, magnetization of a magnetic cluster material and a semiconductor, Faraday rotation in a semiconductor and a magnetic cluster material, and transmission spectroscopy of molecules. Brief descriptions of the experimental setups, magnetic field measurement techniques, field results and various experiments are presented. Magnetic field data and other information on Dirac `97 can be found at .

  15. MANHATTAN: The View From Los Alamos of History's Most Secret Project

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Carr, Alan Brady

    This presentation covers the political and scientific events leading up to the creation of the Manhattan Project. The creation of the Manhattan Project’s three most significant sites--Los Alamos, Oak Ridge, and Hanford--is also discussed. The lecture concludes by exploring the use of the atomic bombs at the end of World War II. The presentation slides include three videos. The first is a short clip of the 100-ton Test. The 100-Ton Test was history’s largest measured blast at that point in time; it was a pre-test for Trinity, the world’s first nuclear detonation. The second clip features views of Trinity followedmore » a short statement by the Laboratory’s first director, J. Robert Oppenheimer. The final clip shows Norris Bradbury talking about arms control.« less

  16. The Los Alamos Seismic Network (LASN): Recent Network Upgrades and Northern New Mexico Earthquake Catalog Updates

    NASA Astrophysics Data System (ADS)

    Roberts, P. M.; House, L. S.; Greene, M.; Ten Cate, J. A.; Schultz-Fellenz, E. S.; Kelley, R.

    2012-12-01

    From the first data recorded in the fall of 1973 to now, the Los Alamos Seismograph Network (LASN) has operated for nearly 40 years. LASN data have been used to locate more than 2,500 earthquakes in north-central New Mexico. The network was installed for seismic verification research, as well as to monitor and locate earthquakes near Los Alamos National Laboratory (LANL). LASN stations are the only earthquake monitoring stations in New Mexico north of Albuquerque. In the late 1970s, LASN included 22 stations spread over a geographic area of 150 km (N-S) by 350 km (E-W), of northern New Mexico. In the early 1980s, the available funding limited the stations that could be operated to a set of 7, located within an area of about 15 km (N-S) by 15 km (E-W), centered on Los Alamos. Over the last 3 years, 6 additional stations have been installed, which have considerably expanded the spatial coverage of the network. These new stations take advantage of broadband state-of-the-art sensors as well as digital recording and telemetry technology. Currently, 7 stations have broadband, three-component seismometers with digital telemetry, and the remaining 6 have traditional 1 Hz short-period seismometers with analog telemetry. In addition, a vertical array of accelerometers was installed in a wellbore on LANL property. This borehole station has 3-component digital strong-motion sensors. In addition, four forensic strong-motion accelerometers (SMA) are operated at LANL facilities. With 3 of the new broadband stations in and around the nearby Valles Caldera, LASN is now able to monitor any very small volcano-seismic events that may be associated with the caldera. We will present a complete description of the current LASN station, instrumentation and telemetry configurations, as well as the data acquisition and event-detection software structure used to record events in Earthworm. More than 2,000 earthquakes were detected and located in north-central New Mexico during the first 11

  17. Los Alamos Shows Airport Security Technology at Work

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Espy, Michelle; Schultz, Larry; Hunter, James

    Los Alamos scientists have advanced a Magnetic Resonance Imaging (MRI) technology that may provide a breakthrough for screening liquids at airport security. They've added low-power X-ray data to the mix, and as a result have unlocked a new detection technology. Funded in part by the Department of Homeland Security's Science and Technology Directorate, the new system is named MagRay. The goal is to quickly and accurately distinguish between liquids that visually appear identical. For example, what appears to be a bottle of white wine could potentially be nitromethane, a liquid that could be used to make an explosive. Both aremore » clear liquids, one would be perfectly safe on a commercial aircraft, the other would be strictly prohibited. How to tell them apart quickly without error at an airport security area is the focus of Michelle Espy, Larry Schultz and their team. In this video, Espy and the MagRay team explain how the new technology works, how they've developed an easy operator interface, and what the next steps might be in transitioning this technology to the private sector.« less

  18. Organizational cultural survey of the Los Alamos Site

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Not Available

    An Organizational Survey (OS) was administered at the Los Alamos Site that queried employees on the subjects of organizational culture, various aspects of communications, employee commitment, work group cohesion, coordination of work, environmental, safety, and health concerns, hazardous nature of work, safety and overall job satisfaction. The purpose of the OS is to measure in a quantitative and objective way the notion of ``culture;`` that is, the values, attitudes, and beliefs of the individuals working within the organization. In addition, through the OS, a broad sample of individuals can be reached that would probably not be interviewed or observed duringmore » the course of a typical assessment. The OS also provides a descriptive profile of the organization at one point in time that can then be compared to a profile taken at a different point in time to assess changes in the culture of the organization. While comparisons among groups are made, it is not the purpose of this report to make evaluative statements of which profile may be positive or negative. However, using the data presented in this report in conjunction with other evaluative activities, may provide useful insight into the organization.« less

  19. Organizational cultural survey of the Los Alamos Site

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Not Available

    An Organizational Survey (OS) was administered at the Los Alamos Site that queried employees on the subjects of organizational culture, various aspects of communications, employee commitment, work group cohesion, coordination of work, environmental, safety, and health concerns, hazardous nature of work, safety and overall job satisfaction. The purpose of the OS is to measure in a quantitative and objective way the notion of culture;'' that is, the values, attitudes, and beliefs of the individuals working within the organization. In addition, through the OS, a broad sample of individuals can be reached that would probably not be interviewed or observed duringmore » the course of a typical assessment. The OS also provides a descriptive profile of the organization at one point in time that can then be compared to a profile taken at a different point in time to assess changes in the culture of the organization. While comparisons among groups are made, it is not the purpose of this report to make evaluative statements of which profile may be positive or negative. However, using the data presented in this report in conjunction with other evaluative activities, may provide useful insight into the organization.« less

  20. Monte Carlo modeling of spatial coherence: free-space diffraction

    PubMed Central

    Fischer, David G.; Prahl, Scott A.; Duncan, Donald D.

    2008-01-01

    We present a Monte Carlo method for propagating partially coherent fields through complex deterministic optical systems. A Gaussian copula is used to synthesize a random source with an arbitrary spatial coherence function. Physical optics and Monte Carlo predictions of the first- and second-order statistics of the field are shown for coherent and partially coherent sources for free-space propagation, imaging using a binary Fresnel zone plate, and propagation through a limiting aperture. Excellent agreement between the physical optics and Monte Carlo predictions is demonstrated in all cases. Convergence criteria are presented for judging the quality of the Monte Carlo predictions. PMID:18830335

  1. Stormwater Pollution Prevention Plan for the TA-03-38 Carpenter's Shop, Los Alamos National Laboratory, Revision 3, January 2018

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Burgin, Jillian Elizabeth

    This Storm Water Pollution Prevention Plan (SWPPP) was developed in accordance with the provisions of the Clean Water Act (33 U.S.C. §§1251 et seq., as amended), and the Multi-Sector General Permit for Storm Water Discharges Associated with Industrial Activity (U.S. EPA, June 2015) issued by the U.S. Environmental Protection Agency (EPA) for the National Pollutant Discharge Elimination System (NPDES) and using the industry specific permit requirements for Sector A–Timber Products, Subsector A4 (Wood Products Facilities not elsewhere classified) as a guide. This SWPPP applies to discharges of stormwater from the operational areas of the TA-03-38 Carpenter’s Shop at Los Alamosmore » National Laboratory. Los Alamos National Laboratory (also referred to as LANL or the “Laboratory”) is owned by the Department of Energy (DOE), and is operated by Los Alamos National Security, LLC (LANS). Throughout this document, the term “facility” refers to the TA-03-38 Carpenter’s Shop and associated areas. The current permit expires at midnight on June 4, 2020.« less

  2. Quantum speedup of Monte Carlo methods

    PubMed Central

    Montanaro, Ashley

    2015-01-01

    Monte Carlo methods use random sampling to estimate numerical quantities which are hard to compute deterministically. One important example is the use in statistical physics of rapidly mixing Markov chains to approximately compute partition functions. In this work, we describe a quantum algorithm which can accelerate Monte Carlo methods in a very general setting. The algorithm estimates the expected output value of an arbitrary randomized or quantum subroutine with bounded variance, achieving a near-quadratic speedup over the best possible classical algorithm. Combining the algorithm with the use of quantum walks gives a quantum speedup of the fastest known classical algorithms with rigorous performance bounds for computing partition functions, which use multiple-stage Markov chain Monte Carlo techniques. The quantum algorithm can also be used to estimate the total variation distance between probability distributions efficiently. PMID:26528079

  3. The X-43A Six Degree of Freedom Monte Carlo Analysis

    NASA Technical Reports Server (NTRS)

    Baumann, Ethan; Bahm, Catherine; Strovers, Brian; Beck, Roger

    2008-01-01

    This report provides an overview of the Hyper-X research vehicle Monte Carlo analysis conducted with the six-degree-of-freedom simulation. The methodology and model uncertainties used for the Monte Carlo analysis are presented as permitted. In addition, the process used to select hardware validation test cases from the Monte Carlo data is described. The preflight Monte Carlo analysis indicated that the X-43A control system was robust to the preflight uncertainties and provided the Hyper-X project an important indication that the vehicle would likely be successful in accomplishing the mission objectives. The X-43A inflight performance is compared to the preflight Monte Carlo predictions and shown to exceed the Monte Carlo bounds in several instances. Possible modeling shortfalls are presented that may account for these discrepancies. The flight control laws and guidance algorithms were robust enough as a result of the preflight Monte Carlo analysis that the unexpected in-flight performance did not have undue consequences. Modeling and Monte Carlo analysis lessons learned are presented.

  4. The X-43A Six Degree of Freedom Monte Carlo Analysis

    NASA Technical Reports Server (NTRS)

    Baumann, Ethan; Bahm, Catherine; Strovers, Brian; Beck, Roger; Richard, Michael

    2007-01-01

    This report provides an overview of the Hyper-X research vehicle Monte Carlo analysis conducted with the six-degree-of-freedom simulation. The methodology and model uncertainties used for the Monte Carlo analysis are presented as permitted. In addition, the process used to select hardware validation test cases from the Monte Carlo data is described. The preflight Monte Carlo analysis indicated that the X-43A control system was robust to the preflight uncertainties and provided the Hyper-X project an important indication that the vehicle would likely be successful in accomplishing the mission objectives. The X-43A in-flight performance is compared to the preflight Monte Carlo predictions and shown to exceed the Monte Carlo bounds in several instances. Possible modeling shortfalls are presented that may account for these discrepancies. The flight control laws and guidance algorithms were robust enough as a result of the preflight Monte Carlo analysis that the unexpected in-flight performance did not have undue consequences. Modeling and Monte Carlo analysis lessons learned are presented.

  5. MontePython 3: Parameter inference code for cosmology

    NASA Astrophysics Data System (ADS)

    Brinckmann, Thejs; Lesgourgues, Julien; Audren, Benjamin; Benabed, Karim; Prunet, Simon

    2018-05-01

    MontePython 3 provides numerous ways to explore parameter space using Monte Carlo Markov Chain (MCMC) sampling, including Metropolis-Hastings, Nested Sampling, Cosmo Hammer, and a Fisher sampling method. This improved version of the Monte Python (ascl:1307.002) parameter inference code for cosmology offers new ingredients that improve the performance of Metropolis-Hastings sampling, speeding up convergence and offering significant time improvement in difficult runs. Additional likelihoods and plotting options are available, as are post-processing algorithms such as Importance Sampling and Adding Derived Parameter.

  6. Historic Manhattan Project Sites at Los Alamos

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    McGehee, Ellen

    The Manhattan Project laboratory constructed at Los Alamos, New Mexico, beginning in 1943, was intended from the start to be temporary and to go up with amazing speed. Because most of those WWII-era facilities were built with minimal materials and so quickly, much of the original infrastructure was torn down in the late '40s and early '50s and replaced by more permanent facilities. However, a few key facilities remained, and are being preserved and maintained for historic significance. Four such sites are visited briefly in this video, taking viewers to V-Site, the buildings where the first nuclear explosive device wasmore » pre-assembled in preparation for the Trinity Test in Southern New Mexico. Included is another WWII area, Gun Site. So named because it was the area where scientists and engineers tested the so-called "gun method" of assembling nuclear materials -- the fundamental design of the Little Boy weapon that was eventually dropped on Hiroshima. The video also goes to Pajarito Site, home of the "Slotin Building" and "Pond Cabin." The Slotin Building is the place where scientist Louis Slotin conducted a criticality experiment that went awry in early 1946, leading to his unfortunate death, and the Pond Cabin served the team of eminent scientist Emilio Segre who did early chemistry work on plutonium that ultimately led to the Fat Man weapon.« less

  7. Historic Manhattan Project Sites at Los Alamos

    ScienceCinema

    McGehee, Ellen

    2018-05-11

    The Manhattan Project laboratory constructed at Los Alamos, New Mexico, beginning in 1943, was intended from the start to be temporary and to go up with amazing speed. Because most of those WWII-era facilities were built with minimal materials and so quickly, much of the original infrastructure was torn down in the late '40s and early '50s and replaced by more permanent facilities. However, a few key facilities remained, and are being preserved and maintained for historic significance. Four such sites are visited briefly in this video, taking viewers to V-Site, the buildings where the first nuclear explosive device was pre-assembled in preparation for the Trinity Test in Southern New Mexico. Included is another WWII area, Gun Site. So named because it was the area where scientists and engineers tested the so-called "gun method" of assembling nuclear materials -- the fundamental design of the Little Boy weapon that was eventually dropped on Hiroshima. The video also goes to Pajarito Site, home of the "Slotin Building" and "Pond Cabin." The Slotin Building is the place where scientist Louis Slotin conducted a criticality experiment that went awry in early 1946, leading to his unfortunate death, and the Pond Cabin served the team of eminent scientist Emilio Segre who did early chemistry work on plutonium that ultimately led to the Fat Man weapon.

  8. Floodplain Assessment for the Middle Los Alamos Canyon Aggregate Area Investigations in Technical Area 02 at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hathcock, Charles Dean

    The proposed action being assessed in this document occurs in TA-02 in the bottom of Los Alamos Canyon. The DOE proposes to conduct soil sampling at AOC 02-011 (d), AOC 02- 011(a)(ii), and SWMU 02-005, and excavate soils in AOC 02-011(a)(ii) as part of a corrective actions effort. Additional shallow surface soil samples (soil grab samples) will be collected throughout the TA-02 area, including within the floodplain, to perform ecotoxicology studies (Figures 1 and 2). The excavation boundaries in AOC 02-011(a)(ii) are slightly within the delineated 100-year floodplain. The project will use a variety of techniques for soil sampling andmore » remediation efforts to include hand/digging, standard hand auger/sampling, excavation using machinery such as backhoe and front end loader and small drill rig. Heavy equipment will traverse the floodplain and spoils piles will be staged in the floodplain within developed or previously disturbed areas (e.g., existing paved roads and parking areas). The project will utilize and maintain appropriate best management practices (BMPs) to contain excavated materials, and all pollutants, including oil from machinery/vehicles. The project will stabilize disturbed areas as appropriate at the end of the project.« less

  9. An unbiased Hessian representation for Monte Carlo PDFs.

    PubMed

    Carrazza, Stefano; Forte, Stefano; Kassabov, Zahari; Latorre, José Ignacio; Rojo, Juan

    We develop a methodology for the construction of a Hessian representation of Monte Carlo sets of parton distributions, based on the use of a subset of the Monte Carlo PDF replicas as an unbiased linear basis, and of a genetic algorithm for the determination of the optimal basis. We validate the methodology by first showing that it faithfully reproduces a native Monte Carlo PDF set (NNPDF3.0), and then, that if applied to Hessian PDF set (MMHT14) which was transformed into a Monte Carlo set, it gives back the starting PDFs with minimal information loss. We then show that, when applied to a large Monte Carlo PDF set obtained as combination of several underlying sets, the methodology leads to a Hessian representation in terms of a rather smaller set of parameters (MC-H PDFs), thereby providing an alternative implementation of the recently suggested Meta-PDF idea and a Hessian version of the recently suggested PDF compression algorithm (CMC-PDFs). The mc2hessian conversion code is made publicly available together with (through LHAPDF6) a Hessian representations of the NNPDF3.0 set, and the MC-H PDF set.

  10. Diffusion Monte Carlo approach versus adiabatic computation for local Hamiltonians

    NASA Astrophysics Data System (ADS)

    Bringewatt, Jacob; Dorland, William; Jordan, Stephen P.; Mink, Alan

    2018-02-01

    Most research regarding quantum adiabatic optimization has focused on stoquastic Hamiltonians, whose ground states can be expressed with only real non-negative amplitudes and thus for whom destructive interference is not manifest. This raises the question of whether classical Monte Carlo algorithms can efficiently simulate quantum adiabatic optimization with stoquastic Hamiltonians. Recent results have given counterexamples in which path-integral and diffusion Monte Carlo fail to do so. However, most adiabatic optimization algorithms, such as for solving MAX-k -SAT problems, use k -local Hamiltonians, whereas our previous counterexample for diffusion Monte Carlo involved n -body interactions. Here we present a 6-local counterexample which demonstrates that even for these local Hamiltonians there are cases where diffusion Monte Carlo cannot efficiently simulate quantum adiabatic optimization. Furthermore, we perform empirical testing of diffusion Monte Carlo on a standard well-studied class of permutation-symmetric tunneling problems and similarly find large advantages for quantum optimization over diffusion Monte Carlo.

  11. Monte Carlo simulation of aorta autofluorescence

    NASA Astrophysics Data System (ADS)

    Kuznetsova, A. A.; Pushkareva, A. E.

    2016-08-01

    Results of numerical simulation of autofluorescence of the aorta by the method of Monte Carlo are reported. Two states of the aorta, normal and with atherosclerotic lesions, are studied. A model of the studied tissue is developed on the basis of information about optical, morphological, and physico-chemical properties. It is shown that the data obtained by numerical Monte Carlo simulation are in good agreement with experimental results indicating adequacy of the developed model of the aorta autofluorescence.

  12. Chemical decontamination technical resources at Los Alamos National Laboratory (2008)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Moore, Murray E

    This document supplies information resources for a person seeking to create planning or pre-planning documents for chemical decontamination operations. A building decontamination plan can be separated into four different sections: Pre-planning, Characterization, Decontamination (Initial response and also complete cleanup), and Clearance. Of the identified Los Alamos resources, they can be matched with these four sections: Pre-planning -- Dave Seidel, EO-EPP, Emergency Planning and Preparedness; David DeCroix and Bruce Letellier, D-3, Computational fluids modeling of structures; Murray E. Moore, RP-2, Aerosol sampling and ventilation engineering. Characterization (this can include development projects) -- Beth Perry, IAT-3, Nuclear Counterterrorism Response (SNIPER database); Fernandomore » Garzon, MPA-11, Sensors and Electrochemical Devices (development); George Havrilla, C-CDE, Chemical Diagnostics and Engineering; Kristen McCabe, B-7, Biosecurity and Public Health. Decontamination -- Adam Stively, EO-ER, Emergency Response; Dina Matz, IHS-IP, Industrial hygiene; Don Hickmott, EES-6, Chemical cleanup. Clearance (validation) -- Larry Ticknor, CCS-6, Statistical Sciences.« less

  13. Environmental Assessment and Finding of No Significant Impact: The Proposed Issuance of an Easement to Public Service Company of New Mexico for the Construction and Operation of a 12-inch Natural Gas Pipeline within Los Alamos National Laboratory, Los Alamos, New Mexico

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    N /A

    2002-07-30

    The National Nuclear Security Administration (NNSA) has assigned a continuing role to Los Alamos National Laboratory (LANL) in carrying out NNSAs national security mission. To enable LANL to continue this enduring responsibility requires that NNSA maintain the capabilities and capacities required in support of its national mission assignments at LANL. To carry out its Congressionally assigned mission requirements, NNSA must maintain a safe and reliable infrastructure at LANL. Upgrades to the various utility services at LANL have been ongoing together with routine maintenance activities over the years. However, the replacement of a certain portion of natural gas service transmission pipelinemore » is now necessary as this delivery system element has been operating well beyond its original design life for the past 20 to 30 years and components of the line are suffering from normal stresses, strains, and general failures. The Proposed Action is to grant an easement to the Public Service Company of New Mexico (PNM) to construct, operate, and maintain approximately 15,000 feet (4,500 meters) of 12-inch (in.) (30-centimeter [cm]) coated steel natural gas transmission mainline on NNSA-administered land within LANL along Los Alamos Canyon. The new gas line would begin at the existing valve setting located at the bottom of Los Alamos Canyon near the Los Alamos County water well pump house and adjacent to the existing 12-in. (30-cm) PNM gas transmission mainline. The new gas line (owned by PNM) would then cross the streambed and continue east in a new easement obtained by PNM from the NNSA, paralleling the existing electrical power line along the bottom of the canyon. The gas line would then turn northeast near State Road (SR) 4 and be connected to the existing 12-in. (30-cm) coated steel gas transmission mainline, located within the right-of-way (ROW) of SR 502. The Proposed Action would also involve crossing a streambed twice. PNM would bore under the streambed for pipe

  14. New Developments in Proton Radiography at the Los Alamos Neutron Science Center (LANSCE)

    DOE PAGES

    Morris, C. L.; Brown, E. N.; Agee, C.; ...

    2015-12-30

    An application of nuclear physics, a facility for using protons for flash radiography, was developed at the Los Alamos Neutron Science Center (LANSCE). Protons have proven far superior to high energy x-rays for flash radiography because of their long mean free path, good position resolution, and low scatter background. Although this facility is primarily used for studying very fast phenomena such as high explosive driven experiments, it is finding increasing application to other fields, such as tomography of static objects, phase changes in materials and the dynamics of chemical reactions. The advantages of protons are discussed, data from some recentmore » experiments will be reviewed and concepts for new techniques are introduced.« less

  15. Limited-scope probabilistic safety analysis for the Los Alamos Meson Physics Facility (LAMPF)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sharirli, M.; Rand, J.L.; Sasser, M.K.

    1992-01-01

    The reliability of instrumentation and safety systems is a major issue in the operation of accelerator facilities. A probabilistic safety analysis was performed or the key safety and instrumentation systems at the Los Alamos Meson Physics Facility (LAMPF). in Phase I of this unique study, the Personnel Safety System (PSS) and the Current Limiters (XLs) were analyzed through the use of the fault tree analyses, failure modes and effects analysis, and criticality analysis. Phase II of the program was done to update and reevaluate the safety systems after the Phase I recommendations were implemented. This paper provides a brief reviewmore » of the studies involved in Phases I and II of the program.« less

  16. Limited-scope probabilistic safety analysis for the Los Alamos Meson Physics Facility (LAMPF)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sharirli, M.; Rand, J.L.; Sasser, M.K.

    1992-12-01

    The reliability of instrumentation and safety systems is a major issue in the operation of accelerator facilities. A probabilistic safety analysis was performed or the key safety and instrumentation systems at the Los Alamos Meson Physics Facility (LAMPF). in Phase I of this unique study, the Personnel Safety System (PSS) and the Current Limiters (XLs) were analyzed through the use of the fault tree analyses, failure modes and effects analysis, and criticality analysis. Phase II of the program was done to update and reevaluate the safety systems after the Phase I recommendations were implemented. This paper provides a brief reviewmore » of the studies involved in Phases I and II of the program.« less

  17. Additive manufacturing capabilities applied to inertial confinement confusion at Los Alamos National Laboratory

    DOE PAGES

    Cardenas, Tana; Schmidt, Derek William; Peterson, Dominic S.

    2016-08-01

    We describe the use at Los Alamos National Laboratory of additive manufacturing (AM) for a variety of jigs and coating, assembly, and radiography fixtures. Additive manufacturing has also been used to produce shipping containers of complex design that would be too costly to have fabricated using traditional techniques. The current goal for AM use in target fabrication is to increase target accuracy and rigidity. This has been realized by implementing AM into target stalk fabrication, allowing increased complexity to address target strength and the addition of features for alignment at facilities. As a result, we will describe the fabrication ofmore » these components and our plans to utilize AM in the future.« less

  18. Discrete Diffusion Monte Carlo for Electron Thermal Transport

    NASA Astrophysics Data System (ADS)

    Chenhall, Jeffrey; Cao, Duc; Wollaeger, Ryan; Moses, Gregory

    2014-10-01

    The iSNB (implicit Schurtz Nicolai Busquet electron thermal transport method of Cao et al. is adapted to a Discrete Diffusion Monte Carlo (DDMC) solution method for eventual inclusion in a hybrid IMC-DDMC (Implicit Monte Carlo) method. The hybrid method will combine the efficiency of a diffusion method in short mean free path regions with the accuracy of a transport method in long mean free path regions. The Monte Carlo nature of the approach allows the algorithm to be massively parallelized. Work to date on the iSNB-DDMC method will be presented. This work was supported by Sandia National Laboratory - Albuquerque.

  19. Off-diagonal expansion quantum Monte Carlo

    NASA Astrophysics Data System (ADS)

    Albash, Tameem; Wagenbreth, Gene; Hen, Itay

    2017-12-01

    We propose a Monte Carlo algorithm designed to simulate quantum as well as classical systems at equilibrium, bridging the algorithmic gap between quantum and classical thermal simulation algorithms. The method is based on a decomposition of the quantum partition function that can be viewed as a series expansion about its classical part. We argue that the algorithm not only provides a theoretical advancement in the field of quantum Monte Carlo simulations, but is optimally suited to tackle quantum many-body systems that exhibit a range of behaviors from "fully quantum" to "fully classical," in contrast to many existing methods. We demonstrate the advantages, sometimes by orders of magnitude, of the technique by comparing it against existing state-of-the-art schemes such as path integral quantum Monte Carlo and stochastic series expansion. We also illustrate how our method allows for the unification of quantum and classical thermal parallel tempering techniques into a single algorithm and discuss its practical significance.

  20. Off-diagonal expansion quantum Monte Carlo.

    PubMed

    Albash, Tameem; Wagenbreth, Gene; Hen, Itay

    2017-12-01

    We propose a Monte Carlo algorithm designed to simulate quantum as well as classical systems at equilibrium, bridging the algorithmic gap between quantum and classical thermal simulation algorithms. The method is based on a decomposition of the quantum partition function that can be viewed as a series expansion about its classical part. We argue that the algorithm not only provides a theoretical advancement in the field of quantum Monte Carlo simulations, but is optimally suited to tackle quantum many-body systems that exhibit a range of behaviors from "fully quantum" to "fully classical," in contrast to many existing methods. We demonstrate the advantages, sometimes by orders of magnitude, of the technique by comparing it against existing state-of-the-art schemes such as path integral quantum Monte Carlo and stochastic series expansion. We also illustrate how our method allows for the unification of quantum and classical thermal parallel tempering techniques into a single algorithm and discuss its practical significance.

  1. RECORDS: improved Reporting of montE CarlO RaDiation transport Studies: Report of the AAPM Research Committee Task Group 268.

    PubMed

    Sechopoulos, Ioannis; Rogers, D W O; Bazalova-Carter, Magdalena; Bolch, Wesley E; Heath, Emily C; McNitt-Gray, Michael F; Sempau, Josep; Williamson, Jeffrey F

    2018-01-01

    Studies involving Monte Carlo simulations are common in both diagnostic and therapy medical physics research, as well as other fields of basic and applied science. As with all experimental studies, the conditions and parameters used for Monte Carlo simulations impact their scope, validity, limitations, and generalizability. Unfortunately, many published peer-reviewed articles involving Monte Carlo simulations do not provide the level of detail needed for the reader to be able to properly assess the quality of the simulations. The American Association of Physicists in Medicine Task Group #268 developed guidelines to improve reporting of Monte Carlo studies in medical physics research. By following these guidelines, manuscripts submitted for peer-review will include a level of relevant detail that will increase the transparency, the ability to reproduce results, and the overall scientific value of these studies. The guidelines include a checklist of the items that should be included in the Methods, Results, and Discussion sections of manuscripts submitted for peer-review. These guidelines do not attempt to replace the journal reviewer, but rather to be a tool during the writing and review process. Given the varied nature of Monte Carlo studies, it is up to the authors and the reviewers to use this checklist appropriately, being conscious of how the different items apply to each particular scenario. It is envisioned that this list will be useful both for authors and for reviewers, to help ensure the adequate description of Monte Carlo studies in the medical physics literature. © 2017 American Association of Physicists in Medicine.

  2. Threatened and Endangered Species Habitat Management Plan for Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Keller, David Charles; Hathcock, Charles Dean

    Los Alamos National Laboratory’s (LANL) Threatened and Endangered Species Habitat Management Plan (HMP) fulfills a commitment made to the U.S. Department of Energy (DOE) in the “Final Environmental Impact Statement for the Dual-Axis Radiographic Hydrodynamic Test Facility Mitigation Action Plan” (DOE 1996). The HMP received concurrence from the U.S. Fish and Wildlife Service (USFWS) in 1999 (USFWS consultation numbers 2-22-98-I-336 and 2-22-95-I-108). This 2015 update retains the management guidelines from the 1999 HMP for listed species, updates some descriptive information, and adds the New Mexico Meadow Jumping Mouse (Zapus hudsonius luteus) and Yellow-billed Cuckoo (Coccyzus americanus) which were federally listedmore » in 2014 (Keller 2015: USFWS consultation number 02ENNM00- 2015-I-0538).« less

  3. Monte Carlo simulation: Its status and future

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Murtha, J.A.

    1997-04-01

    Monte Carlo simulation is a statistics-based analysis tool that yields probability-vs.-value relationships for key parameters, including oil and gas reserves, capital exposure, and various economic yardsticks, such as net present value (NPV) and return on investment (ROI). Monte Carlo simulation is a part of risk analysis and is sometimes performed in conjunction with or as an alternative to decision [tree] analysis. The objectives are (1) to define Monte Carlo simulation in a more general context of risk and decision analysis; (2) to provide some specific applications, which can be interrelated; (3) to respond to some of the criticisms; (4) tomore » offer some cautions about abuses of the method and recommend how to avoid the pitfalls; and (5) to predict what the future has in store.« less

  4. Reconnaissance assessment of erosion and sedimentation in the Canada de los Alamos basin, Los Angeles and Ventura Counties, California

    USGS Publications Warehouse

    Knott, J.M.

    1980-01-01

    An assessment of present erosion and sedimentation conditions in the Ca?ada de los Alamos basin was made to aid in estimating the impact of off-road-vehicle use on the sediment yield of the basin. Impacts of off-road vehicles were evaluated by reconnaissance techniques and by comparing the study area with other offroad-vehicle sites in California. Major-storm sediment yields for the basin were estimated using empirical equations developed for the Transverse Ranges and measurements of gully erosion in a representative off-road-vehicle basin. Normal major-storm yields of 73,200 cubic yards would have to be increased to about 98,000 cubic yards to account for the existing level of accelerated erosion caused by off-road vehicles. Long-term sediment yield of the Ca?ada de los Alamos basin upstream from its confluence with Gorman Creek, under present conditions of off-road-vehicle use, is approximately 420 cubic yards per square mile per year--a rate that is considerably lower than a previous estimate of 1,270 cubic yards per square mile per year for the total catchment area above Pyramid Lake.

  5. Hybrid Monte Carlo/deterministic methods for radiation shielding problems

    NASA Astrophysics Data System (ADS)

    Becker, Troy L.

    For the past few decades, the most common type of deep-penetration (shielding) problem simulated using Monte Carlo methods has been the source-detector problem, in which a response is calculated at a single location in space. Traditionally, the nonanalog Monte Carlo methods used to solve these problems have required significant user input to generate and sufficiently optimize the biasing parameters necessary to obtain a statistically reliable solution. It has been demonstrated that this laborious task can be replaced by automated processes that rely on a deterministic adjoint solution to set the biasing parameters---the so-called hybrid methods. The increase in computational power over recent years has also led to interest in obtaining the solution in a region of space much larger than a point detector. In this thesis, we propose two methods for solving problems ranging from source-detector problems to more global calculations---weight windows and the Transform approach. These techniques employ sonic of the same biasing elements that have been used previously; however, the fundamental difference is that here the biasing techniques are used as elements of a comprehensive tool set to distribute Monte Carlo particles in a user-specified way. The weight window achieves the user-specified Monte Carlo particle distribution by imposing a particular weight window on the system, without altering the particle physics. The Transform approach introduces a transform into the neutron transport equation, which results in a complete modification of the particle physics to produce the user-specified Monte Carlo distribution. These methods are tested in a three-dimensional multigroup Monte Carlo code. For a basic shielding problem and a more realistic one, these methods adequately solved source-detector problems and more global calculations. Furthermore, they confirmed that theoretical Monte Carlo particle distributions correspond to the simulated ones, implying that these methods

  6. Overview and Status of the Los Alamos PSR Injection Upgrade Project

    NASA Astrophysics Data System (ADS)

    Fitzgerald, D. H.; Ahn, H.; Blind, B.; Borden, M. J.; Macek, R. J.; Neri, F.; Rose, C. R.; Thiessen, H. A.; Wilkinson, C. A.; Zumbro, M. V.

    1997-05-01

    An upgrade is in progress to the Los Alamos Proton Storage Ring (PSR) to allow direct injection of the H^- beam into the ring and provide a beam bump system to move the circulating beam off the stripper foil. The primary benefits of this upgrade are matching the transverse phase space of the injected beam to the PSR acceptance and reduction of foil hits by the circulating beam by a factor of ten. Foil thickness is optimized to minimize the combination of circulating-beam losses plus losses due to excited H^0 states produced at injection. An overall factor of four reduction in losses is expected. The project comprises extensive modifications of the injection line, the injection section of the ring, and the waste-beam transport line. We will discuss the goals of the project, present an overview of the technical design, and describe the status of the implementation plan.

  7. Electrical Engineering in Los Alamos Neutron Science Center Accelerator

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Silva, Michael James

    The field of electrical engineering plays a significant role in particle accelerator design and operations. Los Alamos National Laboratories LANSCE facility utilizes the electrical energy concepts of power distribution, plasma generation, radio frequency energy, electrostatic acceleration, signals and diagnostics. The culmination of these fields produces a machine of incredible potential with uses such as isotope production, neutron spallation, neutron imaging and particle analysis. The key isotope produced in LANSCE isotope production facility is Strontium-82 which is utilized for medical uses such as cancer treatment and positron emission tomography also known as PET scans. Neutron spallation is one of the verymore » few methods used to produce neutrons for scientific research the other methods are natural decay of transuranic elements from nuclear reactors. Accelerator produce neutrons by accelerating charged particles into neutron dense elements such as tungsten imparting a neutral particle with kinetic energy, this has the benefit of producing a large number of neutrons as well as minimizing the waste generated. Utilizing the accelerator scientist can gain an understanding of how various particles behave and interact with matter to better understand the natural laws of physics and the universe around us.« less

  8. The performance of the upgraded Los Alamos Neutron Source

    NASA Astrophysics Data System (ADS)

    Ito, Takeyasu; LANL UCN Source Collaboration

    2017-09-01

    Los Alamos National Laboratory has been operating an ultracold (UCN) source based on a solid deuterium (SD2) UCN converter driven by spallation neutrons for over 10 years. It has recently been successfully upgraded, by replacing the cryostat that contains the cold neutron moderator, SD2 volume, and vertical UCN guide. The horizontal UCN guide that transports UCN out of the radiation shield was also replaced. The new design reflects lessons learned from the 10+ year long operation of the previous version of the UCN source and is optimized to maximize the cold neutron flux at the SD2 volume, featuring a close coupled cold neutron moderator, and maximize the transport of the UCN to experiments. During the commissioning of the upgraded UCN source, data were collected to measure its performance, including cold neutron spectra as a function of the cold moderator temperature, and the UCN density in a vessel outside the source. In this talk, after a brief overview of the design of the upgraded source, the results of the performance tests and comparison to prediction will be presented. This work was funded by LANL LDRD.

  9. Wet-based glaciation in Phlegra Montes, Mars.

    NASA Astrophysics Data System (ADS)

    Gallagher, Colman; Balme, Matt

    2016-04-01

    Eskers are sinuous landforms composed of sediments deposited from meltwaters in ice-contact glacial conduits. This presentation describes the first definitive identification of eskers on Mars still physically linked with their parent system (1), a Late Amazonian-age glacier (~150 Ma) in Phlegra Montes. Previously described Amazonian-age glaciers on Mars are generally considered to have been dry based, having moved by creep in the absence of subglacial water required for sliding, but our observations indicate significant sub-glacial meltwater routing. The confinement of the Phlegra Montes glacial system to a regionally extensive graben is evidence that the esker formed due to sub-glacial melting in response to an elevated, but spatially restricted, geothermal heat flux rather than climate-induced warming. Now, however, new observations reveal the presence of many assemblages of glacial abrasion forms and associated channels that could be evidence of more widespread wet-based glaciation in Phlegra Montes, including the collapse of several distinct ice domes. This landform assemblage has not been described in other glaciated, mid-latitude regions of the martian northern hemisphere. Moreover, Phlegra Montes are flanked by lowlands displaying evidence of extensive volcanism, including contact between plains lava and piedmont glacial ice. These observations provide a rationale for investigating non-climatic forcing of glacial melting and associated landscape development on Mars, and can build on insights from Earth into the importance of geothermally-induced destabilisation of glaciers as a key amplifier of climate change. (1) Gallagher, C. and Balme, M. (2015). Eskers in a complete, wet-based glacial system in the Phlegra Montes region, Mars, Earth and Planetary Science Letters, 431, 96-109.

  10. 2016 Results for Avian Monitoring at the TA-36 Minie Site, TA-39 Point 6, and TA-16 Burn Ground at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hathcock, Charles Dean; Thompson, Brent E.; Berryhill, Jesse Tobias

    Los Alamos National Security, LLC (LANS) biologists in the Environmental Compliance and Protection Division at Los Alamos National Laboratory (LANL) initiated a multi-year program in 2013 to monitor avifauna at two open detonation sites and one open burn site on LANL property. Monitoring results from these efforts are compared among years and with avifauna monitoring conducted at other areas across LANL. The objectives of this study are to determine whether LANL firing site operations impact bird abundance or diversity. LANS biologists completed the fourth year of this effort in 2016. The overall results from 2016 continue to indicate that operationsmore » are not negatively affecting bird populations. Data suggest that community structure may be changing at some sites and this trend will continue to be monitored.« less

  11. A Primer in Monte Carlo Integration Using Mathcad

    ERIC Educational Resources Information Center

    Hoyer, Chad E.; Kegerreis, Jeb S.

    2013-01-01

    The essentials of Monte Carlo integration are presented for use in an upper-level physical chemistry setting. A Mathcad document that aids in the dissemination and utilization of this information is described and is available in the Supporting Information. A brief outline of Monte Carlo integration is given, along with ideas and pedagogy for…

  12. Lattice modeling and application of independent component analysis to high power, long bunch beams in the Los Alamos Proton Storage Ring

    NASA Astrophysics Data System (ADS)

    Kolski, Jeffrey

    The linear lattice properties of the Proton Storage Ring (PSR) at the Los Alamos Neutron Science Center (LANSCE) in Los Alamos, NM were measured and applied to determine a better linear accelerator model. We found that the initial model was deficient in predicting the vertical focusing strength. The additional vertical focusing was located through fundamental understanding of experiment and statistically rigorous analysis. An improved model was constructed and compared against the initial model and measurement at operation set points and set points far away from nominal and was shown to indeed be an enhanced model. Independent component analysis (ICA) is a tool for data mining in many fields of science. Traditionally, ICA is applied to turn-by-turn beam position data as a means to measure the lattice functions of the real machine. Due to the diagnostic setup for the PSR, this method is not applicable. A new application method for ICA is derived, ICA applied along the length of the bunch. The ICA modes represent motions within the beam pulse. Several of the dominate ICA modes are experimentally identified.

  13. Stormwater Pollution Prevention Plan for the TA-03-22 Power and Steam Plant, Los Alamos National Laboratory, Revision 3, January 2018

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Burgin, Jillian Elizabeth

    This Storm Water Pollution Prevention Plan (SWPPP) was developed in accordance with the provisions of the Clean Water Act (33 U.S.C. §§1251 et seq., as amended), and the Multi-Sector General Permit for Storm Water Discharges Associated with Industrial Activity (U.S. EPA, June 2015) issued by the U.S. Environmental Protection Agency (EPA) for the National Pollutant Discharge Elimination System (NPDES) and using the industry specific permit requirements for Sector O-Steam Electric Generating Facilities as a guide. This SWPPP applies to discharges of stormwater from the operational areas of the TA-03-22 Power and Steam Plant at Los Alamos National Laboratory. Los Alamosmore » National Laboratory (also referred to as LANL or the “Laboratory”) is owned by the Department of Energy (DOE), and is operated by Los Alamos National Security, LLC (LANS). Throughout this document, the term “facility” refers to the TA-03-22 Power and Steam Plant and associated areas. The current permit expires at midnight on June 4, 2020.« less

  14. Baseball Monte Carlo Style.

    ERIC Educational Resources Information Center

    Houser, Larry L.

    1981-01-01

    Monte Carlo methods are used to simulate activities in baseball such as a team's "hot streak" and a hitter's "batting slump." Student participation in such simulations is viewed as a useful method of giving pupils a better understanding of the probability concepts involved. (MP)

  15. The Los Alamos Seismic Network (LASN): Improved Network Instrumentation, Local Earthquake Catalog Updates, and Peculiar Types of Data

    NASA Astrophysics Data System (ADS)

    Roberts, P. M.; Ten Cate, J. A.; House, L. S.; Greene, M. K.; Morton, E.; Kelley, R. E.

    2013-12-01

    The Los Alamos Seismic Network (LASN) has operated for 41 years, and provided the data to locate more than 2,500 earthquakes in north-central New Mexico. The network was installed for seismic verification research, as well as to monitor and locate earthquakes near Los Alamos National Laboratory (LANL). LASN stations are the only monitoring stations in New Mexico north of Albuquerque. The original network once included 22 stations in northern Mew Mexico. With limited funding in the early 1980's, the network was downsized to 7 stations within an area of about 15 km (N-S) by 15 km (E-W), centered on Los Alamos. Over the last four years, eight additional stations have been installed, which have considerably expanded the spatial coverage of the network. Currently, 7 stations have broadband, three-component seismometers with digital telemetry, and the remaining 8 have traditional 1 Hz short-period seismometers with either analog telemetry or on-site digital recording. A vertical array of accelerometers was also installed in a wellbore on LANL property. This borehole array has 3-component digital strong-motion sensors. Recently we began upgrading the local strong-motion accelerometer (SMA) network as well, with the addition of high-resolution digitizers and high-sensitivity force-balance accelerometers (FBA). We will present an updated description of the current LASN station, instrumentation and telemetry configurations, as well as the data acquisition and event-detection software structure used to record events in Earthworm. Although more than 2,000 earthquakes were detected and located in north-central New Mexico during the first 11 years of LASN's operation (1973 to 1984), currently only 1-2 earthquakes per month are detected and located within about 150 km of Los Alamos. Over 850 of these nearby earthquakes have been located from 1973 to present. We recently updated the LASN earthquake catalog for north-central New Mexico up through 2012 and most of 2013. Locations

  16. The Monte Carlo Method. Popular Lectures in Mathematics.

    ERIC Educational Resources Information Center

    Sobol', I. M.

    The Monte Carlo Method is a method of approximately solving mathematical and physical problems by the simulation of random quantities. The principal goal of this booklet is to suggest to specialists in all areas that they will encounter problems which can be solved by the Monte Carlo Method. Part I of the booklet discusses the simulation of random…

  17. Population array and agricultural data arrays for the Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Jacobson, K.W.; Duffy, S.; Kowalewsky, K.

    1998-07-01

    To quantify or estimate the environmental and radiological impacts from man-made sources of radioactive effluents, certain dose assessment procedures were developed by various government and regulatory agencies. Some of these procedures encourage the use of computer simulations (models) to calculate air dispersion, environmental transport, and subsequent human exposure to radioactivity. Such assessment procedures are frequently used to demonstrate compliance with Department of Energy (DOE) and US Environmental Protection Agency (USEPA) regulations. Knowledge of the density and distribution of the population surrounding a source is an essential component in assessing the impacts from radioactive effluents. Also, as an aid to calculatingmore » the dose to a given population, agricultural data relevant to the dose assessment procedure (or computer model) are often required. This report provides such population and agricultural data for the area surrounding Los Alamos National Laboratory.« less

  18. User Guide for the Plotting Software for the Los Alamos National Laboratory Nuclear Weapons Analysis Tools Version 2.0

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cleland, Timothy James

    The Los Alamos National Laboratory Plotting Software for the Nuclear Weapons Analysis Tools is a Java™ application based upon the open source library JFreeChart. The software provides a capability for plotting data on graphs with a rich variety of display options while allowing the viewer interaction via graph manipulation and scaling to best view the data. The graph types include XY plots, Date XY plots, Bar plots and Histogram plots.

  19. New Approaches and Applications for Monte Carlo Perturbation Theory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Aufiero, Manuele; Bidaud, Adrien; Kotlyar, Dan

    2017-02-01

    This paper presents some of the recent and new advancements in the extension of Monte Carlo Perturbation Theory methodologies and application. In particular, the discussed problems involve Brunup calculation, perturbation calculation based on continuous energy functions, and Monte Carlo Perturbation Theory in loosely coupled systems.

  20. Cell-veto Monte Carlo algorithm for long-range systems.

    PubMed

    Kapfer, Sebastian C; Krauth, Werner

    2016-09-01

    We present a rigorous efficient event-chain Monte Carlo algorithm for long-range interacting particle systems. Using a cell-veto scheme within the factorized Metropolis algorithm, we compute each single-particle move with a fixed number of operations. For slowly decaying potentials such as Coulomb interactions, screening line charges allow us to take into account periodic boundary conditions. We discuss the performance of the cell-veto Monte Carlo algorithm for general inverse-power-law potentials, and illustrate how it provides a new outlook on one of the prominent bottlenecks in large-scale atomistic Monte Carlo simulations.

  1. Recommender engine for continuous-time quantum Monte Carlo methods

    NASA Astrophysics Data System (ADS)

    Huang, Li; Yang, Yi-feng; Wang, Lei

    2017-03-01

    Recommender systems play an essential role in the modern business world. They recommend favorable items such as books, movies, and search queries to users based on their past preferences. Applying similar ideas and techniques to Monte Carlo simulations of physical systems boosts their efficiency without sacrificing accuracy. Exploiting the quantum to classical mapping inherent in the continuous-time quantum Monte Carlo methods, we construct a classical molecular gas model to reproduce the quantum distributions. We then utilize powerful molecular simulation techniques to propose efficient quantum Monte Carlo updates. The recommender engine approach provides a general way to speed up the quantum impurity solvers.

  2. Nuclide Depletion Capabilities in the Shift Monte Carlo Code

    DOE PAGES

    Davidson, Gregory G.; Pandya, Tara M.; Johnson, Seth R.; ...

    2017-12-21

    A new depletion capability has been developed in the Exnihilo radiation transport code suite. This capability enables massively parallel domain-decomposed coupling between the Shift continuous-energy Monte Carlo solver and the nuclide depletion solvers in ORIGEN to perform high-performance Monte Carlo depletion calculations. This paper describes this new depletion capability and discusses its various features, including a multi-level parallel decomposition, high-order transport-depletion coupling, and energy-integrated power renormalization. Several test problems are presented to validate the new capability against other Monte Carlo depletion codes, and the parallel performance of the new capability is analyzed.

  3. Rapid Monte Carlo Simulation of Gravitational Wave Galaxies

    NASA Astrophysics Data System (ADS)

    Breivik, Katelyn; Larson, Shane L.

    2015-01-01

    With the detection of gravitational waves on the horizon, astrophysical catalogs produced by gravitational wave observatories can be used to characterize the populations of sources and validate different galactic population models. Efforts to simulate gravitational wave catalogs and source populations generally focus on population synthesis models that require extensive time and computational power to produce a single simulated galaxy. Monte Carlo simulations of gravitational wave source populations can also be used to generate observation catalogs from the gravitational wave source population. Monte Carlo simulations have the advantes of flexibility and speed, enabling rapid galactic realizations as a function of galactic binary parameters with less time and compuational resources required. We present a Monte Carlo method for rapid galactic simulations of gravitational wave binary populations.

  4. French Alps, Mont Blanc, French/Italian Border

    NASA Image and Video Library

    1992-04-02

    In this southeast looking view, Mont Blanc, on the French/Italian border, (48.0N, 4.5E) the highest mountain peak in all of Europe, is just below and right of center (below the end of the prominent valley of the Aosta River, in the center of the photo. The rivers flow out of the Alps into Italy toward Turin. Chamonix, the famous resort town and center of Alpine mountain climbing, lies in the valley just below Mont Blanc.

  5. Explosively driven two-shockwave tools with application to ejecta formation at the Los Alamos National Laboratory Proton Radiography Facility

    NASA Astrophysics Data System (ADS)

    Buttler, William

    2013-06-01

    We present the development of an explosively driven physics tool to generate two mostly uniaxial shockwaves. The tool is being used to extend single shockwave ejecta models to a subsequent shockwave event separated by a time interval on the order of a few microseconds. We explore the possibility of varying the amplitude of both the first and second shockwaves, and we apply the tool in experimental geometries on Sn with a surface roughness of Ra = 0 . 8 μ m. We then evaluate the tool further at the Los Alamos National Laboratory Proton Radiography (pRad) Facility in an application to Sn with larger scale perturbations of wavelength 550 μ m, and various amplitudes that gave wave-number amplitude products of η0 2 π / λ = { 3 / 4 , 1 / 2 , 1 / 4 , 1 / 8 } , where the perturbation amplitude is η0, and the wave-number k = 2 π / λ . The pRad data and velocimetry imply it should be possible to develop a second shock ejecta model based on unstable Richtmyer-Meshkov physics. In collaboration with David Oro, Fesseha Mariam, Alexander Saunders, Malcolm Andrews, Frank Cherne, James Hammerberg. Robert Hixson, Christopher Morris, Russell Olson, Dean Preston, Joseph Stone, Dale Tupa, and Wendy Vogan-McNeil, Los Alamos National Laboratory,

  6. Monte Carlo capabilities of the SCALE code system

    DOE PAGES

    Rearden, Bradley T.; Petrie, Jr., Lester M.; Peplow, Douglas E.; ...

    2014-09-12

    SCALE is a broadly used suite of tools for nuclear systems modeling and simulation that provides comprehensive, verified and validated, user-friendly capabilities for criticality safety, reactor physics, radiation shielding, and sensitivity and uncertainty analysis. For more than 30 years, regulators, licensees, and research institutions around the world have used SCALE for nuclear safety analysis and design. SCALE provides a “plug-and-play” framework that includes three deterministic and three Monte Carlo radiation transport solvers that can be selected based on the desired solution, including hybrid deterministic/Monte Carlo simulations. SCALE includes the latest nuclear data libraries for continuous-energy and multigroup radiation transport asmore » well as activation, depletion, and decay calculations. SCALE’s graphical user interfaces assist with accurate system modeling, visualization, and convenient access to desired results. SCALE 6.2 will provide several new capabilities and significant improvements in many existing features, especially with expanded continuous-energy Monte Carlo capabilities for criticality safety, shielding, depletion, and sensitivity and uncertainty analysis. Finally, an overview of the Monte Carlo capabilities of SCALE is provided here, with emphasis on new features for SCALE 6.2.« less

  7. Geodesic Monte Carlo on Embedded Manifolds

    PubMed Central

    Byrne, Simon; Girolami, Mark

    2013-01-01

    Markov chain Monte Carlo methods explicitly defined on the manifold of probability distributions have recently been established. These methods are constructed from diffusions across the manifold and the solution of the equations describing geodesic flows in the Hamilton–Jacobi representation. This paper takes the differential geometric basis of Markov chain Monte Carlo further by considering methods to simulate from probability distributions that themselves are defined on a manifold, with common examples being classes of distributions describing directional statistics. Proposal mechanisms are developed based on the geodesic flows over the manifolds of support for the distributions, and illustrative examples are provided for the hypersphere and Stiefel manifold of orthonormal matrices. PMID:25309024

  8. The Rational Hybrid Monte Carlo algorithm

    NASA Astrophysics Data System (ADS)

    Clark, Michael

    2006-12-01

    The past few years have seen considerable progress in algorithmic development for the generation of gauge fields including the effects of dynamical fermions. The Rational Hybrid Monte Carlo (RHMC) algorithm, where Hybrid Monte Carlo is performed using a rational approximation in place the usual inverse quark matrix kernel is one of these developments. This algorithm has been found to be extremely beneficial in many areas of lattice QCD (chiral fermions, finite temperature, Wilson fermions etc.). We review the algorithm and some of these benefits, and we compare against other recent algorithm developements. We conclude with an update of the Berlin wall plot comparing costs of all popular fermion formulations.

  9. Storm Water Quality in Los Alamos Canyon following the Cerro Grande Fire

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    M. Johansen; B. Enz; B. Gallaher

    In May 2000, the Cerro Grande Fire burned about 7400 acres of forest on the Los Alamos National Laboratory (LANL) and about 10,000 acres in watersheds above LANL on Santa Fe National Forest lands. The resulting burned landscapes raised concerns of increased storm water runoff and transport of contaminants by runoff in the canyons traversing LANL. On June 2 and 3, 2000, rain fell in the Los Alamos Canyon watershed generating storm water runoff in the canyon bottom. This event was important in that it was the first significant runoff on LANL following the fire and occurred in a canyonmore » containing known legacy waste sites. Samples from this runoff were analyzed for radionuclide, metal, inorganic, and organic constituents. Results show radionuclide concentrations at or below previous (pre-fire) maximum levels at locations on LANL and downstream. However, greater concentrations of some fallout-associated radionuclides (cesium-137 and strontium-90) were seen arriving on LANL from upstream areas compared to pre-fire conditions. Tests indicate most of the radionuclides in the samples were bound to sediments, not dissolved in water. Most radionuclide concentrations in sediments were below LANL Screening Action Levels, with cesium-137 and strontium-90 as exceptions. Most radionuclide concentrations in samples taken at LANL's downstream boundary were greater than those taken upstream, indicating the presence of contributing sources on LANL. For comparison purposes, doses were calculated on a mrem per liter of unfiltered water basis for 11 radionuclides commonly associated with atmospheric fallout and with LANL operations. The maximum dose was 0.094 mrem per liter unfiltered water and was largely associated with plutonium-239/240. In contrast, all filtered samples had total doses less than 0.001 mrem per liter. Compared to past data, potential doses were not increased by the fire during this initial runoff event. Of the 25 metals tested for, seven were above pre

  10. University of New Mexico-Los Alamos National Laboratory Program in Volcanology

    NASA Astrophysics Data System (ADS)

    Goff, F.; Fischer, T.; Baldridge, W.; Wohletz, K.; Smith, G.; Heiken, G.; Valentine, G.; Elston, W.

    2002-05-01

    The UNM-LANL Program in Volcanology was a vision of Wolf Elston in the late 1980s. Finally established in mid-1992, the program takes advantage of the extensive volcanic record preserved in northern New Mexico, and of the unique expertise and exceptional research facilities existing at the two institutions. Courses are directed toward upper division and graduate level students. The Los Alamos participants are adjunct professors and they take an active role in creating courses, advising thesis candidates, and providing research support. The curriculum is flexible but has a core upper division class in Physical Volcanology. Other classes offered in various years have included Volcanology and Human Affairs; Magmatic and Geothermal Systems; Tectonics and Magma Generation; Volcanoes of North America; Instrumentation for Volcanology; and Advanced Igneous Petrology. Perhaps the most renowned class in the program is the Volcanology Summer Field Course offered in even numbered years. This 3.5-week class is based in the Jemez Mountains volcanic field, which contains the famous Valles caldera (1.2 Ma to 50 ka). All types of calc-alkaline to alkalic domes, flows, tuffs, and intrusions, plus derivative sediments, mineralized zones, and thermal fluids are available for instructional purposes. Students are required to complete nine rigorous field exercises starting with basic instruction in pyroclastic fall, flow, and surge, then progressing towards hydrothermally altered, intracaldera resurgent dome and moat deposits in an active hot spring and fumarole system. The class is open to graduate students, advanced undergraduates, and private sector employees with special needs. Enrollment is competitive with limited financial support and limited space for 17 students. Evening lectures, study time, lodging, and meals are provided at the UNM-owned Young's Ranch built in the 1920s, nestled in a canyon flanked by orange cliffs of Bandelier Tuff. About 120 students from 12 countries have

  11. RNA folding kinetics using Monte Carlo and Gillespie algorithms.

    PubMed

    Clote, Peter; Bayegan, Amir H

    2018-04-01

    RNA secondary structure folding kinetics is known to be important for the biological function of certain processes, such as the hok/sok system in E. coli. Although linear algebra provides an exact computational solution of secondary structure folding kinetics with respect to the Turner energy model for tiny ([Formula: see text]20 nt) RNA sequences, the folding kinetics for larger sequences can only be approximated by binning structures into macrostates in a coarse-grained model, or by repeatedly simulating secondary structure folding with either the Monte Carlo algorithm or the Gillespie algorithm. Here we investigate the relation between the Monte Carlo algorithm and the Gillespie algorithm. We prove that asymptotically, the expected time for a K-step trajectory of the Monte Carlo algorithm is equal to [Formula: see text] times that of the Gillespie algorithm, where [Formula: see text] denotes the Boltzmann expected network degree. If the network is regular (i.e. every node has the same degree), then the mean first passage time (MFPT) computed by the Monte Carlo algorithm is equal to MFPT computed by the Gillespie algorithm multiplied by [Formula: see text]; however, this is not true for non-regular networks. In particular, RNA secondary structure folding kinetics, as computed by the Monte Carlo algorithm, is not equal to the folding kinetics, as computed by the Gillespie algorithm, although the mean first passage times are roughly correlated. Simulation software for RNA secondary structure folding according to the Monte Carlo and Gillespie algorithms is publicly available, as is our software to compute the expected degree of the network of secondary structures of a given RNA sequence-see http://bioinformatics.bc.edu/clote/RNAexpNumNbors .

  12. MONTE: the next generation of mission design and navigation software

    NASA Astrophysics Data System (ADS)

    Evans, Scott; Taber, William; Drain, Theodore; Smith, Jonathon; Wu, Hsi-Cheng; Guevara, Michelle; Sunseri, Richard; Evans, James

    2018-03-01

    The Mission analysis, Operations and Navigation Toolkit Environment (MONTE) (Sunseri et al. in NASA Tech Briefs 36(9), 2012) is an astrodynamic toolkit produced by the Mission Design and Navigation Software Group at the Jet Propulsion Laboratory. It provides a single integrated environment for all phases of deep space and Earth orbiting missions. Capabilities include: trajectory optimization and analysis, operational orbit determination, flight path control, and 2D/3D visualization. MONTE is presented to the user as an importable Python language module. This allows a simple but powerful user interface via CLUI or script. In addition, the Python interface allows MONTE to be used seamlessly with other canonical scientific programming tools such as SciPy, NumPy, and Matplotlib. MONTE is the prime operational orbit determination software for all JPL navigated missions.

  13. Wider pulsation instability regions for β Cephei and SPB stars calculated using new Los Alamos opacities

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Walczak, Przemysław; Fontes, Christopher John; Colgan, James Patrick

    Here, our goal is to test the newly developed OPLIB opacity tables from Los Alamos National Laboratory and check their influence on the pulsation properties of B-type stars. We calculated models using MESA and Dziembowski codes for stellar evolution and linear, nonadiabatic pulsations, respectively. We derived the instability domains of β Cephei and SPB-types for different opacity tables OPLIB, OP, and OPAL. As a result, the new OPLIB opacities have the highest Rosseland mean opacity coefficient near the so-called Z-bump. Therefore, the OPLIB instability domains are wider than in the case of OP and OPAL data.

  14. Wider pulsation instability regions for β Cephei and SPB stars calculated using new Los Alamos opacities

    DOE PAGES

    Walczak, Przemysław; Fontes, Christopher John; Colgan, James Patrick; ...

    2015-08-13

    Here, our goal is to test the newly developed OPLIB opacity tables from Los Alamos National Laboratory and check their influence on the pulsation properties of B-type stars. We calculated models using MESA and Dziembowski codes for stellar evolution and linear, nonadiabatic pulsations, respectively. We derived the instability domains of β Cephei and SPB-types for different opacity tables OPLIB, OP, and OPAL. As a result, the new OPLIB opacities have the highest Rosseland mean opacity coefficient near the so-called Z-bump. Therefore, the OPLIB instability domains are wider than in the case of OP and OPAL data.

  15. 76 FR 70977 - Proposed Collection; Comment Request

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-11-16

    ... Intelligence. DSS functions and responsibilities include the administration and implementation of the Defense... will be handled and marked ``For Official Use Only.'' Affected Public: Businesses, universities... will promote the implementation of the NISP. Contractors operating under DoD security cognizance are...

  16. 32 CFR 2004.22 - Operational Responsibilities [202(a)].

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ...; (3) Sign agreements with the Department of Defense as the Executive Agent for industrial security services; and, (4) Ensure applicable department and agency personnel having NISP implementation... of the Director of National Intelligence (ODNI) for Sensitive Compartmented Information, and DOE for...

  17. PyMercury: Interactive Python for the Mercury Monte Carlo Particle Transport Code

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Iandola, F N; O'Brien, M J; Procassini, R J

    2010-11-29

    Monte Carlo particle transport applications are often written in low-level languages (C/C++) for optimal performance on clusters and supercomputers. However, this development approach often sacrifices straightforward usability and testing in the interest of fast application performance. To improve usability, some high-performance computing applications employ mixed-language programming with high-level and low-level languages. In this study, we consider the benefits of incorporating an interactive Python interface into a Monte Carlo application. With PyMercury, a new Python extension to the Mercury general-purpose Monte Carlo particle transport code, we improve application usability without diminishing performance. In two case studies, we illustrate how PyMercury improvesmore » usability and simplifies testing and validation in a Monte Carlo application. In short, PyMercury demonstrates the value of interactive Python for Monte Carlo particle transport applications. In the future, we expect interactive Python to play an increasingly significant role in Monte Carlo usage and testing.« less

  18. Annual Report for Los Alamos National Laboratory Technical Area 54, Area G Disposal Facility - Fiscal Year 2016

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Birdsell, Kay Hanson; Stauffer, Philip H.; Atchley, Adam Lee

    As a condition to the disposal authorization statement issued to Los Alamos National Laboratory (LANL or the Laboratory) on March 17, 2010, a comprehensive performance assessment and composite analysis (PA/CA) maintenance program must be implemented for the Technical Area 54, Area G disposal facility. Annual determinations of the adequacy of the PA/CA are to be conducted under the maintenance program to ensure that the conclusions reached by those analyses continue to be valid. This report summarizes the results of the fiscal year (FY) 2016 annual review for Area G.

  19. Space-charge-sustained microbunch structure in the Los Alamos Proton Storage Ring

    NASA Astrophysics Data System (ADS)

    Cousineau, S.; Danilov, V.; Holmes, J.; Macek, R.

    2004-09-01

    We present experimental data from the Los Alamos Proton Storage Ring (PSR) showing long-lived linac microbunch structure during beam storage with no rf bunching. Analysis of the experimental data and particle-in-cell simulations of the experiments indicate that space charge, coupled with energy spread effects, is responsible for the sustained microbunch structure. The simulated longitudinal phase space of the beam reveals a well-defined separatrix in the phase space between linac microbunches, with particles executing unbounded motion outside of the separatrix. We show that the longitudinal phase space of the beam was near steady state during the PSR experiments, such that the separatrix persisted for long periods of time. Our simulations indicate that the steady state is very sensitive to the experimental conditions. Finally, we solve the steady-state problem in an analytic, self-consistent fashion for a set of periodic longitudinal space-charge potentials.

  20. Monte Carlo algorithms for Brownian phylogenetic models.

    PubMed

    Horvilleur, Benjamin; Lartillot, Nicolas

    2014-11-01

    Brownian models have been introduced in phylogenetics for describing variation in substitution rates through time, with applications to molecular dating or to the comparative analysis of variation in substitution patterns among lineages. Thus far, however, the Monte Carlo implementations of these models have relied on crude approximations, in which the Brownian process is sampled only at the internal nodes of the phylogeny or at the midpoints along each branch, and the unknown trajectory between these sampled points is summarized by simple branchwise average substitution rates. A more accurate Monte Carlo approach is introduced, explicitly sampling a fine-grained discretization of the trajectory of the (potentially multivariate) Brownian process along the phylogeny. Generic Monte Carlo resampling algorithms are proposed for updating the Brownian paths along and across branches. Specific computational strategies are developed for efficient integration of the finite-time substitution probabilities across branches induced by the Brownian trajectory. The mixing properties and the computational complexity of the resulting Markov chain Monte Carlo sampler scale reasonably with the discretization level, allowing practical applications with up to a few hundred discretization points along the entire depth of the tree. The method can be generalized to other Markovian stochastic processes, making it possible to implement a wide range of time-dependent substitution models with well-controlled computational precision. The program is freely available at www.phylobayes.org. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  1. High-efficiency wavefunction updates for large scale Quantum Monte Carlo

    NASA Astrophysics Data System (ADS)

    Kent, Paul; McDaniel, Tyler; Li, Ying Wai; D'Azevedo, Ed

    Within ab intio Quantum Monte Carlo (QMC) simulations, the leading numerical cost for large systems is the computation of the values of the Slater determinants in the trial wavefunctions. The evaluation of each Monte Carlo move requires finding the determinant of a dense matrix, which is traditionally iteratively evaluated using a rank-1 Sherman-Morrison updating scheme to avoid repeated explicit calculation of the inverse. For calculations with thousands of electrons, this operation dominates the execution profile. We propose a novel rank- k delayed update scheme. This strategy enables probability evaluation for multiple successive Monte Carlo moves, with application of accepted moves to the matrices delayed until after a predetermined number of moves, k. Accepted events grouped in this manner are then applied to the matrices en bloc with enhanced arithmetic intensity and computational efficiency. This procedure does not change the underlying Monte Carlo sampling or the sampling efficiency. For large systems and algorithms such as diffusion Monte Carlo where the acceptance ratio is high, order of magnitude speedups can be obtained on both multi-core CPU and on GPUs, making this algorithm highly advantageous for current petascale and future exascale computations.

  2. A modified Monte Carlo model for the ionospheric heating rates

    NASA Technical Reports Server (NTRS)

    Mayr, H. G.; Fontheim, E. G.; Robertson, S. C.

    1972-01-01

    A Monte Carlo method is adopted as a basis for the derivation of the photoelectron heat input into the ionospheric plasma. This approach is modified in an attempt to minimize the computation time. The heat input distributions are computed for arbitrarily small source elements that are spaced at distances apart corresponding to the photoelectron dissipation range. By means of a nonlinear interpolation procedure their individual heating rate distributions are utilized to produce synthetic ones that fill the gaps between the Monte Carlo generated distributions. By varying these gaps and the corresponding number of Monte Carlo runs the accuracy of the results is tested to verify the validity of this procedure. It is concluded that this model can reduce the computation time by more than a factor of three, thus improving the feasibility of including Monte Carlo calculations in self-consistent ionosphere models.

  3. Implementation of Monte Carlo Dose calculation for CyberKnife treatment planning

    NASA Astrophysics Data System (ADS)

    Ma, C.-M.; Li, J. S.; Deng, J.; Fan, J.

    2008-02-01

    Accurate dose calculation is essential to advanced stereotactic radiosurgery (SRS) and stereotactic radiotherapy (SRT) especially for treatment planning involving heterogeneous patient anatomy. This paper describes the implementation of a fast Monte Carlo dose calculation algorithm in SRS/SRT treatment planning for the CyberKnife® SRS/SRT system. A superposition Monte Carlo algorithm is developed for this application. Photon mean free paths and interaction types for different materials and energies as well as the tracks of secondary electrons are pre-simulated using the MCSIM system. Photon interaction forcing and splitting are applied to the source photons in the patient calculation and the pre-simulated electron tracks are repeated with proper corrections based on the tissue density and electron stopping powers. Electron energy is deposited along the tracks and accumulated in the simulation geometry. Scattered and bremsstrahlung photons are transported, after applying the Russian roulette technique, in the same way as the primary photons. Dose calculations are compared with full Monte Carlo simulations performed using EGS4/MCSIM and the CyberKnife treatment planning system (TPS) for lung, head & neck and liver treatments. Comparisons with full Monte Carlo simulations show excellent agreement (within 0.5%). More than 10% differences in the target dose are found between Monte Carlo simulations and the CyberKnife TPS for SRS/SRT lung treatment while negligible differences are shown in head and neck and liver for the cases investigated. The calculation time using our superposition Monte Carlo algorithm is reduced up to 62 times (46 times on average for 10 typical clinical cases) compared to full Monte Carlo simulations. SRS/SRT dose distributions calculated by simple dose algorithms may be significantly overestimated for small lung target volumes, which can be improved by accurate Monte Carlo dose calculations.

  4. Induction of Micronuclei in Human Fibroblasts from the Los Alamos High Energy Neutron Beam

    NASA Technical Reports Server (NTRS)

    Cox, Bradley

    2009-01-01

    The space radiation field includes a broad spectrum of high energy neutrons. Interactions between these neutrons and a spacecraft, or other material, significantly contribute to the dose equivalent for astronauts. The 15 degree beam line in the Weapons Neutron Research beam at Los Alamos Nuclear Science Center generates a neutron spectrum relatively similar to that seen in space. Human foreskin fibroblast (AG1522) samples were irradiated behind 0 to 20 cm of water equivalent shielding. The cells were exposed to either a 0.05 or 0.2 Gy entrance dose. Following irradiation, micronuclei were counted to see how the water shield affects the beam and its damage to cell nuclei. Micronuclei induction was then compared with dose equivalent data provided from a tissue equivalent proportional counter.

  5. Early years of Computational Statistical Mechanics

    NASA Astrophysics Data System (ADS)

    Mareschal, Michel

    2018-05-01

    Evidence that a model of hard spheres exhibits a first-order solid-fluid phase transition was provided in the late fifties by two new numerical techniques known as Monte Carlo and Molecular Dynamics. This result can be considered as the starting point of computational statistical mechanics: at the time, it was a confirmation of a counter-intuitive (and controversial) theoretical prediction by J. Kirkwood. It necessitated an intensive collaboration between the Los Alamos team, with Bill Wood developing the Monte Carlo approach, and the Livermore group, where Berni Alder was inventing Molecular Dynamics. This article tells how it happened.

  6. Calculating Potential Energy Curves with Quantum Monte Carlo

    NASA Astrophysics Data System (ADS)

    Powell, Andrew D.; Dawes, Richard

    2014-06-01

    Quantum Monte Carlo (QMC) is a computational technique that can be applied to the electronic Schrödinger equation for molecules. QMC methods such as Variational Monte Carlo (VMC) and Diffusion Monte Carlo (DMC) have demonstrated the capability of capturing large fractions of the correlation energy, thus suggesting their possible use for high-accuracy quantum chemistry calculations. QMC methods scale particularly well with respect to parallelization making them an attractive consideration in anticipation of next-generation computing architectures which will involve massive parallelization with millions of cores. Due to the statistical nature of the approach, in contrast to standard quantum chemistry methods, uncertainties (error-bars) are associated with each calculated energy. This study focuses on the cost, feasibility and practical application of calculating potential energy curves for small molecules with QMC methods. Trial wave functions were constructed with the multi-configurational self-consistent field (MCSCF) method from GAMESS-US.[1] The CASINO Monte Carlo quantum chemistry package [2] was used for all of the DMC calculations. An overview of our progress in this direction will be given. References: M. W. Schmidt et al. J. Comput. Chem. 14, 1347 (1993). R. J. Needs et al. J. Phys.: Condensed Matter 22, 023201 (2010).

  7. Monte Carlo verification of radiotherapy treatments with CloudMC.

    PubMed

    Miras, Hector; Jiménez, Rubén; Perales, Álvaro; Terrón, José Antonio; Bertolet, Alejandro; Ortiz, Antonio; Macías, José

    2018-06-27

    A new implementation has been made on CloudMC, a cloud-based platform presented in a previous work, in order to provide services for radiotherapy treatment verification by means of Monte Carlo in a fast, easy and economical way. A description of the architecture of the application and the new developments implemented is presented together with the results of the tests carried out to validate its performance. CloudMC has been developed over Microsoft Azure cloud. It is based on a map/reduce implementation for Monte Carlo calculations distribution over a dynamic cluster of virtual machines in order to reduce calculation time. CloudMC has been updated with new methods to read and process the information related to radiotherapy treatment verification: CT image set, treatment plan, structures and dose distribution files in DICOM format. Some tests have been designed in order to determine, for the different tasks, the most suitable type of virtual machines from those available in Azure. Finally, the performance of Monte Carlo verification in CloudMC is studied through three real cases that involve different treatment techniques, linac models and Monte Carlo codes. Considering computational and economic factors, D1_v2 and G1 virtual machines were selected as the default type for the Worker Roles and the Reducer Role respectively. Calculation times up to 33 min and costs of 16 € were achieved for the verification cases presented when a statistical uncertainty below 2% (2σ) was required. The costs were reduced to 3-6 € when uncertainty requirements are relaxed to 4%. Advantages like high computational power, scalability, easy access and pay-per-usage model, make Monte Carlo cloud-based solutions, like the one presented in this work, an important step forward to solve the long-lived problem of truly introducing the Monte Carlo algorithms in the daily routine of the radiotherapy planning process.

  8. Libya Montes

    NASA Image and Video Library

    2002-11-23

    This image by NASA Mars Odyssey spacecraft shows the rugged cratered highland region of Libya Montes, which forms part of the rim of an ancient impact basin called Isidis. This region of the highlands is fairly dissected with valley networks. There is still debate within the scientific community as to how valley networks themselves form: surface runoff (rainfall/snowmelt) or headward erosion via groundwater sapping. The degree of dissection here in this region suggests surface runoff rather than groundwater sapping. Small dunes are also visible on the floors of some of these channels. http://photojournal.jpl.nasa.gov/catalog/PIA04008

  9. Exploring cluster Monte Carlo updates with Boltzmann machines

    NASA Astrophysics Data System (ADS)

    Wang, Lei

    2017-11-01

    Boltzmann machines are physics informed generative models with broad applications in machine learning. They model the probability distribution of an input data set with latent variables and generate new samples accordingly. Applying the Boltzmann machines back to physics, they are ideal recommender systems to accelerate the Monte Carlo simulation of physical systems due to their flexibility and effectiveness. More intriguingly, we show that the generative sampling of the Boltzmann machines can even give different cluster Monte Carlo algorithms. The latent representation of the Boltzmann machines can be designed to mediate complex interactions and identify clusters of the physical system. We demonstrate these findings with concrete examples of the classical Ising model with and without four-spin plaquette interactions. In the future, automatic searches in the algorithm space parametrized by Boltzmann machines may discover more innovative Monte Carlo updates.

  10. Development of the Los Alamos National Laboratory Cryogenic Pressure Loader

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ebey, Peter S.; Dole, James M.; Hoffer, James K.

    2003-05-15

    Targets for inertial fusion research and ignition at OMEGA, the National Ignition Facility, LMJ, and future facilities rely on beta-radiation-driven layering of spherical cryogenic DT ice layers contained within plastic or metal shells. Plastic shells will be permeation filled at room temperature then cooled to cryogenic temperatures before removal of the overpressure. The cryogenic pressure loader (CPL) was recently developed at Los Alamos National Laboratory as a testbed for studying the filling and layering of plastic target shells with DT. A technical description of the CPL is provided. The CPL consists of a cryostat, which contains a high-pressure permeation cell,more » and has optical access for investigating beta layering. The cryostat is housed within a tritium glovebox that contains manifolds for supplying high-pressure DT. The CPL shares some design elements with the cryogenic target handling system at the OMEGA facility to allow testing of tritium issues related to that system. The CPL has the capability to fill plastic targets by permeation to pressures up to 100 MPa and to cool them to 15 K. The CPL will accommodate a range of targets and may be modified for future experiments.« less

  11. Four decades of implicit Monte Carlo

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wollaber, Allan B.

    In 1971, Fleck and Cummings derived a system of equations to enable robust Monte Carlo simulations of time-dependent, thermal radiative transfer problems. Denoted the “Implicit Monte Carlo” (IMC) equations, their solution remains the de facto standard of high-fidelity radiative transfer simulations. Over the course of 44 years, their numerical properties have become better understood, and accuracy enhancements, novel acceleration methods, and variance reduction techniques have been suggested. In this review, we rederive the IMC equations—explicitly highlighting assumptions as they are made—and outfit the equations with a Monte Carlo interpretation. We put the IMC equations in context with other approximate formsmore » of the radiative transfer equations and present a new demonstration of their equivalence to another well-used linearization solved with deterministic transport methods for frequency-independent problems. We discuss physical and numerical limitations of the IMC equations for asymptotically small time steps, stability characteristics and the potential of maximum principle violations for large time steps, and solution behaviors in an asymptotically thick diffusive limit. We provide a new stability analysis for opacities with general monomial dependence on temperature. Here, we consider spatial accuracy limitations of the IMC equations and discussion acceleration and variance reduction techniques.« less

  12. jTracker and Monte Carlo Comparison

    NASA Astrophysics Data System (ADS)

    Selensky, Lauren; SeaQuest/E906 Collaboration

    2015-10-01

    SeaQuest is designed to observe the characteristics and behavior of `sea-quarks' in a proton by reconstructing them from the subatomic particles produced in a collision. The 120 GeV beam from the main injector collides with a fixed target and then passes through a series of detectors which records information about the particles produced in the collision. However, this data becomes meaningful only after it has been processed, stored, analyzed, and interpreted. Several programs are involved in this process. jTracker (sqerp) reads wire or hodoscope hits and reconstructs the tracks of potential dimuon pairs from a run, and Geant4 Monte Carlo simulates dimuon production and background noise from the beam. During track reconstruction, an event must meet the criteria set by the tracker to be considered a viable dimuon pair; this ensures that relevant data is retained. As a check, a comparison between a new version of jTracker and Monte Carlo was made in order to see how accurately jTracker could reconstruct the events created by Monte Carlo. In this presentation, the results of the inquest and their potential effects on the programming will be shown. This work is supported by U.S. DOE MENP Grant DE-FG02-03ER41243.

  13. Four decades of implicit Monte Carlo

    DOE PAGES

    Wollaber, Allan B.

    2016-02-23

    In 1971, Fleck and Cummings derived a system of equations to enable robust Monte Carlo simulations of time-dependent, thermal radiative transfer problems. Denoted the “Implicit Monte Carlo” (IMC) equations, their solution remains the de facto standard of high-fidelity radiative transfer simulations. Over the course of 44 years, their numerical properties have become better understood, and accuracy enhancements, novel acceleration methods, and variance reduction techniques have been suggested. In this review, we rederive the IMC equations—explicitly highlighting assumptions as they are made—and outfit the equations with a Monte Carlo interpretation. We put the IMC equations in context with other approximate formsmore » of the radiative transfer equations and present a new demonstration of their equivalence to another well-used linearization solved with deterministic transport methods for frequency-independent problems. We discuss physical and numerical limitations of the IMC equations for asymptotically small time steps, stability characteristics and the potential of maximum principle violations for large time steps, and solution behaviors in an asymptotically thick diffusive limit. We provide a new stability analysis for opacities with general monomial dependence on temperature. Here, we consider spatial accuracy limitations of the IMC equations and discussion acceleration and variance reduction techniques.« less

  14. 48 CFR 4.402 - General.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ..., January 6, 1993 (58 FR 3479, January 8, 1993), entitled “National Industrial Security Program” (NISP... National Industrial Security Program Operating Manual (NISPOM) incorporates the requirements of these... Central Intelligence, is responsible for issuance and maintenance of this Manual. The following DOD...

  15. Monte Carlo Particle Lists: MCPL

    NASA Astrophysics Data System (ADS)

    Kittelmann, T.; Klinkby, E.; Knudsen, E. B.; Willendrup, P.; Cai, X. X.; Kanaki, K.

    2017-09-01

    A binary format with lists of particle state information, for interchanging particles between various Monte Carlo simulation applications, is presented. Portable C code for file manipulation is made available to the scientific community, along with converters and plugins for several popular simulation packages.

  16. Calculations of skyshine from an intense portable electron linac

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Estes, G.P.; Hughes, H.G.; Fry, D.A.

    1994-12-31

    The MCNP Monte carlo code has been used at Los Alamos to calculate skyshine and terrain albedo efects from an intense portable electron linear accelerator that is to be used by the Russian Federation to radiograph nuclear weapons that may have been damaged by accidents. Relative dose rate profiles have been calculated. The design of the accelerator, along with a diagram, is presented.

  17. NOTE: Monte Carlo evaluation of kerma in an HDR brachytherapy bunker

    NASA Astrophysics Data System (ADS)

    Pérez-Calatayud, J.; Granero, D.; Ballester, F.; Casal, E.; Crispin, V.; Puchades, V.; León, A.; Verdú, G.

    2004-12-01

    In recent years, the use of high dose rate (HDR) after-loader machines has greatly increased due to the shift from traditional Cs-137/Ir-192 low dose rate (LDR) to HDR brachytherapy. The method used to calculate the required concrete and, where appropriate, lead shielding in the door is based on analytical methods provided by documents published by the ICRP, the IAEA and the NCRP. The purpose of this study is to perform a more realistic kerma evaluation at the entrance maze door of an HDR bunker using the Monte Carlo code GEANT4. The Monte Carlo results were validated experimentally. The spectrum at the maze entrance door, obtained with Monte Carlo, has an average energy of about 110 keV, maintaining a similar value along the length of the maze. The comparison of results from the aforementioned values with the Monte Carlo ones shows that results obtained using the albedo coefficient from the ICRP document more closely match those given by the Monte Carlo method, although the maximum value given by MC calculations is 30% greater.

  18. Monte Carlo Calculations of Polarized Microwave Radiation Emerging from Cloud Structures

    NASA Technical Reports Server (NTRS)

    Kummerow, Christian; Roberti, Laura

    1998-01-01

    The last decade has seen tremendous growth in cloud dynamical and microphysical models that are able to simulate storms and storm systems with very high spatial resolution, typically of the order of a few kilometers. The fairly realistic distributions of cloud and hydrometeor properties that these models generate has in turn led to a renewed interest in the three-dimensional microwave radiative transfer modeling needed to understand the effect of cloud and rainfall inhomogeneities upon microwave observations. Monte Carlo methods, and particularly backwards Monte Carlo methods have shown themselves to be very desirable due to the quick convergence of the solutions. Unfortunately, backwards Monte Carlo methods are not well suited to treat polarized radiation. This study reviews the existing Monte Carlo methods and presents a new polarized Monte Carlo radiative transfer code. The code is based on a forward scheme but uses aliasing techniques to keep the computational requirements equivalent to the backwards solution. Radiative transfer computations have been performed using a microphysical-dynamical cloud model and the results are presented together with the algorithm description.

  19. LA-UR-14-27684, Analysis of Wildland Fire Hazard to the TWF at Los Alamos National Labs

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gilbertson, Sarah

    Wildfires represent an Anticipated Natural Phenomena Hazard for LANL and the surrounding area. The TWF facility is located in a cleared area and is surrounded on three sides by roadway pavement. Therefore, direct propagation of flames to the facility is not considered the most credible means of ignition. Rather, fires started by airborne transport of burning brands constitute the most significant wildland fire threat to the TWF. The purpose of this document is to update LA-UR-13-24529, Airborne Projection of Burning Embers – Planning and Controls for Los Alamos National Laboratory Facilities, to be specific to the TWF site and operations.

  20. Status of the flora of the Los Alamos National Environmental Research Park: a historical perspective. Volume 2

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Foxx, T.S.; Tierney, G.D.

    1984-09-01

    Studies of the flora of the Los Alamos National Environmental Research Park (LA/NERP) are continued in Water and Pajarito Canyons and their extensions to natural boundaries outside the LA/NERP. Six plant communities and sixteen plant habitats are described for the plant communities and sixteen plant habitats are described for the study area. The status of endangered, threatened, and rare plant species in the study area is reviewed, and land-use history of the Pajarito Plateau is related to the levels of apparent anthropogenic disturbance in the study areas' six plant communities. 66 references, 20 figures.

  1. Annual Report for Los Alamos National Laboratory Technical Area 54, Area G Disposal Facility – Fiscal Year 2015

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    French, Sean B.; Stauffer, Philip H.; Birdsell, Kay H.

    As a condition to the disposal authorization statement issued to Los Alamos National Laboratory (LANL or the Laboratory) on March 17, 2010, a comprehensive performance assessment and composite analysis maintenance program must be implemented for the Technical Area 54, Area G disposal facility. Annual determinations of the adequacy of the performance assessment and composite analysis (PA/CA) are to be conducted under the maintenance program to ensure that the conclusions reached by those analyses continue to be valid. This report summarizes the results of the fiscal year (FY) 2015 annual review for Area G.

  2. TH-E-18A-01: Developments in Monte Carlo Methods for Medical Imaging

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Badal, A; Zbijewski, W; Bolch, W

    Monte Carlo simulation methods are widely used in medical physics research and are starting to be implemented in clinical applications such as radiation therapy planning systems. Monte Carlo simulations offer the capability to accurately estimate quantities of interest that are challenging to measure experimentally while taking into account the realistic anatomy of an individual patient. Traditionally, practical application of Monte Carlo simulation codes in diagnostic imaging was limited by the need for large computational resources or long execution times. However, recent advancements in high-performance computing hardware, combined with a new generation of Monte Carlo simulation algorithms and novel postprocessing methods,more » are allowing for the computation of relevant imaging parameters of interest such as patient organ doses and scatter-to-primaryratios in radiographic projections in just a few seconds using affordable computational resources. Programmable Graphics Processing Units (GPUs), for example, provide a convenient, affordable platform for parallelized Monte Carlo executions that yield simulation times on the order of 10{sup 7} xray/ s. Even with GPU acceleration, however, Monte Carlo simulation times can be prohibitive for routine clinical practice. To reduce simulation times further, variance reduction techniques can be used to alter the probabilistic models underlying the x-ray tracking process, resulting in lower variance in the results without biasing the estimates. Other complementary strategies for further reductions in computation time are denoising of the Monte Carlo estimates and estimating (scoring) the quantity of interest at a sparse set of sampling locations (e.g. at a small number of detector pixels in a scatter simulation) followed by interpolation. Beyond reduction of the computational resources required for performing Monte Carlo simulations in medical imaging, the use of accurate representations of patient anatomy is crucial to

  3. Quasi-Monte Carlo Methods Applied to Tau-Leaping in Stochastic Biological Systems.

    PubMed

    Beentjes, Casper H L; Baker, Ruth E

    2018-05-25

    Quasi-Monte Carlo methods have proven to be effective extensions of traditional Monte Carlo methods in, amongst others, problems of quadrature and the sample path simulation of stochastic differential equations. By replacing the random number input stream in a simulation procedure by a low-discrepancy number input stream, variance reductions of several orders have been observed in financial applications. Analysis of stochastic effects in well-mixed chemical reaction networks often relies on sample path simulation using Monte Carlo methods, even though these methods suffer from typical slow [Formula: see text] convergence rates as a function of the number of sample paths N. This paper investigates the combination of (randomised) quasi-Monte Carlo methods with an efficient sample path simulation procedure, namely [Formula: see text]-leaping. We show that this combination is often more effective than traditional Monte Carlo simulation in terms of the decay of statistical errors. The observed convergence rate behaviour is, however, non-trivial due to the discrete nature of the models of chemical reactions. We explain how this affects the performance of quasi-Monte Carlo methods by looking at a test problem in standard quadrature.

  4. 78 FR 21116 - Proposed Collection; Comment Request

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-04-09

    ...-0472. Needs and Uses: Executive Order 12829, ``National Industrial Security Program (NISP)'' Section...; Comment Request AGENCY: Defense Security Service, DoD. ACTION: Notice. In compliance with Section 3506(c)(2)(A) of the Paperwork Reduction Act of 1995, the Defense Security Service (DSS) announces the...

  5. Integrated sequence and immunology filovirus database at Los Alamos

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Yusim, Karina; Yoon, Hyejin; Foley, Brian

    The Ebola outbreak of 2013–15 infected more than 28,000 people and claimed more lives than all previous filovirus outbreaks combined. Governmental agencies, clinical teams, and the world scientific community pulled together in a multifaceted response ranging from prevention and disease control, to evaluating vaccines and therapeutics in human trials. We report that as this epidemic is finally coming to a close, refocusing on long-term prevention strategies becomes paramount. Given the very real threat of future filovirus outbreaks, and the inherent uncertainty of the next outbreak virus and geographic location, it is prudent to consider the extent and implications of knownmore » natural diversity in advancing vaccines and therapeutic approaches. To facilitate such consideration, we have updated and enhanced the content of the filovirus portion of Los Alamos Hemorrhagic Fever Viruses Database. We have integrated and performed baseline analysis of all family Filoviridae sequences deposited into GenBank, with associated immune response data, and metadata, and we have added new computational tools with web-interfaces to assist users with analysis. Here, we (i) describe the main features of updated database, (ii) provide integrated views and some basic analyses summarizing evolutionary patterns as they relate to geo-temporal data captured in the database and (iii) highlight the most conserved regions in the proteome that may be useful for a T cell vaccine strategy.« less

  6. Integrated sequence and immunology filovirus database at Los Alamos

    PubMed Central

    Yoon, Hyejin; Foley, Brian; Feng, Shihai; Macke, Jennifer; Dimitrijevic, Mira; Abfalterer, Werner; Szinger, James; Fischer, Will; Kuiken, Carla; Korber, Bette

    2016-01-01

    The Ebola outbreak of 2013–15 infected more than 28 000 people and claimed more lives than all previous filovirus outbreaks combined. Governmental agencies, clinical teams, and the world scientific community pulled together in a multifaceted response ranging from prevention and disease control, to evaluating vaccines and therapeutics in human trials. As this epidemic is finally coming to a close, refocusing on long-term prevention strategies becomes paramount. Given the very real threat of future filovirus outbreaks, and the inherent uncertainty of the next outbreak virus and geographic location, it is prudent to consider the extent and implications of known natural diversity in advancing vaccines and therapeutic approaches. To facilitate such consideration, we have updated and enhanced the content of the filovirus portion of Los Alamos Hemorrhagic Fever Viruses Database. We have integrated and performed baseline analysis of all family Filoviridae sequences deposited into GenBank, with associated immune response data, and metadata, and we have added new computational tools with web-interfaces to assist users with analysis. Here, we (i) describe the main features of updated database, (ii) provide integrated views and some basic analyses summarizing evolutionary patterns as they relate to geo-temporal data captured in the database and (iii) highlight the most conserved regions in the proteome that may be useful for a T cell vaccine strategy. Database URL: www.hfv.lanl.gov PMID:27103629

  7. Integrated sequence and immunology filovirus database at Los Alamos

    DOE PAGES

    Yusim, Karina; Yoon, Hyejin; Foley, Brian; ...

    2016-01-01

    The Ebola outbreak of 2013–15 infected more than 28,000 people and claimed more lives than all previous filovirus outbreaks combined. Governmental agencies, clinical teams, and the world scientific community pulled together in a multifaceted response ranging from prevention and disease control, to evaluating vaccines and therapeutics in human trials. We report that as this epidemic is finally coming to a close, refocusing on long-term prevention strategies becomes paramount. Given the very real threat of future filovirus outbreaks, and the inherent uncertainty of the next outbreak virus and geographic location, it is prudent to consider the extent and implications of knownmore » natural diversity in advancing vaccines and therapeutic approaches. To facilitate such consideration, we have updated and enhanced the content of the filovirus portion of Los Alamos Hemorrhagic Fever Viruses Database. We have integrated and performed baseline analysis of all family Filoviridae sequences deposited into GenBank, with associated immune response data, and metadata, and we have added new computational tools with web-interfaces to assist users with analysis. Here, we (i) describe the main features of updated database, (ii) provide integrated views and some basic analyses summarizing evolutionary patterns as they relate to geo-temporal data captured in the database and (iii) highlight the most conserved regions in the proteome that may be useful for a T cell vaccine strategy.« less

  8. Calculations of the conditions for bunch leakage in the Los Alamos proton storage ring

    NASA Astrophysics Data System (ADS)

    Neuffer, D.; Ohmori, C.

    1994-04-01

    Observations are consistent with the possibility of an "ep" instability in the Los Alamos Proton Storage Ring (PSR) with both bunched and unbunched beam. The instability requires electrons to be trapped within the beam, and calculations have shown that such trapping requires leakage of beam into the interbunch gap. Observationally, leakage of beam into the gap appears necessary for the onset of the instability. In this paper we present results of studies of the longitudinal beam dynamics at PSR parameters. The studies indicate that the combined effects of the rf buncher, longitudinal space charge, and injection mismatch are sufficient to cause the observed bunch leakage. Simulation results are presented and compared with PSR observations. Variations of PSR performance parameters are considered, and methods of improving bunch confinement are suggested and studied.

  9. NUEN-618 Class Project: Actually Implicit Monte Carlo

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Vega, R. M.; Brunner, T. A.

    2017-12-14

    This research describes a new method for the solution of the thermal radiative transfer (TRT) equations that is implicit in time which will be called Actually Implicit Monte Carlo (AIMC). This section aims to introduce the TRT equations, as well as the current workhorse method which is known as Implicit Monte Carlo (IMC). As the name of the method proposed here indicates, IMC is a misnomer in that it is only semi-implicit, which will be shown in this section as well.

  10. Geometry and Dynamics for Markov Chain Monte Carlo

    NASA Astrophysics Data System (ADS)

    Barp, Alessandro; Briol, François-Xavier; Kennedy, Anthony D.; Girolami, Mark

    2018-03-01

    Markov Chain Monte Carlo methods have revolutionised mathematical computation and enabled statistical inference within many previously intractable models. In this context, Hamiltonian dynamics have been proposed as an efficient way of building chains which can explore probability densities efficiently. The method emerges from physics and geometry and these links have been extensively studied by a series of authors through the last thirty years. However, there is currently a gap between the intuitions and knowledge of users of the methodology and our deep understanding of these theoretical foundations. The aim of this review is to provide a comprehensive introduction to the geometric tools used in Hamiltonian Monte Carlo at a level accessible to statisticians, machine learners and other users of the methodology with only a basic understanding of Monte Carlo methods. This will be complemented with some discussion of the most recent advances in the field which we believe will become increasingly relevant to applied scientists.

  11. Tool for Rapid Analysis of Monte Carlo Simulations

    NASA Technical Reports Server (NTRS)

    Restrepo, Carolina; McCall, Kurt E.; Hurtado, John E.

    2011-01-01

    Designing a spacecraft, or any other complex engineering system, requires extensive simulation and analysis work. Oftentimes, the large amounts of simulation data generated are very di cult and time consuming to analyze, with the added risk of overlooking potentially critical problems in the design. The authors have developed a generic data analysis tool that can quickly sort through large data sets and point an analyst to the areas in the data set that cause specific types of failures. The Tool for Rapid Analysis of Monte Carlo simulations (TRAM) has been used in recent design and analysis work for the Orion vehicle, greatly decreasing the time it takes to evaluate performance requirements. A previous version of this tool was developed to automatically identify driving design variables in Monte Carlo data sets. This paper describes a new, parallel version, of TRAM implemented on a graphical processing unit, and presents analysis results for NASA's Orion Monte Carlo data to demonstrate its capabilities.

  12. Monte Carlo tests of the ELIPGRID-PC algorithm

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Davidson, J.R.

    1995-04-01

    The standard tool for calculating the probability of detecting pockets of contamination called hot spots has been the ELIPGRID computer code of Singer and Wickman. The ELIPGRID-PC program has recently made this algorithm available for an IBM{reg_sign} PC. However, no known independent validation of the ELIPGRID algorithm exists. This document describes a Monte Carlo simulation-based validation of a modified version of the ELIPGRID-PC code. The modified ELIPGRID-PC code is shown to match Monte Carlo-calculated hot-spot detection probabilities to within {plus_minus}0.5% for 319 out of 320 test cases. The one exception, a very thin elliptical hot spot located within a rectangularmore » sampling grid, differed from the Monte Carlo-calculated probability by about 1%. These results provide confidence in the ability of the modified ELIPGRID-PC code to accurately predict hot-spot detection probabilities within an acceptable range of error.« less

  13. Two proposed convergence criteria for Monte Carlo solutions

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Forster, R.A.; Pederson, S.P.; Booth, T.E.

    1992-01-01

    The central limit theorem (CLT) can be applied to a Monte Carlo solution if two requirements are satisfied: (1) The random variable has a finite mean and a finite variance; and (2) the number N of independent observations grows large. When these two conditions are satisfied, a confidence interval (CI) based on the normal distribution with a specified coverage probability can be formed. The first requirement is generally satisfied by the knowledge of the Monte Carlo tally being used. The Monte Carlo practitioner has a limited number of marginal methods to assess the fulfillment of the second requirement, such asmore » statistical error reduction proportional to 1/[radical]N with error magnitude guidelines. Two proposed methods are discussed in this paper to assist in deciding if N is large enough: estimating the relative variance of the variance (VOV) and examining the empirical history score probability density function (pdf).« less

  14. Venus - Lakshmi Planum and Maxwell Montes

    NASA Image and Video Library

    1996-03-07

    This full resolution radar image from NASA Magellan spacecraft is centered along the eastern edge of Lakshmi Planum and the western edge of Maxwell Montes. http://photojournal.jpl.nasa.gov/catalog/PIA00241

  15. Fixed forced detection for fast SPECT Monte-Carlo simulation

    NASA Astrophysics Data System (ADS)

    Cajgfinger, T.; Rit, S.; Létang, J. M.; Halty, A.; Sarrut, D.

    2018-03-01

    Monte-Carlo simulations of SPECT images are notoriously slow to converge due to the large ratio between the number of photons emitted and detected in the collimator. This work proposes a method to accelerate the simulations based on fixed forced detection (FFD) combined with an analytical response of the detector. FFD is based on a Monte-Carlo simulation but forces the detection of a photon in each detector pixel weighted by the probability of emission (or scattering) and transmission to this pixel. The method was evaluated with numerical phantoms and on patient images. We obtained differences with analog Monte Carlo lower than the statistical uncertainty. The overall computing time gain can reach up to five orders of magnitude. Source code and examples are available in the Gate V8.0 release.

  16. Fixed forced detection for fast SPECT Monte-Carlo simulation.

    PubMed

    Cajgfinger, T; Rit, S; Létang, J M; Halty, A; Sarrut, D

    2018-03-02

    Monte-Carlo simulations of SPECT images are notoriously slow to converge due to the large ratio between the number of photons emitted and detected in the collimator. This work proposes a method to accelerate the simulations based on fixed forced detection (FFD) combined with an analytical response of the detector. FFD is based on a Monte-Carlo simulation but forces the detection of a photon in each detector pixel weighted by the probability of emission (or scattering) and transmission to this pixel. The method was evaluated with numerical phantoms and on patient images. We obtained differences with analog Monte Carlo lower than the statistical uncertainty. The overall computing time gain can reach up to five orders of magnitude. Source code and examples are available in the Gate V8.0 release.

  17. Monte Carlo treatment planning with modulated electron radiotherapy: framework development and application

    NASA Astrophysics Data System (ADS)

    Alexander, Andrew William

    Within the field of medical physics, Monte Carlo radiation transport simulations are considered to be the most accurate method for the determination of dose distributions in patients. The McGill Monte Carlo treatment planning system (MMCTP), provides a flexible software environment to integrate Monte Carlo simulations with current and new treatment modalities. A developing treatment modality called energy and intensity modulated electron radiotherapy (MERT) is a promising modality, which has the fundamental capabilities to enhance the dosimetry of superficial targets. An objective of this work is to advance the research and development of MERT with the end goal of clinical use. To this end, we present the MMCTP system with an integrated toolkit for MERT planning and delivery of MERT fields. Delivery is achieved using an automated "few leaf electron collimator" (FLEC) and a controller. Aside from the MERT planning toolkit, the MMCTP system required numerous add-ons to perform the complex task of large-scale autonomous Monte Carlo simulations. The first was a DICOM import filter, followed by the implementation of DOSXYZnrc as a dose calculation engine and by logic methods for submitting and updating the status of Monte Carlo simulations. Within this work we validated the MMCTP system with a head and neck Monte Carlo recalculation study performed by a medical dosimetrist. The impact of MMCTP lies in the fact that it allows for systematic and platform independent large-scale Monte Carlo dose calculations for different treatment sites and treatment modalities. In addition to the MERT planning tools, various optimization algorithms were created external to MMCTP. The algorithms produced MERT treatment plans based on dose volume constraints that employ Monte Carlo pre-generated patient-specific kernels. The Monte Carlo kernels are generated from patient-specific Monte Carlo dose distributions within MMCTP. The structure of the MERT planning toolkit software and

  18. Fast quantum Monte Carlo on a GPU

    NASA Astrophysics Data System (ADS)

    Lutsyshyn, Y.

    2015-02-01

    We present a scheme for the parallelization of quantum Monte Carlo method on graphical processing units, focusing on variational Monte Carlo simulation of bosonic systems. We use asynchronous execution schemes with shared memory persistence, and obtain an excellent utilization of the accelerator. The CUDA code is provided along with a package that simulates liquid helium-4. The program was benchmarked on several models of Nvidia GPU, including Fermi GTX560 and M2090, and the Kepler architecture K20 GPU. Special optimization was developed for the Kepler cards, including placement of data structures in the register space of the Kepler GPUs. Kepler-specific optimization is discussed.

  19. Diagnosing Undersampling Biases in Monte Carlo Eigenvalue and Flux Tally Estimates

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Perfetti, Christopher M.; Rearden, Bradley T.; Marshall, William J.

    2017-02-08

    Here, this study focuses on understanding the phenomena in Monte Carlo simulations known as undersampling, in which Monte Carlo tally estimates may not encounter a sufficient number of particles during each generation to obtain unbiased tally estimates. Steady-state Monte Carlo simulations were performed using the KENO Monte Carlo tools within the SCALE code system for models of several burnup credit applications with varying degrees of spatial and isotopic complexities, and the incidence and impact of undersampling on eigenvalue and flux estimates were examined. Using an inadequate number of particle histories in each generation was found to produce a maximum bias of ~100 pcm in eigenvalue estimates and biases that exceeded 10% in fuel pin flux tally estimates. Having quantified the potential magnitude of undersampling biases in eigenvalue and flux tally estimates in these systems, this study then investigated whether Markov Chain Monte Carlo convergence metrics could be integrated into Monte Carlo simulations to predict the onset and magnitude of undersampling biases. Five potential metrics for identifying undersampling biases were implemented in the SCALE code system and evaluated for their ability to predict undersampling biases by comparing the test metric scores with the observed undersampling biases. Finally, of the five convergence metrics that were investigated, three (the Heidelberger-Welch relative half-width, the Gelman-Rubin more » $$\\hat{R}_c$$ diagnostic, and tally entropy) showed the potential to accurately predict the behavior of undersampling biases in the responses examined.« less

  20. Dunes in Nectaris Montes

    NASA Image and Video Library

    2018-05-14

    This image from NASA's Mars Reconnaissance Orbiter (MRO) shows some of these on the slopes of Nectaris Montes within Coprates Chasma. Sand dunes in Valles Marineris can be impressive in size, with steep slopes that seem to climb and descend. The brighter bedforms are inactive while the bigger dunes move over the landscape, burying and exhuming the surface. https://photojournal.jpl.nasa.gov/catalog/PIA22455

  1. Accuracy of Monte Carlo simulations compared to in-vivo MDCT dosimetry.

    PubMed

    Bostani, Maryam; Mueller, Jonathon W; McMillan, Kyle; Cody, Dianna D; Cagnon, Chris H; DeMarco, John J; McNitt-Gray, Michael F

    2015-02-01

    The purpose of this study was to assess the accuracy of a Monte Carlo simulation-based method for estimating radiation dose from multidetector computed tomography (MDCT) by comparing simulated doses in ten patients to in-vivo dose measurements. MD Anderson Cancer Center Institutional Review Board approved the acquisition of in-vivo rectal dose measurements in a pilot study of ten patients undergoing virtual colonoscopy. The dose measurements were obtained by affixing TLD capsules to the inner lumen of rectal catheters. Voxelized patient models were generated from the MDCT images of the ten patients, and the dose to the TLD for all exposures was estimated using Monte Carlo based simulations. The Monte Carlo simulation results were compared to the in-vivo dose measurements to determine accuracy. The calculated mean percent difference between TLD measurements and Monte Carlo simulations was -4.9% with standard deviation of 8.7% and a range of -22.7% to 5.7%. The results of this study demonstrate very good agreement between simulated and measured doses in-vivo. Taken together with previous validation efforts, this work demonstrates that the Monte Carlo simulation methods can provide accurate estimates of radiation dose in patients undergoing CT examinations.

  2. Automated System Calibration and Verification of the Position Measurements for the Los Alamos Isotope Production Facility and the Switchyard Kicker Facilities

    NASA Astrophysics Data System (ADS)

    Barr, D.; Gilpatrick, J. D.; Martinez, D.; Shurter, R. B.

    2004-11-01

    The Los Alamos Neutron Science Center (LANSCE) facility at Los Alamos National Laboratory has constructed both an Isotope Production Facility (IPF) and a Switchyard Kicker (XDK) as additions to the H+ and H- accelerator. These additions contain eleven Beam Position Monitors (BPMs) that measure the beam's position throughout the transport. The analog electronics within each processing module determines the beam position using the log-ratio technique. For system reliability, calibrations compensate for various temperature drifts and other imperfections in the processing electronics components. Additionally, verifications are periodically implemented by a PC running a National Instruments LabVIEW virtual instrument (VI) to verify continued system and cable integrity. The VI communicates with the processor cards via a PCI/MXI-3 VXI-crate communication module. Previously, accelerator operators performed BPM system calibrations typically once per day while beam was explicitly turned off. One of this new measurement system's unique achievements is its automated calibration and verification capability. Taking advantage of the pulsed nature of the LANSCE-facility beams, the integrated electronics hardware and VI perform calibration and verification operations between beam pulses without interrupting production beam delivery. The design, construction, and performance results of the automated calibration and verification portion of this position measurement system will be the topic of this paper.

  3. A checklist of plant and animal species at Los Alamos National Laboratory and surrounding areas

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hinojosa, H.

    Past and current members of the Biology Team (BT) of the Ecology Group have completed biological assessments (BAs) for all of the land that comprises Los Alamos National Laboratory (LANL). Within these assessments are lists of plant and animal species with the potential to exist on LANL lands and the surrounding areas. To compile these lists, BT members examined earlier published and unpublished reports, surveys, and data bases that pertained to the biota of this area or to areas that are similar. The species lists that are contained herein are compilations of the lists from these BAs, other lists thatmore » were a part of the initial research for the performance of these BAs, and more recent surveys.« less

  4. Analytical Applications of Monte Carlo Techniques.

    ERIC Educational Resources Information Center

    Guell, Oscar A.; Holcombe, James A.

    1990-01-01

    Described are analytical applications of the theory of random processes, in particular solutions obtained by using statistical procedures known as Monte Carlo techniques. Supercomputer simulations, sampling, integration, ensemble, annealing, and explicit simulation are discussed. (CW)

  5. SU-E-T-188: Film Dosimetry Verification of Monte Carlo Generated Electron Treatment Plans

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Enright, S; Asprinio, A; Lu, L

    2014-06-01

    Purpose: The purpose of this study was to compare dose distributions from film measurements to Monte Carlo generated electron treatment plans. Irradiation with electrons offers the advantages of dose uniformity in the target volume and of minimizing the dose to deeper healthy tissue. Using the Monte Carlo algorithm will improve dose accuracy in regions with heterogeneities and irregular surfaces. Methods: Dose distributions from GafChromic{sup ™} EBT3 films were compared to dose distributions from the Electron Monte Carlo algorithm in the Eclipse{sup ™} radiotherapy treatment planning system. These measurements were obtained for 6MeV, 9MeV and 12MeV electrons at two depths. Allmore » phantoms studied were imported into Eclipse by CT scan. A 1 cm thick solid water template with holes for bonelike and lung-like plugs was used. Different configurations were used with the different plugs inserted into the holes. Configurations with solid-water plugs stacked on top of one another were also used to create an irregular surface. Results: The dose distributions measured from the film agreed with those from the Electron Monte Carlo treatment plan. Accuracy of Electron Monte Carlo algorithm was also compared to that of Pencil Beam. Dose distributions from Monte Carlo had much higher pass rates than distributions from Pencil Beam when compared to the film. The pass rate for Monte Carlo was in the 80%–99% range, where the pass rate for Pencil Beam was as low as 10.76%. Conclusion: The dose distribution from Monte Carlo agreed with the measured dose from the film. When compared to the Pencil Beam algorithm, pass rates for Monte Carlo were much higher. Monte Carlo should be used over Pencil Beam for regions with heterogeneities and irregular surfaces.« less

  6. Multiple-time-stepping generalized hybrid Monte Carlo methods

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Escribano, Bruno, E-mail: bescribano@bcamath.org; Akhmatskaya, Elena; IKERBASQUE, Basque Foundation for Science, E-48013 Bilbao

    2015-01-01

    Performance of the generalized shadow hybrid Monte Carlo (GSHMC) method [1], which proved to be superior in sampling efficiency over its predecessors [2–4], molecular dynamics and hybrid Monte Carlo, can be further improved by combining it with multi-time-stepping (MTS) and mollification of slow forces. We demonstrate that the comparatively simple modifications of the method not only lead to better performance of GSHMC itself but also allow for beating the best performed methods, which use the similar force splitting schemes. In addition we show that the same ideas can be successfully applied to the conventional generalized hybrid Monte Carlo method (GHMC).more » The resulting methods, MTS-GHMC and MTS-GSHMC, provide accurate reproduction of thermodynamic and dynamical properties, exact temperature control during simulation and computational robustness and efficiency. MTS-GHMC uses a generalized momentum update to achieve weak stochastic stabilization to the molecular dynamics (MD) integrator. MTS-GSHMC adds the use of a shadow (modified) Hamiltonian to filter the MD trajectories in the HMC scheme. We introduce a new shadow Hamiltonian formulation adapted to force-splitting methods. The use of such Hamiltonians improves the acceptance rate of trajectories and has a strong impact on the sampling efficiency of the method. Both methods were implemented in the open-source MD package ProtoMol and were tested on a water and a protein systems. Results were compared to those obtained using a Langevin Molly (LM) method [5] on the same systems. The test results demonstrate the superiority of the new methods over LM in terms of stability, accuracy and sampling efficiency. This suggests that putting the MTS approach in the framework of hybrid Monte Carlo and using the natural stochasticity offered by the generalized hybrid Monte Carlo lead to improving stability of MTS and allow for achieving larger step sizes in the simulation of complex systems.« less

  7. Markov Chain Monte Carlo Methods for Bayesian Data Analysis in Astronomy

    NASA Astrophysics Data System (ADS)

    Sharma, Sanjib

    2017-08-01

    Markov Chain Monte Carlo based Bayesian data analysis has now become the method of choice for analyzing and interpreting data in almost all disciplines of science. In astronomy, over the last decade, we have also seen a steady increase in the number of papers that employ Monte Carlo based Bayesian analysis. New, efficient Monte Carlo based methods are continuously being developed and explored. In this review, we first explain the basics of Bayesian theory and discuss how to set up data analysis problems within this framework. Next, we provide an overview of various Monte Carlo based methods for performing Bayesian data analysis. Finally, we discuss advanced ideas that enable us to tackle complex problems and thus hold great promise for the future. We also distribute downloadable computer software (available at https://github.com/sanjibs/bmcmc/ ) that implements some of the algorithms and examples discussed here.

  8. Monte Carlos of the new generation: status and progress

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Frixione, Stefano

    2005-03-22

    Standard parton shower monte carlos are designed to give reliable descriptions of low-pT physics. In the very high-energy regime of modern colliders, this is may lead to largely incorrect predictions of the basic reaction processes. This motivated the recent theoretical efforts aimed at improving monte carlos through the inclusion of matrix elements computed beyond the leading order in QCD. I briefly review the progress made, and discuss bottom production at the Tevatron.

  9. Accurately modeling Gaussian beam propagation in the context of Monte Carlo techniques

    NASA Astrophysics Data System (ADS)

    Hokr, Brett H.; Winblad, Aidan; Bixler, Joel N.; Elpers, Gabriel; Zollars, Byron; Scully, Marlan O.; Yakovlev, Vladislav V.; Thomas, Robert J.

    2016-03-01

    Monte Carlo simulations are widely considered to be the gold standard for studying the propagation of light in turbid media. However, traditional Monte Carlo methods fail to account for diffraction because they treat light as a particle. This results in converging beams focusing to a point instead of a diffraction limited spot, greatly effecting the accuracy of Monte Carlo simulations near the focal plane. Here, we present a technique capable of simulating a focusing beam in accordance to the rules of Gaussian optics, resulting in a diffraction limited focal spot. This technique can be easily implemented into any traditional Monte Carlo simulation allowing existing models to be converted to include accurate focusing geometries with minimal effort. We will present results for a focusing beam in a layered tissue model, demonstrating that for different scenarios the region of highest intensity, thus the greatest heating, can change from the surface to the focus. The ability to simulate accurate focusing geometries will greatly enhance the usefulness of Monte Carlo for countless applications, including studying laser tissue interactions in medical applications and light propagation through turbid media.

  10. Review of Rover fuel element protective coating development at Los Alamos

    NASA Technical Reports Server (NTRS)

    Wallace, Terry C.

    1991-01-01

    The Los Alamos Scientific Laboratory (LASL) entered the nuclear propulsion field in 1955 and began work on all aspects of a nuclear propulsion program with a target exhaust temperature of about 2750 K. A very extensive chemical vapor deposition coating technology for preventing catastrophic corrosion of reactor core components by the high temperature, high pressure hydrogen propellant gas was developed. Over the 17-year term of the program, more than 50,000 fuel elements were coated and evaluated. Advances in performance were achieved only through closely coupled interaction between the developing fuel element fabrication and protective coating technologies. The endurance of fuel elements in high temperature, high pressure hydrogen environment increased from several minutes at 2000 K exit gas temperature to 2 hours at 2440 K exit gas temperature in a reactor test and 10 hours at 2350 K exit gas temperature in a hot gas test. The purpose of this paper is to highlight the rationale for selection of coating materials used (NbC and ZrC), identify critical fuel element-coat interactions that had to be modified to increase system performance, and review the evolution of protective coating technology.

  11. Fission matrix-based Monte Carlo criticality analysis of fuel storage pools

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Farlotti, M.; Ecole Polytechnique, Palaiseau, F 91128; Larsen, E. W.

    2013-07-01

    Standard Monte Carlo transport procedures experience difficulties in solving criticality problems in fuel storage pools. Because of the strong neutron absorption between fuel assemblies, source convergence can be very slow, leading to incorrect estimates of the eigenvalue and the eigenfunction. This study examines an alternative fission matrix-based Monte Carlo transport method that takes advantage of the geometry of a storage pool to overcome this difficulty. The method uses Monte Carlo transport to build (essentially) a fission matrix, which is then used to calculate the criticality and the critical flux. This method was tested using a test code on a simplemore » problem containing 8 assemblies in a square pool. The standard Monte Carlo method gave the expected eigenfunction in 5 cases out of 10, while the fission matrix method gave the expected eigenfunction in all 10 cases. In addition, the fission matrix method provides an estimate of the error in the eigenvalue and the eigenfunction, and it allows the user to control this error by running an adequate number of cycles. Because of these advantages, the fission matrix method yields a higher confidence in the results than standard Monte Carlo. We also discuss potential improvements of the method, including the potential for variance reduction techniques. (authors)« less

  12. Accuracy of Monte Carlo simulations compared to in-vivo MDCT dosimetry

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bostani, Maryam, E-mail: mbostani@mednet.ucla.edu; McMillan, Kyle; Cagnon, Chris H.

    Purpose: The purpose of this study was to assess the accuracy of a Monte Carlo simulation-based method for estimating radiation dose from multidetector computed tomography (MDCT) by comparing simulated doses in ten patients to in-vivo dose measurements. Methods: MD Anderson Cancer Center Institutional Review Board approved the acquisition of in-vivo rectal dose measurements in a pilot study of ten patients undergoing virtual colonoscopy. The dose measurements were obtained by affixing TLD capsules to the inner lumen of rectal catheters. Voxelized patient models were generated from the MDCT images of the ten patients, and the dose to the TLD for allmore » exposures was estimated using Monte Carlo based simulations. The Monte Carlo simulation results were compared to the in-vivo dose measurements to determine accuracy. Results: The calculated mean percent difference between TLD measurements and Monte Carlo simulations was −4.9% with standard deviation of 8.7% and a range of −22.7% to 5.7%. Conclusions: The results of this study demonstrate very good agreement between simulated and measured doses in-vivo. Taken together with previous validation efforts, this work demonstrates that the Monte Carlo simulation methods can provide accurate estimates of radiation dose in patients undergoing CT examinations.« less

  13. Monte Carlo Methodology Serves Up a Software Success

    NASA Technical Reports Server (NTRS)

    2003-01-01

    Widely used for the modeling of gas flows through the computation of the motion and collisions of representative molecules, the Direct Simulation Monte Carlo method has become the gold standard for producing research and engineering predictions in the field of rarefied gas dynamics. Direct Simulation Monte Carlo was first introduced in the early 1960s by Dr. Graeme Bird, a professor at the University of Sydney, Australia. It has since proved to be a valuable tool to the aerospace and defense industries in providing design and operational support data, as well as flight data analysis. In 2002, NASA brought to the forefront a software product that maintains the same basic physics formulation of Dr. Bird's method, but provides effective modeling of complex, three-dimensional, real vehicle simulations and parallel processing capabilities to handle additional computational requirements, especially in areas where computational fluid dynamics (CFD) is not applicable. NASA's Direct Simulation Monte Carlo Analysis Code (DAC) software package is now considered the Agency s premier high-fidelity simulation tool for predicting vehicle aerodynamics and aerothermodynamic environments in rarified, or low-density, gas flows.

  14. Dynamic Monte Carlo description of thermal desorption processes

    NASA Astrophysics Data System (ADS)

    Weinketz, Sieghard

    1994-07-01

    The applicability of the dynamic Monte Carlo method of Fichthorn and Weinberg, in which the time evolution of a system is described in terms of the absolute number of different microscopic possible events and their associated transition rates, is discussed for the case of thermal desorption simulations. It is shown that the definition of the time increment at each successful event leads naturally to the macroscopic differential equation of desorption, in the case of simple first- and second-order processes in which the only possible events are desorption and diffusion. This equivalence is numerically demonstrated for a second-order case. In the sequence, the equivalence of this method with the Monte Carlo method of Sales and Zgrablich for more complex desorption processes, allowing for lateral interactions between adsorbates, is shown, even though the dynamic Monte Carlo method does not bear their limitation of a rapid surface diffusion condition, thus being able to describe a more complex ``kinetics'' of surface reactive processes, and therefore be applied to a wider class of phenomena, such as surface catalysis.

  15. Semi-stochastic full configuration interaction quantum Monte Carlo

    NASA Astrophysics Data System (ADS)

    Holmes, Adam; Petruzielo, Frank; Khadilkar, Mihir; Changlani, Hitesh; Nightingale, M. P.; Umrigar, C. J.

    2012-02-01

    In the recently proposed full configuration interaction quantum Monte Carlo (FCIQMC) [1,2], the ground state is projected out stochastically, using a population of walkers each of which represents a basis state in the Hilbert space spanned by Slater determinants. The infamous fermion sign problem manifests itself in the fact that walkers of either sign can be spawned on a given determinant. We propose an improvement on this method in the form of a hybrid stochastic/deterministic technique, which we expect will improve the efficiency of the algorithm by ameliorating the sign problem. We test the method on atoms and molecules, e.g., carbon, carbon dimer, N2 molecule, and stretched N2. [4pt] [1] Fermion Monte Carlo without fixed nodes: a Game of Life, death and annihilation in Slater Determinant space. George Booth, Alex Thom, Ali Alavi. J Chem Phys 131, 050106, (2009).[0pt] [2] Survival of the fittest: Accelerating convergence in full configuration-interaction quantum Monte Carlo. Deidre Cleland, George Booth, and Ali Alavi. J Chem Phys 132, 041103 (2010).

  16. pyNSMC: A Python Module for Null-Space Monte Carlo Uncertainty Analysis

    NASA Astrophysics Data System (ADS)

    White, J.; Brakefield, L. K.

    2015-12-01

    The null-space monte carlo technique is a non-linear uncertainty analyses technique that is well-suited to high-dimensional inverse problems. While the technique is powerful, the existing workflow for completing null-space monte carlo is cumbersome, requiring the use of multiple commandline utilities, several sets of intermediate files and even a text editor. pyNSMC is an open-source python module that automates the workflow of null-space monte carlo uncertainty analyses. The module is fully compatible with the PEST and PEST++ software suites and leverages existing functionality of pyEMU, a python framework for linear-based uncertainty analyses. pyNSMC greatly simplifies the existing workflow for null-space monte carlo by taking advantage of object oriented design facilities in python. The core of pyNSMC is the ensemble class, which draws and stores realized random vectors and also provides functionality for exporting and visualizing results. By relieving users of the tedium associated with file handling and command line utility execution, pyNSMC instead focuses the user on the important steps and assumptions of null-space monte carlo analysis. Furthermore, pyNSMC facilitates learning through flow charts and results visualization, which are available at many points in the algorithm. The ease-of-use of the pyNSMC workflow is compared to the existing workflow for null-space monte carlo for a synthetic groundwater model with hundreds of estimable parameters.

  17. Studies of fission fragment properties at the Los Alamos Neutron Science Center (LANSCE)

    NASA Astrophysics Data System (ADS)

    Tovesson, Fredrik; Mayorov, Dmitriy; Duke, Dana; Manning, Brett; Geppert-Kleinrath, Verena

    2017-09-01

    Nuclear data related to the fission process are needed for a wide variety of research areas, including fundamental science, nuclear energy and non-proliferation. While some of the relevant data have been measured to the required accuracies there are still many aspects of fission that need further investigation. One such aspect is how Total Kinetic Energy (TKE), fragment yields, angular distributions and other fission observables depend on excitation energy of the fissioning system. Another question is the correlation between mass, charge and energy of fission fragments. At the Los Alamos Neutron Science Center (LANSCE) we are studying neutron-induced fission at incident energies from thermal up to hundreds of MeV using the Lujan Center and Weapons Neutron Research (WNR) facilities. Advanced instruments such as SPIDER (time-of-flight and kinetic energy spectrometer), the NIFFTE Time Projection Chamber (TPC), and Frisch grid Ionization Chambers (FGIC) are used to investigate the properties of fission fragments, and some important results for the major actinides have been obtained.

  18. Options Assessment Report: Treatment of Nitrate Salt Waste at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Robinson, Bruce Alan; Stevens, Patrice Ann

    2015-12-17

    This report documents the methodology used to select a method of treatment for the remediated nitrate salt (RNS) and unremediated nitrate salt (UNS) waste containers at Los Alamos National Laboratory (LANL). The method selected should treat the containerized waste in a manner that renders the waste safe and suitable for transport and final disposal in the Waste Isolation Pilot Plant (WIPP) repository, under specifications listed in the WIPP Waste Acceptance Criteria (DOE/CBFO, 2013). LANL recognizes that the results must be thoroughly vetted with the New Mexico Environment Department (NMED) and that a modification to the LANL Hazardous Waste Facility Permitmore » is a necessary step before implementation of this or any treatment option. Likewise, facility readiness and safety basis approvals must be received from the Department of Energy (DOE). This report presents LANL’s preferred option, and the documentation of the process for reaching the recommended treatment option for RNS and UNS waste, and is presented for consideration by NMED and DOE.« less

  19. 75 FR 5994 - Submission for OMB Review; Comment Request

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-02-05

    ... and OMB Number: Voice of Industry Survey; OMB Control Number 0704-TBD. Type of Request: New. Number of... Security Program (NISP)'' Section 202(a) stipulates that the Secretary of Defense shall serve as the... Security Agency (CSA), designated by the NISPOM, is responsible for determining the frequency of Security...

  20. Wireline-rotary air coring of the Bandelier Tuff, Los Alamos, New Mexico

    USGS Publications Warehouse

    Teasdale, W.E.; Pemberton, R.R.

    1984-01-01

    This paper describes experiments using wireline-rotary air-coring techniques conducted in the Bandelier Tuff using a modified standard wireline core-barrel system. The modified equipment was used to collect uncontaminated cores of unconsolidated ash and indurated tuff at Los Alamos, New Mexico. Core recovery obtained from the 210-foot deep test hole was about 92 percent. A standard HQ-size, triple-tube wireline core barrel (designed for the passage of liquid drilling fluids) was modified for air coring as follows: (1) Air passages were milled in the latch body part of the head assembly; (2) the inside dimension of the outer core barrel tube was machined and honed to provide greater clearance between the inner and outer barrels; (3) oversized reaming devices were added to the outer core barrel and the coring bit to allow more clearance for air and cuttings return; (4) the eight discharge ports in the coring bit were enlarged. To control airborne-dust pollution, a dust-and-cuttings discharge subassembly, designed and built by project personnel, was used. (USGS)

  1. Final Progress Report: Internship at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Dunham, Ryan Q.

    2012-08-10

    Originally I was tasked fluidized bed modeling, however, I changed projects. While still working with ANSYS Fluent, I performed a study of particle tracks in glove boxes. This is useful from a Health-Physics perspective, dealing respirable particles that can be hazardous to the human body. I iteratively tested different amounts of turbulent particles in a steady-state flow. The goal of this testing was to discover how Fluent handles built-in Rosin-Rammler distributions for particle injections. I worked on the health physics flow problems and distribution analysis under the direction of two mentors, Bruce Letellier and Dave Decroix. I set up andmore » ran particle injection calculations using Fluent. I tried different combinations of input parameters to produce sets of 500,000, 1 million, and 1.5 million particles to determine what a good test case would be for future experiments. I performed a variety of tasks in my work as an Undergraduate Student Intern at LANL this summer, and learned how to use a powerful CFD application in addition to expanding my skills in MATLAB. I enjoyed my work at LANL and hope to be able to use the experience here to further my career in the future working in a security-conscious environment. My mentors provided guidance and help with all of my projects and I am grateful for the opportunity to work at Los Alamos National Laboratory.« less

  2. A Monte Carlo Simulation of Brownian Motion in the Freshman Laboratory

    ERIC Educational Resources Information Center

    Anger, C. D.; Prescott, J. R.

    1970-01-01

    Describes a dry- lab" experiment for the college freshman laboratory, in which the essential features of Browian motion are given principles, using the Monte Carlo technique. Calculations principles, using the Monte Carlo technique. Calculations are carried out by a computation sheme based on computer language. Bibliography. (LC)

  3. Multiscale Monte Carlo equilibration: Pure Yang-Mills theory

    DOE PAGES

    Endres, Michael G.; Brower, Richard C.; Orginos, Kostas; ...

    2015-12-29

    In this study, we present a multiscale thermalization algorithm for lattice gauge theory, which enables efficient parallel generation of uncorrelated gauge field configurations. The algorithm combines standard Monte Carlo techniques with ideas drawn from real space renormalization group and multigrid methods. We demonstrate the viability of the algorithm for pure Yang-Mills gauge theory for both heat bath and hybrid Monte Carlo evolution, and show that it ameliorates the problem of topological freezing up to controllable lattice spacing artifacts.

  4. Improved Monte Carlo Renormalization Group Method

    DOE R&D Accomplishments Database

    Gupta, R.; Wilson, K. G.; Umrigar, C.

    1985-01-01

    An extensive program to analyze critical systems using an Improved Monte Carlo Renormalization Group Method (IMCRG) being undertaken at LANL and Cornell is described. Here we first briefly review the method and then list some of the topics being investigated.

  5. Optimised Iteration in Coupled Monte Carlo - Thermal-Hydraulics Calculations

    NASA Astrophysics Data System (ADS)

    Hoogenboom, J. Eduard; Dufek, Jan

    2014-06-01

    This paper describes an optimised iteration scheme for the number of neutron histories and the relaxation factor in successive iterations of coupled Monte Carlo and thermal-hydraulic reactor calculations based on the stochastic iteration method. The scheme results in an increasing number of neutron histories for the Monte Carlo calculation in successive iteration steps and a decreasing relaxation factor for the spatial power distribution to be used as input to the thermal-hydraulics calculation. The theoretical basis is discussed in detail and practical consequences of the scheme are shown, among which a nearly linear increase per iteration of the number of cycles in the Monte Carlo calculation. The scheme is demonstrated for a full PWR type fuel assembly. Results are shown for the axial power distribution during several iteration steps. A few alternative iteration method are also tested and it is concluded that the presented iteration method is near optimal.

  6. Reconstruction of Human Monte Carlo Geometry from Segmented Images

    NASA Astrophysics Data System (ADS)

    Zhao, Kai; Cheng, Mengyun; Fan, Yanchang; Wang, Wen; Long, Pengcheng; Wu, Yican

    2014-06-01

    Human computational phantoms have been used extensively for scientific experimental analysis and experimental simulation. This article presented a method for human geometry reconstruction from a series of segmented images of a Chinese visible human dataset. The phantom geometry could actually describe detailed structure of an organ and could be converted into the input file of the Monte Carlo codes for dose calculation. A whole-body computational phantom of Chinese adult female has been established by FDS Team which is named Rad-HUMAN with about 28.8 billion voxel number. For being processed conveniently, different organs on images were segmented with different RGB colors and the voxels were assigned with positions of the dataset. For refinement, the positions were first sampled. Secondly, the large sums of voxels inside the organ were three-dimensional adjacent, however, there were not thoroughly mergence methods to reduce the cell amounts for the description of the organ. In this study, the voxels on the organ surface were taken into consideration of the mergence which could produce fewer cells for the organs. At the same time, an indexed based sorting algorithm was put forward for enhancing the mergence speed. Finally, the Rad-HUMAN which included a total of 46 organs and tissues was described by the cuboids into the Monte Carlo Monte Carlo Geometry for the simulation. The Monte Carlo geometry was constructed directly from the segmented images and the voxels was merged exhaustively. Each organ geometry model was constructed without ambiguity and self-crossing, its geometry information could represent the accuracy appearance and precise interior structure of the organs. The constructed geometry largely retaining the original shape of organs could easily be described into different Monte Carlo codes input file such as MCNP. Its universal property was testified and high-performance was experimentally verified

  7. Accelerated Monte Carlo Simulation for Safety Analysis of the Advanced Airspace Concept

    NASA Technical Reports Server (NTRS)

    Thipphavong, David

    2010-01-01

    Safe separation of aircraft is a primary objective of any air traffic control system. An accelerated Monte Carlo approach was developed to assess the level of safety provided by a proposed next-generation air traffic control system. It combines features of fault tree and standard Monte Carlo methods. It runs more than one order of magnitude faster than the standard Monte Carlo method while providing risk estimates that only differ by about 10%. It also preserves component-level model fidelity that is difficult to maintain using the standard fault tree method. This balance of speed and fidelity allows sensitivity analysis to be completed in days instead of weeks or months with the standard Monte Carlo method. Results indicate that risk estimates are sensitive to transponder, pilot visual avoidance, and conflict detection failure probabilities.

  8. Los Alamos National Security, LLC Request for Interest (RFI) for Investment Mentors to participate in the Laboratory’s Entrepreneurial Postdoctoral Pilot.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Clow, Shandra Deann

    Los Alamos National Laboratory (LANL) is committed to understanding how the role of venture funding, new investment mechanisms, and fostering the development of a culture of entrepreneurship may enhance the Laboratory and bring strength and creativity to its people. LANL, in partnership with the University of California (UC), has created the Entrepreneurial Postdoctoral Fellowship Pilot (Pilot) to provide an immersion-based learning opportunity to post-doctoral researchers to develop and practice skills in entrepreneurship and comercialization.

  9. Final tolerancing approach and the value of short-cutting tolerances by measurement

    NASA Astrophysics Data System (ADS)

    Grupp, Frank; Prieto, Eric; Geis, Norbert; Bode, Andreas; Bodendorf, Christof; Costille, Anne; Katterloher, Reinhard; Penka, Daniela; Bender, Ralf

    2016-07-01

    Within the ESAs 2015 - 2025 Cosmic Vision framework the 1.2 m aperture EUCLID space telescope addresses cosmological questions related to dark matter and dark energy. Being equipped with two instruments that are simultaneously observing patches of > 0.5 square degree on the sky EUCLID is aiming at major cosmological probes in a large seven years survey scanning the entire extragalactic sky. These two instruments, the visual light high spacial resolution imager (VIS) and the near infrared spectrometer and photometer (NISP) are separated by a dichroic beam splitter. Its huge field of view (FoV) - larger than the full moon disk - together with high demands on the optical performance and strong requirements on in flight stability lead to very challenging demands on alignment and post launch - post cool-down optical element position. The role of an accurate and trust-worthy tolerance analysis which is well adopted to the stepwise integration and alignment concept, as well as to the missions stability properties is therefore crucial for the missions success. While the previous contributions of this series of papers (e.g.[1])was addressing the technical aspects of tolerancing, the mechanical challenges and the answers of the NISP instrument to these challenges, this paper will focus on our concept of shortcutting the tolerance chain by measurement wherever useful and possible. The NISP instrument is only possible, due to the innovative use of technologies such as computer generated hologram (CGH) based manufacturing and alignment. Expanding this concept, certain steps in the assembly process, such as focal length determination before detector placement allow to reduce the overall tolerance induced imaging errors. With this papers we show three major examples of this shortcutting strategy.

  10. Path integral Monte Carlo and the electron gas

    NASA Astrophysics Data System (ADS)

    Brown, Ethan W.

    Path integral Monte Carlo is a proven method for accurately simulating quantum mechanical systems at finite-temperature. By stochastically sampling Feynman's path integral representation of the quantum many-body density matrix, path integral Monte Carlo includes non-perturbative effects like thermal fluctuations and particle correlations in a natural way. Over the past 30 years, path integral Monte Carlo has been successfully employed to study the low density electron gas, high-pressure hydrogen, and superfluid helium. For systems where the role of Fermi statistics is important, however, traditional path integral Monte Carlo simulations have an exponentially decreasing efficiency with decreased temperature and increased system size. In this thesis, we work towards improving this efficiency, both through approximate and exact methods, as specifically applied to the homogeneous electron gas. We begin with a brief overview of the current state of atomic simulations at finite-temperature before we delve into a pedagogical review of the path integral Monte Carlo method. We then spend some time discussing the one major issue preventing exact simulation of Fermi systems, the sign problem. Afterwards, we introduce a way to circumvent the sign problem in PIMC simulations through a fixed-node constraint. We then apply this method to the homogeneous electron gas at a large swatch of densities and temperatures in order to map out the warm-dense matter regime. The electron gas can be a representative model for a host of real systems, from simple medals to stellar interiors. However, its most common use is as input into density functional theory. To this end, we aim to build an accurate representation of the electron gas from the ground state to the classical limit and examine its use in finite-temperature density functional formulations. The latter half of this thesis focuses on possible routes beyond the fixed-node approximation. As a first step, we utilize the variational

  11. Monte Carlo methods for multidimensional integration for European option pricing

    NASA Astrophysics Data System (ADS)

    Todorov, V.; Dimov, I. T.

    2016-10-01

    In this paper, we illustrate examples of highly accurate Monte Carlo and quasi-Monte Carlo methods for multiple integrals related to the evaluation of European style options. The idea is that the value of the option is formulated in terms of the expectation of some random variable; then the average of independent samples of this random variable is used to estimate the value of the option. First we obtain an integral representation for the value of the option using the risk neutral valuation formula. Then with an appropriations change of the constants we obtain a multidimensional integral over the unit hypercube of the corresponding dimensionality. Then we compare a specific type of lattice rules over one of the best low discrepancy sequence of Sobol for numerical integration. Quasi-Monte Carlo methods are compared with Adaptive and Crude Monte Carlo techniques for solving the problem. The four approaches are completely different thus it is a question of interest to know which one of them outperforms the other for evaluation multidimensional integrals in finance. Some of the advantages and disadvantages of the developed algorithms are discussed.

  12. Harnessing graphical structure in Markov chain Monte Carlo learning

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Stolorz, P.E.; Chew P.C.

    1996-12-31

    The Monte Carlo method is recognized as a useful tool in learning and probabilistic inference methods common to many datamining problems. Generalized Hidden Markov Models and Bayes nets are especially popular applications. However, the presence of multiple modes in many relevant integrands and summands often renders the method slow and cumbersome. Recent mean field alternatives designed to speed things up have been inspired by experience gleaned from physics. The current work adopts an approach very similar to this in spirit, but focusses instead upon dynamic programming notions as a basis for producing systematic Monte Carlo improvements. The idea is tomore » approximate a given model by a dynamic programming-style decomposition, which then forms a scaffold upon which to build successively more accurate Monte Carlo approximations. Dynamic programming ideas alone fail to account for non-local structure, while standard Monte Carlo methods essentially ignore all structure. However, suitably-crafted hybrids can successfully exploit the strengths of each method, resulting in algorithms that combine speed with accuracy. The approach relies on the presence of significant {open_quotes}local{close_quotes} information in the problem at hand. This turns out to be a plausible assumption for many important applications. Example calculations are presented, and the overall strengths and weaknesses of the approach are discussed.« less

  13. PEPSI — a Monte Carlo generator for polarized leptoproduction

    NASA Astrophysics Data System (ADS)

    Mankiewicz, L.; Schäfer, A.; Veltri, M.

    1992-09-01

    We describe PEPSI (Polarized Electron Proton Scattering Interactions), a Monte Carlo program for polarized deep inelastic leptoproduction mediated by electromagnetic interaction, and explain how to use it. The code is a modification of the LEPTO 4.3 Lund Monte Carlo for unpolarized scattering. The hard virtual gamma-parton scattering is generated according to the polarization-dependent QCD cross-section of the first order in α S. PEPSI requires the standard polarization-independent JETSET routines to simulate the fragmentation into final hadrons.

  14. Hypothesis testing of scientific Monte Carlo calculations.

    PubMed

    Wallerberger, Markus; Gull, Emanuel

    2017-11-01

    The steadily increasing size of scientific Monte Carlo simulations and the desire for robust, correct, and reproducible results necessitates rigorous testing procedures for scientific simulations in order to detect numerical problems and programming bugs. However, the testing paradigms developed for deterministic algorithms have proven to be ill suited for stochastic algorithms. In this paper we demonstrate explicitly how the technique of statistical hypothesis testing, which is in wide use in other fields of science, can be used to devise automatic and reliable tests for Monte Carlo methods, and we show that these tests are able to detect some of the common problems encountered in stochastic scientific simulations. We argue that hypothesis testing should become part of the standard testing toolkit for scientific simulations.

  15. Hypothesis testing of scientific Monte Carlo calculations

    NASA Astrophysics Data System (ADS)

    Wallerberger, Markus; Gull, Emanuel

    2017-11-01

    The steadily increasing size of scientific Monte Carlo simulations and the desire for robust, correct, and reproducible results necessitates rigorous testing procedures for scientific simulations in order to detect numerical problems and programming bugs. However, the testing paradigms developed for deterministic algorithms have proven to be ill suited for stochastic algorithms. In this paper we demonstrate explicitly how the technique of statistical hypothesis testing, which is in wide use in other fields of science, can be used to devise automatic and reliable tests for Monte Carlo methods, and we show that these tests are able to detect some of the common problems encountered in stochastic scientific simulations. We argue that hypothesis testing should become part of the standard testing toolkit for scientific simulations.

  16. Mission Analysis, Operations, and Navigation Toolkit Environment (Monte) Version 040

    NASA Technical Reports Server (NTRS)

    Sunseri, Richard F.; Wu, Hsi-Cheng; Evans, Scott E.; Evans, James R.; Drain, Theodore R.; Guevara, Michelle M.

    2012-01-01

    Monte is a software set designed for use in mission design and spacecraft navigation operations. The system can process measurement data, design optimal trajectories and maneuvers, and do orbit determination, all in one application. For the first time, a single software set can be used for mission design and navigation operations. This eliminates problems due to different models and fidelities used in legacy mission design and navigation software. The unique features of Monte 040 include a blowdown thruster model for GRAIL (Gravity Recovery and Interior Laboratory) with associated pressure models, as well as an updated, optimalsearch capability (COSMIC) that facilitated mission design for ARTEMIS. Existing legacy software lacked the capabilities necessary for these two missions. There is also a mean orbital element propagator and an osculating to mean element converter that allows long-term orbital stability analysis for the first time in compiled code. The optimized trajectory search tool COSMIC allows users to place constraints and controls on their searches without any restrictions. Constraints may be user-defined and depend on trajectory information either forward or backwards in time. In addition, a long-term orbit stability analysis tool (morbiter) existed previously as a set of scripts on top of Monte. Monte is becoming the primary tool for navigation operations, a core competency at JPL. The mission design capabilities in Monte are becoming mature enough for use in project proposals as well as post-phase A mission design. Monte has three distinct advantages over existing software. First, it is being developed in a modern paradigm: object- oriented C++ and Python. Second, the software has been developed as a toolkit, which allows users to customize their own applications and allows the development team to implement requirements quickly, efficiently, and with minimal bugs. Finally, the software is managed in accordance with the CMMI (Capability Maturity Model

  17. PRELIMINARY COUPLING OF THE MONTE CARLO CODE OPENMC AND THE MULTIPHYSICS OBJECT-ORIENTED SIMULATION ENVIRONMENT (MOOSE) FOR ANALYZING DOPPLER FEEDBACK IN MONTE CARLO SIMULATIONS

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Matthew Ellis; Derek Gaston; Benoit Forget

    In recent years the use of Monte Carlo methods for modeling reactors has become feasible due to the increasing availability of massively parallel computer systems. One of the primary challenges yet to be fully resolved, however, is the efficient and accurate inclusion of multiphysics feedback in Monte Carlo simulations. The research in this paper presents a preliminary coupling of the open source Monte Carlo code OpenMC with the open source Multiphysics Object-Oriented Simulation Environment (MOOSE). The coupling of OpenMC and MOOSE will be used to investigate efficient and accurate numerical methods needed to include multiphysics feedback in Monte Carlo codes.more » An investigation into the sensitivity of Doppler feedback to fuel temperature approximations using a two dimensional 17x17 PWR fuel assembly is presented in this paper. The results show a functioning multiphysics coupling between OpenMC and MOOSE. The coupling utilizes Functional Expansion Tallies to accurately and efficiently transfer pin power distributions tallied in OpenMC to unstructured finite element meshes used in MOOSE. The two dimensional PWR fuel assembly case also demonstrates that for a simplified model the pin-by-pin doppler feedback can be adequately replicated by scaling a representative pin based on pin relative powers.« less

  18. COMPARISON OF MONTE CARLO METHODS FOR NONLINEAR RADIATION TRANSPORT

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    W. R. MARTIN; F. B. BROWN

    2001-03-01

    Five Monte Carlo methods for solving the nonlinear thermal radiation transport equations are compared. The methods include the well-known Implicit Monte Carlo method (IMC) developed by Fleck and Cummings, an alternative to IMC developed by Carter and Forest, an ''exact'' method recently developed by Ahrens and Larsen, and two methods recently proposed by Martin and Brown. The five Monte Carlo methods are developed and applied to the radiation transport equation in a medium assuming local thermodynamic equilibrium. Conservation of energy is derived and used to define appropriate material energy update equations for each of the methods. Details of the Montemore » Carlo implementation are presented, both for the random walk simulation and the material energy update. Simulation results for all five methods are obtained for two infinite medium test problems and a 1-D test problem, all of which have analytical solutions. Conclusions regarding the relative merits of the various schemes are presented.« less

  19. Ground state of excitonic molecules by the Green's-function Monte Carlo method

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lee, M.A.; Vashishta, P.; Kalia, R.K.

    1983-12-26

    The ground-state energy of excitonic molecules is evaluated as a function of the ratio of electron and hole masses, sigma, with use of the Green's-function Monte Carlo method. For all sigma, the Green's-function Monte Carlo energies are significantly lower than the variational estimates and in favorable agreement with experiments. In excitonic rydbergs, the binding energy of the positronium molecule (sigma = 1) is predicted to be -0.06 and for sigma<<1, the Green's-function Monte Carlo energies agree with the ''exact'' limiting behavior, E = -2.346+0.764sigma.

  20. Structural Reliability and Monte Carlo Simulation.

    ERIC Educational Resources Information Center

    Laumakis, P. J.; Harlow, G.

    2002-01-01

    Analyzes a simple boom structure and assesses its reliability using elementary engineering mechanics. Demonstrates the power and utility of Monte-Carlo simulation by showing that such a simulation can be implemented more readily with results that compare favorably to the theoretical calculations. (Author/MM)

  1. Physical time scale in kinetic Monte Carlo simulations of continuous-time Markov chains.

    PubMed

    Serebrinsky, Santiago A

    2011-03-01

    We rigorously establish a physical time scale for a general class of kinetic Monte Carlo algorithms for the simulation of continuous-time Markov chains. This class of algorithms encompasses rejection-free (or BKL) and rejection (or "standard") algorithms. For rejection algorithms, it was formerly considered that the availability of a physical time scale (instead of Monte Carlo steps) was empirical, at best. Use of Monte Carlo steps as a time unit now becomes completely unnecessary.

  2. Correlated uncertainties in Monte Carlo reaction rate calculations

    NASA Astrophysics Data System (ADS)

    Longland, Richard

    2017-07-01

    Context. Monte Carlo methods have enabled nuclear reaction rates from uncertain inputs to be presented in a statistically meaningful manner. However, these uncertainties are currently computed assuming no correlations between the physical quantities that enter those calculations. This is not always an appropriate assumption. Astrophysically important reactions are often dominated by resonances, whose properties are normalized to a well-known reference resonance. This insight provides a basis from which to develop a flexible framework for including correlations in Monte Carlo reaction rate calculations. Aims: The aim of this work is to develop and test a method for including correlations in Monte Carlo reaction rate calculations when the input has been normalized to a common reference. Methods: A mathematical framework is developed for including correlations between input parameters in Monte Carlo reaction rate calculations. The magnitude of those correlations is calculated from the uncertainties typically reported in experimental papers, where full correlation information is not available. The method is applied to four illustrative examples: a fictional 3-resonance reaction, 27Al(p, γ)28Si, 23Na(p, α)20Ne, and 23Na(α, p)26Mg. Results: Reaction rates at low temperatures that are dominated by a few isolated resonances are found to minimally impacted by correlation effects. However, reaction rates determined from many overlapping resonances can be significantly affected. Uncertainties in the 23Na(α, p)26Mg reaction, for example, increase by up to a factor of 5. This highlights the need to take correlation effects into account in reaction rate calculations, and provides insight into which cases are expected to be most affected by them. The impact of correlation effects on nucleosynthesis is also investigated.

  3. Los Alamos National Laboratory: A guide to records series supporting epidemiologic studies conducted for the Department of Energy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    NONE

    1997-01-01

    The purpose of this guide is to describe each series of records that pertains to the epidemiologic studies conducted by the Epidemiology Section of the Occupational Medicine Group (ESH-2) at the Department of Energy`s (DOE) Los Alamos National Laboratory (LANL) in Los Alamos, New Mexico. The records described in this guide relate to occupational studies performed by the Epidemiology Section, including those pertaining to workers at LANL, Mound Plant, Oak Ridge Reservation, Pantex Plant, Rocky Flats Plant, and Savannah River Site. Also included are descriptions of other health-related records generated or collected by the Epidemiology Section and a small setmore » of records collected by the Industrial Hygiene and Safety Group. This guide is not designed to describe the universe of records generated by LANL which may be used for epidemiologic studies of the LANL work force. History Associates Incorporated (HAI) prepared this guide as part of its work as the support services contractor for DOE`s Epidemiologic Records Inventory Project. This introduction briefly describes the Epidemiologic Records Inventory Project, HAI`s role in the project, the history of LANL the history and functions of LANL`s Health Division and Epidemiology Section, and the various epidemiologic studies performed by the Epidemiology Section. It provides information on the methodology that HAI used to inventory and describe records housed in the offices of the LANL Epidemiology Section in Technical Area 59 and at the LANL Records Center. Other topics include the methodology used to produce the guide, the arrangement of the detailed record series descriptions, and information concerning access to records repositories.« less

  4. A Monte Carlo simulation study of associated liquid crystals

    NASA Astrophysics Data System (ADS)

    Berardi, R.; Fehervari, M.; Zannoni, C.

    We have performed a Monte Carlo simulation study of a system of ellipsoidal particles with donor-acceptor sites modelling complementary hydrogen-bonding groups in real molecules. We have considered elongated Gay-Berne particles with terminal interaction sites allowing particles to associate and form dimers. The changes in the phase transitions and in the molecular organization and the interplay between orientational ordering and dimer formation are discussed. Particle flip and dimer moves have been used to increase the convergency rate of the Monte Carlo (MC) Markov chain.

  5. Efficient Monte Carlo Methods for Biomolecular Simulations.

    NASA Astrophysics Data System (ADS)

    Bouzida, Djamal

    A new approach to efficient Monte Carlo simulations of biological molecules is presented. By relaxing the usual restriction to Markov processes, we are able to optimize performance while dealing directly with the inhomogeneity and anisotropy inherent in these systems. The advantage of this approach is that we can introduce a wide variety of Monte Carlo moves to deal with complicated motions of the molecule, while maintaining full optimization at every step. This enables the use of a variety of collective rotational moves that relax long-wavelength modes. We were able to show by explicit simulations that the resulting algorithms substantially increase the speed of the simulation while reproducing the correct equilibrium behavior. This approach is particularly intended for simulations of macromolecules, although we expect it to be useful in other situations. The dynamic optimization of the new Monte Carlo methods makes them very suitable for simulated annealing experiments on all systems whose state space is continuous in general, and to the protein folding problem in particular. We introduce an efficient annealing schedule using preferential bias moves. Our simulated annealing experiments yield structures whose free energies were lower than the equilibrated X-ray structure, which leads us to believe that the empirical energy function used does not fully represent the interatomic interactions. Furthermore, we believe that the largest discrepancies involve the solvent effects in particular.

  6. Sediment Transport in the Bill Williams River and Turbidity in Lake Havasu During and Following Two High Releases from Alamo Dam, Arizona, in 2005 and 2006

    USGS Publications Warehouse

    Wiele, Stephen M.; Hart, Robert J.; Darling, Hugh L.; Hautzinger, Andrew B.

    2009-01-01

    Discharges higher than are typically released from Alamo Dam in west-central Arizona were planned and released in 2005, 2006, 2007, and 2008 to study the effects of these releases on the Bill Williams River and Lake Havasu, into which the river debouches. Sediment concentrations and water discharges were measured in the Bill Williams River, and turbidity, temperature, and dissolved oxygen were measured in Lake Havasu during and after experimental releases in 2005 and 2006 from Alamo Dam. Data from such releases will support ongoing ecological studies, improve environmentally sensitive management of the river corridor, and support the development of a predictive relationship between the operation of Alamo Dam and downstream flows and their impact on Lake Havasu and the Colorado River. Elevated discharges in the Bill Williams River mobilize more sediment than during more typical dam operation and can generate a turbidity plume in Lake Havasu. The intakes for the Central Arizona Project, which transfers Colorado River water to central and southern Arizona, are near the mouth of the Bill Williams River. Measurement of the turbidity and the development of the plume over time consequently were important components of the study. In this report, the measurements of suspended sediment concentration and discharges in the Bill Williams River and of turbidity in Lake Havasu are presented along with calculations of silt and sand loads in the Bill Williams River. Sediment concentrations were varied and likely dependent on a variable supply. Sediment loads were calculated at the mouth of the river and near Planet, about 10 km upstream from the mouth for the 2005 release, and they indicate that a net increase in transport of silt and a net decrease in the transport of sand occurred in the reach between the two sites.

  7. Analytic continuation of quantum Monte Carlo data by stochastic analytical inference.

    PubMed

    Fuchs, Sebastian; Pruschke, Thomas; Jarrell, Mark

    2010-05-01

    We present an algorithm for the analytic continuation of imaginary-time quantum Monte Carlo data which is strictly based on principles of Bayesian statistical inference. Within this framework we are able to obtain an explicit expression for the calculation of a weighted average over possible energy spectra, which can be evaluated by standard Monte Carlo simulations, yielding as by-product also the distribution function as function of the regularization parameter. Our algorithm thus avoids the usual ad hoc assumptions introduced in similar algorithms to fix the regularization parameter. We apply the algorithm to imaginary-time quantum Monte Carlo data and compare the resulting energy spectra with those from a standard maximum-entropy calculation.

  8. Preliminary volcanic hazards evaluation for Los Alamos National Laboratory Facilities and Operations : current state of knowledge and proposed path forward

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Keating, Gordon N.; Schultz-Fellenz, Emily S.; Miller, Elizabeth D.

    2010-09-01

    The integration of available information on the volcanic history of the region surrounding Los Alamos National Laboratory indicates that the Laboratory is at risk from volcanic hazards. Volcanism in the vicinity of the Laboratory is unlikely within the lifetime of the facility (ca. 50–100 years) but cannot be ruled out. This evaluation provides a preliminary estimate of recurrence rates for volcanic activity. If further assessment of the hazard is deemed beneficial to reduce risk uncertainty, the next step would be to convene a formal probabilistic volcanic hazards assessment.

  9. Electric fields, electron production, and electron motion at the stripper foil in the Los Alamos Proton Storage Ring

    NASA Astrophysics Data System (ADS)

    Plum, M.

    The beam instability at the Los Alamos Proton Storage Ring (PSR) most likely involves coupled oscillations between electrons and protons. For this instability to occur, there must be a strong source of electrons. Investigation of the various sources of electrons in the PSR had begun. Copious electron production is expected in the injection section because this section contains the stripper foil. This foil is mounted near the center of the beam pipe, and both circulating and injected protons pass through it, thus allowing ample opportunity for electron production. This paper discusses various mechanisms for electron production, beam-induced electric fields, and electron motion in the vicinity of the foil.

  10. Evaluation of the Likelihood for Thermal Runaway for Nitrate Salt Containers in Storage at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Heatwole, Eric Mann; Gunderson, Jake Alfred; Parker, Gary Robert

    2016-03-25

    In order to handle and process the existing Los Alamos National Laboratory (LANL) Nitrate Salt drums it is necessary to quantify the risk. One of the most obvious dangers is a repeat of the original violent reaction (2015), which would endanger nearby workers, not only with radioactive contamination, but also with large amounts of heat, dangerous corrosive gases and the physical dangers associated with a bursting drum. If there still existed a high probability of violent reaction, then these drums should only be accessed remotely. The objective of the work reported herein is to determine the likelihood of a similarmore » violent event occurring.« less

  11. Monte Carlo Simulation of THz Multipliers

    NASA Technical Reports Server (NTRS)

    East, J.; Blakey, P.

    1997-01-01

    Schottky Barrier diode frequency multipliers are critical components in submillimeter and Thz space based earth observation systems. As the operating frequency of these multipliers has increased, the agreement between design predictions and experimental results has become poorer. The multiplier design is usually based on a nonlinear model using a form of harmonic balance and a model for the Schottky barrier diode. Conventional voltage dependent lumped element models do a poor job of predicting THz frequency performance. This paper will describe a large signal Monte Carlo simulation of Schottky barrier multipliers. The simulation is a time dependent particle field Monte Carlo simulation with ohmic and Schottky barrier boundary conditions included that has been combined with a fixed point solution for the nonlinear circuit interaction. The results in the paper will point out some important time constants in varactor operation and will describe the effects of current saturation and nonlinear resistances on multiplier operation.

  12. Exact Dynamics via Poisson Process: a unifying Monte Carlo paradigm

    NASA Astrophysics Data System (ADS)

    Gubernatis, James

    2014-03-01

    A common computational task is solving a set of ordinary differential equations (o.d.e.'s). A little known theorem says that the solution of any set of o.d.e.'s is exactly solved by the expectation value over a set of arbitary Poisson processes of a particular function of the elements of the matrix that defines the o.d.e.'s. The theorem thus provides a new starting point to develop real and imaginary-time continous-time solvers for quantum Monte Carlo algorithms, and several simple observations enable various quantum Monte Carlo techniques and variance reduction methods to transfer to a new context. I will state the theorem, note a transformation to a very simple computational scheme, and illustrate the use of some techniques from the directed-loop algorithm in context of the wavefunction Monte Carlo method that is used to solve the Lindblad master equation for the dynamics of open quantum systems. I will end by noting that as the theorem does not depend on the source of the o.d.e.'s coming from quantum mechanics, it also enables the transfer of continuous-time methods from quantum Monte Carlo to the simulation of various classical equations of motion heretofore only solved deterministically.

  13. Monte Carlo simulation of proton track structure in biological matter

    DOE PAGES

    Quinto, Michele A.; Monti, Juan M.; Weck, Philippe F.; ...

    2017-05-25

    Here, understanding the radiation-induced effects at the cellular and subcellular levels remains crucial for predicting the evolution of irradiated biological matter. In this context, Monte Carlo track-structure simulations have rapidly emerged among the most suitable and powerful tools. However, most existing Monte Carlo track-structure codes rely heavily on the use of semi-empirical cross sections as well as water as a surrogate for biological matter. In the current work, we report on the up-to-date version of our homemade Monte Carlo code TILDA-V – devoted to the modeling of the slowing-down of 10 keV–100 MeV protons in both water and DNA –more » where the main collisional processes are described by means of an extensive set of ab initio differential and total cross sections.« less

  14. Monte Carlo-based Reconstruction in Water Cherenkov Detectors using Chroma

    NASA Astrophysics Data System (ADS)

    Seibert, Stanley; Latorre, Anthony

    2012-03-01

    We demonstrate the feasibility of event reconstruction---including position, direction, energy and particle identification---in water Cherenkov detectors with a purely Monte Carlo-based method. Using a fast optical Monte Carlo package we have written, called Chroma, in combination with several variance reduction techniques, we can estimate the value of a likelihood function for an arbitrary event hypothesis. The likelihood can then be maximized over the parameter space of interest using a form of gradient descent designed for stochastic functions. Although slower than more traditional reconstruction algorithms, this completely Monte Carlo-based technique is universal and can be applied to a detector of any size or shape, which is a major advantage during the design phase of an experiment. As a specific example, we focus on reconstruction results from a simulation of the 200 kiloton water Cherenkov far detector option for LBNE.

  15. Monte Carlo simulation of proton track structure in biological matter

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Quinto, Michele A.; Monti, Juan M.; Weck, Philippe F.

    Here, understanding the radiation-induced effects at the cellular and subcellular levels remains crucial for predicting the evolution of irradiated biological matter. In this context, Monte Carlo track-structure simulations have rapidly emerged among the most suitable and powerful tools. However, most existing Monte Carlo track-structure codes rely heavily on the use of semi-empirical cross sections as well as water as a surrogate for biological matter. In the current work, we report on the up-to-date version of our homemade Monte Carlo code TILDA-V – devoted to the modeling of the slowing-down of 10 keV–100 MeV protons in both water and DNA –more » where the main collisional processes are described by means of an extensive set of ab initio differential and total cross sections.« less

  16. Ex Post Facto Monte Carlo Variance Reduction

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Booth, Thomas E.

    The variance in Monte Carlo particle transport calculations is often dominated by a few particles whose importance increases manyfold on a single transport step. This paper describes a novel variance reduction method that uses a large importance change as a trigger to resample the offending transport step. That is, the method is employed only after (ex post facto) a random walk attempts a transport step that would otherwise introduce a large variance in the calculation.Improvements in two Monte Carlo transport calculations are demonstrated empirically using an ex post facto method. First, the method is shown to reduce the variance inmore » a penetration problem with a cross-section window. Second, the method empirically appears to modify a point detector estimator from an infinite variance estimator to a finite variance estimator.« less

  17. 78 FR 14031 - Proposed Amendment of Class D Airspace; El Monte, CA

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-03-04

    ...: Federal Aviation Administration (FAA), DOT. ACTION: Notice of proposed rulemaking (NPRM). SUMMARY: This action proposes to amend Class D Airspace at El Monte Airport, El Monte, CA. This action, initiated by... particularly helpful in developing reasoned regulatory decisions on the proposal. Comments are specifically...

  18. Los Alamos National Laboratory considers the use of biodiesel.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Matlin, M. K.

    2002-01-01

    A new EPA-approved alternative fuel, called biodiesel, may soon be used at Los Alamos National Laboratory in everything from diesel trucks to laboratory equipment. Biodiesel transforms vegetable oils into a renewable, cleaner energy source that can be used in any machinery that uses diesel fuel. For the past couple years, the Laboratory has been exploring the possibility of switching over to soybean-based biodiesel. This change could lead to many health and environmental benefits, as well as help reduce the nation's dependence on foreign oil. Biodiesel is a clean, renewable diesel fuel substitute made from soybean and other vegetable oil crops,more » as well as from recycled cooking oils. A chemical process breaks down the vegetable oil into a usable form. Vegetable oil has a chain of about 18 carbons and ordinary diesel has about 12 or 13 carbons. The process breaks the carbon chains of the vegetable oil and separates out the glycerin (a fatty substance used in creams and soaps). The co-product of glycerin can be used by pharmaceutical and cosmetic companies, as well as many other markets. Once the chains are shortened and the glycerin is removed from the oil, the remaining liquid is similar to petroleum diesel fuel. It can be burned in pure form or in a blend of any proportion with petroleum diesel. To be considered an alternative fuel source by the EPA, the blend must be at least 20 percent biodiesel (B20). According to the U.S. Department of Energy (DOE), biodiesel is America's fastest growing alternative fuel.« less

  19. 2011 Los Alamos National Laboratory Riparian Inventory Results

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Norris, Elizabeth J.; Hansen, Leslie A.; Hathcock, Charles D.

    A total length of 36.7 kilometers of riparian habitat were inventoried within LANL boundaries between 2007 and 2011. The following canyons and lengths of riparian habitat were surveyed and inventoried between 2007 and 2011. Water Canyon (9,669 m), Los Alamos Canyon (7,131 m), Pajarito Canyon (6,009 m), Mortandad Canyon (3,110 m), Two-Mile Canyon (2,680 m), Sandia Canyon (2,181 m), Three-Mile Canyon (1,883 m), Canyon de Valle (1,835 m), Ancho Canyon (1,143 m), Canada del Buey (700 m), Sandia Canyon (221 m), DP Canyon (159 m) and Chaquehui Canyon (50 m). Effluent Canyon, Fence Canyon and Potrillo Canyon were surveyed butmore » no areas of riparian habitat were found. Stretches of inventoried riparian habitat were classified for prioritization of treatment, if any was recommended. High priority sites included stretches of Mortandad Canyon, LA Canyon, Pajarito Canyon, Two-Mile Canyon, Sandia Canyon and Water Canyon. Recommended treatment for high priority sites includes placement of objects into the stream channel to encourage sediment deposition, elimination of channel incision, and to expand and slow water flow across the floodplain. Additional stretches were classified as lower priority, and, for other sites it was recommended that feral cattle and exotic plants be removed to aid in riparian habitat recovery. In June 2011 the Las Conchas Wildfire burned over 150,000 acres of land in the Jemez Mountains and surrounding areas. The watersheds above LA Canyon, Water Canyon and Pajarito Canyon were burned in the Las Conchas Wildfire and flooding and habitat alteration were observed in these canyon bottoms (Wright 2011). Post fire status of lower priority areas may change to higher priority for some of the sites surveyed prior to the Las Conchas Wildfire, due to changes in vegetation cover in the adjacent upland watershed.« less

  20. Study of the Transition Flow Regime using Monte Carlo Methods

    NASA Technical Reports Server (NTRS)

    Hassan, H. A.

    1999-01-01

    This NASA Cooperative Agreement presents a study of the Transition Flow Regime Using Monte Carlo Methods. The topics included in this final report are: 1) New Direct Simulation Monte Carlo (DSMC) procedures; 2) The DS3W and DS2A Programs; 3) Papers presented; 4) Miscellaneous Applications and Program Modifications; 5) Solution of Transitional Wake Flows at Mach 10; and 6) Turbulence Modeling of Shock-Dominated Fows with a k-Enstrophy Formulation.

  1. Monte Carlo charged-particle tracking and energy deposition on a Lagrangian mesh.

    PubMed

    Yuan, J; Moses, G A; McKenty, P W

    2005-10-01

    A Monte Carlo algorithm for alpha particle tracking and energy deposition on a cylindrical computational mesh in a Lagrangian hydrodynamics code used for inertial confinement fusion (ICF) simulations is presented. The straight line approximation is used to follow propagation of "Monte Carlo particles" which represent collections of alpha particles generated from thermonuclear deuterium-tritium (DT) reactions. Energy deposition in the plasma is modeled by the continuous slowing down approximation. The scheme addresses various aspects arising in the coupling of Monte Carlo tracking with Lagrangian hydrodynamics; such as non-orthogonal severely distorted mesh cells, particle relocation on the moving mesh and particle relocation after rezoning. A comparison with the flux-limited multi-group diffusion transport method is presented for a polar direct drive target design for the National Ignition Facility. Simulations show the Monte Carlo transport method predicts about earlier ignition than predicted by the diffusion method, and generates higher hot spot temperature. Nearly linear speed-up is achieved for multi-processor parallel simulations.

  2. Collision of Physics and Software in the Monte Carlo Application Toolkit (MCATK)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sweezy, Jeremy Ed

    2016-01-21

    The topic is presented in a series of slides organized as follows: MCATK overview, development strategy, available algorithms, problem modeling (sources, geometry, data, tallies), parallelism, miscellaneous tools/features, example MCATK application, recent areas of research, and summary and future work. MCATK is a C++ component-based Monte Carlo neutron-gamma transport software library with continuous energy neutron and photon transport. Designed to build specialized applications and to provide new functionality in existing general-purpose Monte Carlo codes like MCNP, it reads ACE formatted nuclear data generated by NJOY. The motivation behind MCATK was to reduce costs. MCATK physics involves continuous energy neutron & gammamore » transport with multi-temperature treatment, static eigenvalue (k eff and α) algorithms, time-dependent algorithm, and fission chain algorithms. MCATK geometry includes mesh geometries and solid body geometries. MCATK provides verified, unit-test Monte Carlo components, flexibility in Monte Carlo application development, and numerous tools such as geometry and cross section plotters.« less

  3. Teaching Ionic Solvation Structure with a Monte Carlo Liquid Simulation Program

    ERIC Educational Resources Information Center

    Serrano, Agostinho; Santos, Flavia M. T.; Greca, Ileana M.

    2004-01-01

    The use of molecular dynamics and Monte Carlo methods has provided efficient means to stimulate the behavior of molecular liquids and solutions. A Monte Carlo simulation program is used to compute the structure of liquid water and of water as a solvent to Na(super +), Cl(super -), and Ar on a personal computer to show that it is easily feasible to…

  4. Calculation of radiation therapy dose using all particle Monte Carlo transport

    DOEpatents

    Chandler, William P.; Hartmann-Siantar, Christine L.; Rathkopf, James A.

    1999-01-01

    The actual radiation dose absorbed in the body is calculated using three-dimensional Monte Carlo transport. Neutrons, protons, deuterons, tritons, helium-3, alpha particles, photons, electrons, and positrons are transported in a completely coupled manner, using this Monte Carlo All-Particle Method (MCAPM). The major elements of the invention include: computer hardware, user description of the patient, description of the radiation source, physical databases, Monte Carlo transport, and output of dose distributions. This facilitated the estimation of dose distributions on a Cartesian grid for neutrons, photons, electrons, positrons, and heavy charged-particles incident on any biological target, with resolutions ranging from microns to centimeters. Calculations can be extended to estimate dose distributions on general-geometry (non-Cartesian) grids for biological and/or non-biological media.

  5. Calculation of radiation therapy dose using all particle Monte Carlo transport

    DOEpatents

    Chandler, W.P.; Hartmann-Siantar, C.L.; Rathkopf, J.A.

    1999-02-09

    The actual radiation dose absorbed in the body is calculated using three-dimensional Monte Carlo transport. Neutrons, protons, deuterons, tritons, helium-3, alpha particles, photons, electrons, and positrons are transported in a completely coupled manner, using this Monte Carlo All-Particle Method (MCAPM). The major elements of the invention include: computer hardware, user description of the patient, description of the radiation source, physical databases, Monte Carlo transport, and output of dose distributions. This facilitated the estimation of dose distributions on a Cartesian grid for neutrons, photons, electrons, positrons, and heavy charged-particles incident on any biological target, with resolutions ranging from microns to centimeters. Calculations can be extended to estimate dose distributions on general-geometry (non-Cartesian) grids for biological and/or non-biological media. 57 figs.

  6. Active damping of the e-p instability at the Los Alamos Proton Storage Ring

    NASA Astrophysics Data System (ADS)

    Macek, R. J.; Assadi, S.; Byrd, J. M.; Deibele, C. E.; Henderson, S. D.; Lee, S. Y.; McCrady, R. C.; Pivi, M. F. T.; Plum, M. A.; Walbridge, S. B.; Zaugg, T. J.

    2007-12-01

    A prototype of an analog, transverse (vertical) feedback system for active damping of the two-stream (e-p) instability has been developed and successfully tested at the Los Alamos Proton Storage Ring (PSR). This system was able to improve the instability threshold by approximately 30% (as measured by the change in RF buncher voltage at instability threshold). The feedback system configuration, setup procedures, and optimization of performance are described. Results of several experimental tests of system performance are presented including observations of instability threshold improvement and grow-damp experiments, which yield estimates of instability growth and damping rates. A major effort was undertaken to identify and study several factors limiting system performance. Evidence obtained from these tests suggests that performance of the prototype was limited by higher instability growth rates arising from beam leakage into the gap at lower RF buncher voltage and the onset of instability in the horizontal plane, which had no feedback.

  7. Applying Quantum Monte Carlo to the Electronic Structure Problem

    NASA Astrophysics Data System (ADS)

    Powell, Andrew D.; Dawes, Richard

    2016-06-01

    Two distinct types of Quantum Monte Carlo (QMC) calculations are applied to electronic structure problems such as calculating potential energy curves and producing benchmark values for reaction barriers. First, Variational and Diffusion Monte Carlo (VMC and DMC) methods using a trial wavefunction subject to the fixed node approximation were tested using the CASINO code.[1] Next, Full Configuration Interaction Quantum Monte Carlo (FCIQMC), along with its initiator extension (i-FCIQMC) were tested using the NECI code.[2] FCIQMC seeks the FCI energy for a specific basis set. At a reduced cost, the efficient i-FCIQMC method can be applied to systems in which the standard FCIQMC approach proves to be too costly. Since all of these methods are statistical approaches, uncertainties (error-bars) are introduced for each calculated energy. This study tests the performance of the methods relative to traditional quantum chemistry for some benchmark systems. References: [1] R. J. Needs et al., J. Phys.: Condensed Matter 22, 023201 (2010). [2] G. H. Booth et al., J. Chem. Phys. 131, 054106 (2009).

  8. A New Monte Carlo Method for Estimating Marginal Likelihoods.

    PubMed

    Wang, Yu-Bo; Chen, Ming-Hui; Kuo, Lynn; Lewis, Paul O

    2018-06-01

    Evaluating the marginal likelihood in Bayesian analysis is essential for model selection. Estimators based on a single Markov chain Monte Carlo sample from the posterior distribution include the harmonic mean estimator and the inflated density ratio estimator. We propose a new class of Monte Carlo estimators based on this single Markov chain Monte Carlo sample. This class can be thought of as a generalization of the harmonic mean and inflated density ratio estimators using a partition weighted kernel (likelihood times prior). We show that our estimator is consistent and has better theoretical properties than the harmonic mean and inflated density ratio estimators. In addition, we provide guidelines on choosing optimal weights. Simulation studies were conducted to examine the empirical performance of the proposed estimator. We further demonstrate the desirable features of the proposed estimator with two real data sets: one is from a prostate cancer study using an ordinal probit regression model with latent variables; the other is for the power prior construction from two Eastern Cooperative Oncology Group phase III clinical trials using the cure rate survival model with similar objectives.

  9. Shielding calculations and verifications for the new Radiation Instrument Calibration Facility at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    George, G. L.; Olsher, R. H.; Seagraves, D. T.

    2002-01-01

    MCNP-4C1 was used to perform the shielding design for the new Central Health Physics Calibration Facility (CHPCF) at Los Alamos National Laboratory (LANL). The problem of shielding the facility was subdivided into three separate components: (1) Transmission; (2) Skyshine; and (3) Maze Streaming/ Transmission. When possible, actual measurements were taken to verify calculation results. The comparison of calculation versus measurement results shows excellent agreement for neutron calculations. For photon comparisons, calculations resulted in conservative estimates of the Effective Dose Equivalent (EDE) compared to measured results. This disagreement in the photon measurements versus calculations is most likely due to several conservativemore » assumptions regarding shield density and composition. For example, reinforcing steel bars (Rebar) in the concrete shield walls were not included in the shield model.« less

  10. Studies of fission fragment properties at the Los Alamos Neutron Science Center (LANSCE)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tovesson, Fredrik; Mayorov, Dmitriy; Duke, Dana

    Nuclear data related to the fission process are needed for a wide variety of research areas, including fundamental science, nuclear energy and non-proliferation. While some of the relevant data have been measured to the required accuracies there are still many aspects of fission that need further investigation. One such aspect is how Total Kinetic Energy (TKE), fragment yields, angular distributions and other fission observables depend on excitation energy of the fissioning system. Another question is the correlation between mass, charge and energy of fission fragments. At the Los Alamos Neutron Science Center (LANSCE) we are studying neutron-induced fission at incidentmore » energies from thermal up to hundreds of MeV using the Lujan Center and Weapons Neutron Research (WNR) facilities. Advanced instruments such as SPIDER (time-of-flight and kinetic energy spectrometer), the NIFFTE Time Projection Chamber (TPC), and Frisch grid Ionization Chambers (FGIC) are used to investigate the properties of fission fragments, and some important results for the major actinides have been obtained.« less

  11. Studies of fission fragment properties at the Los Alamos Neutron Science Center (LANSCE)

    DOE PAGES

    Tovesson, Fredrik; Mayorov, Dmitriy; Duke, Dana; ...

    2017-09-13

    Nuclear data related to the fission process are needed for a wide variety of research areas, including fundamental science, nuclear energy and non-proliferation. While some of the relevant data have been measured to the required accuracies there are still many aspects of fission that need further investigation. One such aspect is how Total Kinetic Energy (TKE), fragment yields, angular distributions and other fission observables depend on excitation energy of the fissioning system. Another question is the correlation between mass, charge and energy of fission fragments. At the Los Alamos Neutron Science Center (LANSCE) we are studying neutron-induced fission at incidentmore » energies from thermal up to hundreds of MeV using the Lujan Center and Weapons Neutron Research (WNR) facilities. Advanced instruments such as SPIDER (time-of-flight and kinetic energy spectrometer), the NIFFTE Time Projection Chamber (TPC), and Frisch grid Ionization Chambers (FGIC) are used to investigate the properties of fission fragments, and some important results for the major actinides have been obtained.« less

  12. Delayed Slater determinant update algorithms for high efficiency quantum Monte Carlo.

    PubMed

    McDaniel, T; D'Azevedo, E F; Li, Y W; Wong, K; Kent, P R C

    2017-11-07

    Within ab initio Quantum Monte Carlo simulations, the leading numerical cost for large systems is the computation of the values of the Slater determinants in the trial wavefunction. Each Monte Carlo step requires finding the determinant of a dense matrix. This is most commonly iteratively evaluated using a rank-1 Sherman-Morrison updating scheme to avoid repeated explicit calculation of the inverse. The overall computational cost is, therefore, formally cubic in the number of electrons or matrix size. To improve the numerical efficiency of this procedure, we propose a novel multiple rank delayed update scheme. This strategy enables probability evaluation with an application of accepted moves to the matrices delayed until after a predetermined number of moves, K. The accepted events are then applied to the matrices en bloc with enhanced arithmetic intensity and computational efficiency via matrix-matrix operations instead of matrix-vector operations. This procedure does not change the underlying Monte Carlo sampling or its statistical efficiency. For calculations on large systems and algorithms such as diffusion Monte Carlo, where the acceptance ratio is high, order of magnitude improvements in the update time can be obtained on both multi-core central processing units and graphical processing units.

  13. Delayed Slater determinant update algorithms for high efficiency quantum Monte Carlo

    NASA Astrophysics Data System (ADS)

    McDaniel, T.; D'Azevedo, E. F.; Li, Y. W.; Wong, K.; Kent, P. R. C.

    2017-11-01

    Within ab initio Quantum Monte Carlo simulations, the leading numerical cost for large systems is the computation of the values of the Slater determinants in the trial wavefunction. Each Monte Carlo step requires finding the determinant of a dense matrix. This is most commonly iteratively evaluated using a rank-1 Sherman-Morrison updating scheme to avoid repeated explicit calculation of the inverse. The overall computational cost is, therefore, formally cubic in the number of electrons or matrix size. To improve the numerical efficiency of this procedure, we propose a novel multiple rank delayed update scheme. This strategy enables probability evaluation with an application of accepted moves to the matrices delayed until after a predetermined number of moves, K. The accepted events are then applied to the matrices en bloc with enhanced arithmetic intensity and computational efficiency via matrix-matrix operations instead of matrix-vector operations. This procedure does not change the underlying Monte Carlo sampling or its statistical efficiency. For calculations on large systems and algorithms such as diffusion Monte Carlo, where the acceptance ratio is high, order of magnitude improvements in the update time can be obtained on both multi-core central processing units and graphical processing units.

  14. Path integral Monte Carlo ground state approach: formalism, implementation, and applications

    NASA Astrophysics Data System (ADS)

    Yan, Yangqian; Blume, D.

    2017-11-01

    Monte Carlo techniques have played an important role in understanding strongly correlated systems across many areas of physics, covering a wide range of energy and length scales. Among the many Monte Carlo methods applicable to quantum mechanical systems, the path integral Monte Carlo approach with its variants has been employed widely. Since semi-classical or classical approaches will not be discussed in this review, path integral based approaches can for our purposes be divided into two categories: approaches applicable to quantum mechanical systems at zero temperature and approaches applicable to quantum mechanical systems at finite temperature. While these two approaches are related to each other, the underlying formulation and aspects of the algorithm differ. This paper reviews the path integral Monte Carlo ground state (PIGS) approach, which solves the time-independent Schrödinger equation. Specifically, the PIGS approach allows for the determination of expectation values with respect to eigen states of the few- or many-body Schrödinger equation provided the system Hamiltonian is known. The theoretical framework behind the PIGS algorithm, implementation details, and sample applications for fermionic systems are presented.

  15. Monte Carlo method for calculating the radiation skyshine produced by electron accelerators

    NASA Astrophysics Data System (ADS)

    Kong, Chaocheng; Li, Quanfeng; Chen, Huaibi; Du, Taibin; Cheng, Cheng; Tang, Chuanxiang; Zhu, Li; Zhang, Hui; Pei, Zhigang; Ming, Shenjin

    2005-06-01

    Using the MCNP4C Monte Carlo code, the X-ray skyshine produced by 9 MeV, 15 MeV and 21 MeV electron linear accelerators were calculated respectively with a new two-step method combined with the split and roulette variance reduction technique. Results of the Monte Carlo simulation, the empirical formulas used for skyshine calculation and the dose measurements were analyzed and compared. In conclusion, the skyshine dose measurements agreed reasonably with the results computed by the Monte Carlo method, but deviated from computational results given by empirical formulas. The effect on skyshine dose caused by different structures of accelerator head is also discussed in this paper.

  16. [Accuracy Check of Monte Carlo Simulation in Particle Therapy Using Gel Dosimeters].

    PubMed

    Furuta, Takuya

    2017-01-01

    Gel dosimeters are a three-dimensional imaging tool for dose distribution induced by radiations. They can be used for accuracy check of Monte Carlo simulation in particle therapy. An application was reviewed in this article. An inhomogeneous biological sample placing a gel dosimeter behind it was irradiated by carbon beam. The recorded dose distribution in the gel dosimeter reflected the inhomogeneity of the biological sample. Monte Carlo simulation was conducted by reconstructing the biological sample from its CT image. The accuracy of the particle transport by Monte Carlo simulation was checked by comparing the dose distribution in the gel dosimeter between simulation and experiment.

  17. Instantons in Quantum Annealing: Thermally Assisted Tunneling Vs Quantum Monte Carlo Simulations

    NASA Technical Reports Server (NTRS)

    Jiang, Zhang; Smelyanskiy, Vadim N.; Boixo, Sergio; Isakov, Sergei V.; Neven, Hartmut; Mazzola, Guglielmo; Troyer, Matthias

    2015-01-01

    Recent numerical result (arXiv:1512.02206) from Google suggested that the D-Wave quantum annealer may have an asymptotic speed-up than simulated annealing, however, the asymptotic advantage disappears when it is compared to quantum Monte Carlo (a classical algorithm despite its name). We show analytically that the asymptotic scaling of quantum tunneling is exactly the same as the escape rate in quantum Monte Carlo for a class of problems. Thus, the Google result might be explained in our framework. We also found that the transition state in quantum Monte Carlo corresponds to the instanton solution in quantum tunneling problems, which is observed in numerical simulations.

  18. CloudMC: a cloud computing application for Monte Carlo simulation.

    PubMed

    Miras, H; Jiménez, R; Miras, C; Gomà, C

    2013-04-21

    This work presents CloudMC, a cloud computing application-developed in Windows Azure®, the platform of the Microsoft® cloud-for the parallelization of Monte Carlo simulations in a dynamic virtual cluster. CloudMC is a web application designed to be independent of the Monte Carlo code in which the simulations are based-the simulations just need to be of the form: input files → executable → output files. To study the performance of CloudMC in Windows Azure®, Monte Carlo simulations with penelope were performed on different instance (virtual machine) sizes, and for different number of instances. The instance size was found to have no effect on the simulation runtime. It was also found that the decrease in time with the number of instances followed Amdahl's law, with a slight deviation due to the increase in the fraction of non-parallelizable time with increasing number of instances. A simulation that would have required 30 h of CPU on a single instance was completed in 48.6 min when executed on 64 instances in parallel (speedup of 37 ×). Furthermore, the use of cloud computing for parallel computing offers some advantages over conventional clusters: high accessibility, scalability and pay per usage. Therefore, it is strongly believed that cloud computing will play an important role in making Monte Carlo dose calculation a reality in future clinical practice.

  19. Probabilistic learning of nonlinear dynamical systems using sequential Monte Carlo

    NASA Astrophysics Data System (ADS)

    Schön, Thomas B.; Svensson, Andreas; Murray, Lawrence; Lindsten, Fredrik

    2018-05-01

    Probabilistic modeling provides the capability to represent and manipulate uncertainty in data, models, predictions and decisions. We are concerned with the problem of learning probabilistic models of dynamical systems from measured data. Specifically, we consider learning of probabilistic nonlinear state-space models. There is no closed-form solution available for this problem, implying that we are forced to use approximations. In this tutorial we will provide a self-contained introduction to one of the state-of-the-art methods-the particle Metropolis-Hastings algorithm-which has proven to offer a practical approximation. This is a Monte Carlo based method, where the particle filter is used to guide a Markov chain Monte Carlo method through the parameter space. One of the key merits of the particle Metropolis-Hastings algorithm is that it is guaranteed to converge to the "true solution" under mild assumptions, despite being based on a particle filter with only a finite number of particles. We will also provide a motivating numerical example illustrating the method using a modeling language tailored for sequential Monte Carlo methods. The intention of modeling languages of this kind is to open up the power of sophisticated Monte Carlo methods-including particle Metropolis-Hastings-to a large group of users without requiring them to know all the underlying mathematical details.

  20. MC3: Multi-core Markov-chain Monte Carlo code

    NASA Astrophysics Data System (ADS)

    Cubillos, Patricio; Harrington, Joseph; Lust, Nate; Foster, AJ; Stemm, Madison; Loredo, Tom; Stevenson, Kevin; Campo, Chris; Hardin, Matt; Hardy, Ryan

    2016-10-01

    MC3 (Multi-core Markov-chain Monte Carlo) is a Bayesian statistics tool that can be executed from the shell prompt or interactively through the Python interpreter with single- or multiple-CPU parallel computing. It offers Markov-chain Monte Carlo (MCMC) posterior-distribution sampling for several algorithms, Levenberg-Marquardt least-squares optimization, and uniform non-informative, Jeffreys non-informative, or Gaussian-informative priors. MC3 can share the same value among multiple parameters and fix the value of parameters to constant values, and offers Gelman-Rubin convergence testing and correlated-noise estimation with time-averaging or wavelet-based likelihood estimation methods.

  1. Perturbative two- and three-loop coefficients from large β Monte Carlo

    NASA Astrophysics Data System (ADS)

    Lepage, G. P.; Mackenzie, P. B.; Shakespeare, N. H.; Trottier, H. D.

    Perturbative coefficients for Wilson loops and the static quark self-energy are extracted from Monte Carlo simulations at large β on finite volumes, where all the lattice momenta are large. The Monte Carlo results are in excellent agreement with perturbation theory through second order. New results for third order coefficients are reported. Twisted boundary conditions are used to eliminate zero modes and to suppress Z3 tunneling.

  2. A model for microbially induced precipitation of vadose-zone calcites in fractures at LOS Alamos, New Mexico, USA

    NASA Astrophysics Data System (ADS)

    Newman, Brent D.; Campbell, Andrew R.; Norman, David I.; Ringelberg, David B.

    1997-05-01

    Fractures are unique environments that can concentrate the flow of water, nutrients, and contaminants. As such, fractures play an important role in controlling the flux of various substances into and through the vadose zone. Calcite fracture fillings are present in the near surface in the Bandelier Tuff Formation at Los Alamos, New Mexico, and provide a record of the geochemical and hydrologic processes that have occurred in fractures. The objective of this study was to examine calcite fracture fills in order to improve understanding of processes within fractures, and in particular those that lead to precipitation of calcite. Samples of calcite fillings were collected from vertical and horizontal fractures exposed in a shallow waste-burial pit. Scanning electron microscopy show morphologies which suggest that plants, fungi, and bacteria were important in the precipitation process. Quadrupole mass spectrometric analyses of fluid inclusion gases show predominantly methane (17-99%) and little to no oxygen (0-8%), suggesting the development of anaerobic conditions in the fractures. Ester-linked phospholipid biomarkers are evidence for a diverse microbial community in the fractures, and the presence of di-ether lipids indicate that the methane was generated by anaerobic bacteria. The calcite fillings apparently resulted from multiple biological and chemical processes in which plant roots in the fractures were converted to calcite. Roots grew into the fractures, eventually died, and were decomposed by bacteria and fungi. Anaerobic gases were generated from encapsulated organic material within the calcite via microbial decomposition, or were generated by microbes simultaneously with calcite precipitation. It is likely that the biological controls on calcite formation that occurred in the Los Alamos fractures also occurs in soils, and may explain the occurrence of other types of pedogenic calcites.

  3. Archaeological Investigations at Site 45-DO-285, Chief Joseph Dam Project, Washington.

    DTIC Science & Technology

    1984-01-01

    systematic design. Sampling strata were created by dividing the site Into seven sets of grid units, each composed of 25 2 x 2-in units arranged in squares...NISP=23) C-r-s-my pJc (painted turtle) -- 23 elements. Painted turtle Is the only turtle currently I lving in the project area. Clemmys marmoraa

  4. Full Configuration Interaction Quantum Monte Carlo and Diffusion Monte Carlo: A Comparative Study of the 3D Homogeneous Electron Gas

    NASA Astrophysics Data System (ADS)

    Shepherd, James J.; López Ríos, Pablo; Needs, Richard J.; Drummond, Neil D.; Mohr, Jennifer A.-F.; Booth, George H.; Grüneis, Andreas; Kresse, Georg; Alavi, Ali

    2013-03-01

    Full configuration interaction quantum Monte Carlo1 (FCIQMC) and its initiator adaptation2 allow for exact solutions to the Schrödinger equation to be obtained within a finite-basis wavefunction ansatz. In this talk, we explore an application of FCIQMC to the homogeneous electron gas (HEG). In particular we use these exact finite-basis energies to compare with approximate quantum chemical calculations from the VASP code3. After removing the basis set incompleteness error by extrapolation4,5, we compare our energies with state-of-the-art diffusion Monte Carlo calculations from the CASINO package6. Using a combined approach of the two quantum Monte Carlo methods, we present the highest-accuracy thermodynamic (infinite-particle) limit energies for the HEG achieved to date. 1 G. H. Booth, A. Thom, and A. Alavi, J. Chem. Phys. 131, 054106 (2009). 2 D. Cleland, G. H. Booth, and A. Alavi, J. Chem. Phys. 132, 041103 (2010). 3 www.vasp.at (2012). 4 J. J. Shepherd, A. Grüneis, G. H. Booth, G. Kresse, and A. Alavi, Phys. Rev. B. 86, 035111 (2012). 5 J. J. Shepherd, G. H. Booth, and A. Alavi, J. Chem. Phys. 136, 244101 (2012). 6 R. Needs, M. Towler, N. Drummond, and P. L. Ríos, J. Phys.: Condensed Matter 22, 023201 (2010).

  5. Annual Report on the Activities and Publications of the DHS-DNDO-NTNFC Sponsored Post-doctoral Fellow at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Rim, Jung Ho; Tandon, Lav

    This report is a summary of the projects Jung Rim is working on as a DHS postdoctoral fellow at Los Alamos National Laboratory. These research projects are designed to explore different radioanalytical methods to support nuclear forensics applications. The current projects discussed here include development of alpha spectroscopy method for 240/239Pu Isotopic ratio measurement, non-destructive uranium assay method using gamma spectroscopy, and 236U non-destructive uranium analysis using FRAM code. This report documents the work that has been performed since the start of the postdoctoral appointment.

  6. Estimating statistical uncertainty of Monte Carlo efficiency-gain in the context of a correlated sampling Monte Carlo code for brachytherapy treatment planning with non-normal dose distribution.

    PubMed

    Mukhopadhyay, Nitai D; Sampson, Andrew J; Deniz, Daniel; Alm Carlsson, Gudrun; Williamson, Jeffrey; Malusek, Alexandr

    2012-01-01

    Correlated sampling Monte Carlo methods can shorten computing times in brachytherapy treatment planning. Monte Carlo efficiency is typically estimated via efficiency gain, defined as the reduction in computing time by correlated sampling relative to conventional Monte Carlo methods when equal statistical uncertainties have been achieved. The determination of the efficiency gain uncertainty arising from random effects, however, is not a straightforward task specially when the error distribution is non-normal. The purpose of this study is to evaluate the applicability of the F distribution and standardized uncertainty propagation methods (widely used in metrology to estimate uncertainty of physical measurements) for predicting confidence intervals about efficiency gain estimates derived from single Monte Carlo runs using fixed-collision correlated sampling in a simplified brachytherapy geometry. A bootstrap based algorithm was used to simulate the probability distribution of the efficiency gain estimates and the shortest 95% confidence interval was estimated from this distribution. It was found that the corresponding relative uncertainty was as large as 37% for this particular problem. The uncertainty propagation framework predicted confidence intervals reasonably well; however its main disadvantage was that uncertainties of input quantities had to be calculated in a separate run via a Monte Carlo method. The F distribution noticeably underestimated the confidence interval. These discrepancies were influenced by several photons with large statistical weights which made extremely large contributions to the scored absorbed dose difference. The mechanism of acquiring high statistical weights in the fixed-collision correlated sampling method was explained and a mitigation strategy was proposed. Copyright © 2011 Elsevier Ltd. All rights reserved.

  7. LMC: Logarithmantic Monte Carlo

    NASA Astrophysics Data System (ADS)

    Mantz, Adam B.

    2017-06-01

    LMC is a Markov Chain Monte Carlo engine in Python that implements adaptive Metropolis-Hastings and slice sampling, as well as the affine-invariant method of Goodman & Weare, in a flexible framework. It can be used for simple problems, but the main use case is problems where expensive likelihood evaluations are provided by less flexible third-party software, which benefit from parallelization across many nodes at the sampling level. The parallel/adaptive methods use communication through MPI, or alternatively by writing/reading files, and mostly follow the approaches pioneered by CosmoMC (ascl:1106.025).

  8. Structural testing of the Los Alamos National Laboratory Heat Source/Radioisotopic Thermoelectric Generator shipping container

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bronowski, D.R.; Madsen, M.M.

    The Heat Source/Radioisotopic Thermoelectric Generator shipping container is a Type B packaging design currently under development by Los Alamos National Laboratory. Type B packaging for transporting radioactive material is required to maintain containment and shielding after being exposed to the normal and hypothetical accident environments defined in Title 10 Code of Federal Regulations Part 71. A combination of testing and analysis is used to verify the adequacy of this package design. This report documents the test program portion of the design verification, using several prototype packages. Four types of testing were performed: 30-foot hypothetical accident condition drop tests in threemore » orientations, 40-inch hypothetical accident condition puncture tests in five orientations, a 21 psi external overpressure test, and a normal conditions of transport test consisting of a water spray and a 4 foot drop test. 18 refs., 104 figs., 13 tabs.« less

  9. Metallic lithium by quantum Monte Carlo

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sugiyama, G.; Zerah, G.; Alder, B.J.

    Lithium was chosen as the simplest known metal for the first application of quantum Monte Carlo methods in order to evaluate the accuracy of conventional one-electron band theories. Lithium has been extensively studied using such techniques. Band theory calculations have certain limitations in general and specifically in their application to lithium. Results depend on such factors as charge shape approximations (muffin tins), pseudopotentials (a special problem for lithium where the lack of rho core states requires a strong pseudopotential), and the form and parameters chosen for the exchange potential. The calculations are all one-electron methods in which the correlation effectsmore » are included in an ad hoc manner. This approximation may be particularly poor in the high compression regime, where the core states become delocalized. Furthermore, band theory provides only self-consistent results rather than strict limits on the energies. The quantum Monte Carlo method is a totally different technique using a many-body rather than a mean field approach which yields an upper bound on the energies. 18 refs., 4 figs., 1 tab.« less

  10. How Monte Carlo heuristics aid to identify the physical processes of drug release kinetics.

    PubMed

    Lecca, Paola

    2018-01-01

    We implement a Monte Carlo heuristic algorithm to model drug release from a solid dosage form. We show that with Monte Carlo simulations it is possible to identify and explain the causes of the unsatisfactory predictive power of current drug release models. It is well known that the power-law, the exponential models, as well as those derived from or inspired by them accurately reproduce only the first 60% of the release curve of a drug from a dosage form. In this study, by using Monte Carlo simulation approaches, we show that these models fit quite accurately almost the entire release profile when the release kinetics is not governed by the coexistence of different physico-chemical mechanisms. We show that the accuracy of the traditional models are comparable with those of Monte Carlo heuristics when these heuristics approximate and oversimply the phenomenology of drug release. This observation suggests to develop and use novel Monte Carlo simulation heuristics able to describe the complexity of the release kinetics, and consequently to generate data more similar to those observed in real experiments. Implementing Monte Carlo simulation heuristics of the drug release phenomenology may be much straightforward and efficient than hypothesizing and implementing from scratch complex mathematical models of the physical processes involved in drug release. Identifying and understanding through simulation heuristics what processes of this phenomenology reproduce the observed data and then formalize them in mathematics may allow avoiding time-consuming, trial-error based regression procedures. Three bullet points, highlighting the customization of the procedure. •An efficient heuristics based on Monte Carlo methods for simulating drug release from solid dosage form encodes is presented. It specifies the model of the physical process in a simple but accurate way in the formula of the Monte Carlo Micro Step (MCS) time interval.•Given the experimentally observed curve of

  11. Pesticides in Water and Suspended Sediment of the Alamo and New Rivers, Imperial Valley/Salton Sea Basin, California, 2006-2007

    USGS Publications Warehouse

    Orlando, James L.; Smalling, Kelly L.; Kuivila, Kathryn

    2008-01-01

    Water and suspended-sediment samples were collected at eight sites on the Alamo and New Rivers in the Imperial Valley/Salton Sea Basin of California and analyzed for both current-use and organochlorine pesticides by the U.S. Geological Survey. Samples were collected in the fall of 2006 and spring of 2007, corresponding to the seasons of greatest pesticide use in the basin. Large-volume water samples (up to 650 liters) were collected at each site and processed using a flow-through centrifuge to isolate suspended sediments. One-liter water samples were collected from the effluent of the centrifuge for the analysis of dissolved pesticides. Additional samples were collected for analysis of dissolved organic carbon and for suspended-sediment concentrations. Water samples were analyzed for a suite of 61 current-use and organochlorine pesticides using gas chromatography/mass spectrometry. A total of 25 pesticides were detected in the water samples, with seven pesticides detected in more than half of the samples. Dissolved concentrations of pesticides observed in this study ranged from below their respective method detection limits to 8,940 nanograms per liter (EPTC). The most frequently detected compounds in the water samples were chlorpyrifos, DCPA, EPTC, and trifluralin, which were observed in more than 75 percent of the samples. The maximum concentrations of most pesticides were detected in samples from the Alamo River. Maximum dissolved concentrations of carbofuran, chlorpyrifos, diazinon, and malathion exceeded aquatic life benchmarks established by the U.S. Environmental Protection Agency for these pesticides. Suspended sediments were analyzed for 87 current-use and organochlorine pesticides using microwave-assisted extraction, gel permeation chromatography for sulfur removal, and either carbon/alumina stacked solid-phase extraction cartridges or deactivated Florisil for removal of matrix interferences. Twenty current-use pesticides were detected in the suspended

  12. Radiotherapy Monte Carlo simulation using cloud computing technology.

    PubMed

    Poole, C M; Cornelius, I; Trapp, J V; Langton, C M

    2012-12-01

    Cloud computing allows for vast computational resources to be leveraged quickly and easily in bursts as and when required. Here we describe a technique that allows for Monte Carlo radiotherapy dose calculations to be performed using GEANT4 and executed in the cloud, with relative simulation cost and completion time evaluated as a function of machine count. As expected, simulation completion time decreases as 1/n for n parallel machines, and relative simulation cost is found to be optimal where n is a factor of the total simulation time in hours. Using the technique, we demonstrate the potential usefulness of cloud computing as a solution for rapid Monte Carlo simulation for radiotherapy dose calculation without the need for dedicated local computer hardware as a proof of principal.

  13. Monte Carlo modelling the dosimetric effects of electrode material on diamond detectors.

    PubMed

    Baluti, Florentina; Deloar, Hossain M; Lansley, Stuart P; Meyer, Juergen

    2015-03-01

    Diamond detectors for radiation dosimetry were modelled using the EGSnrc Monte Carlo code to investigate the influence of electrode material and detector orientation on the absorbed dose. The small dimensions of the electrode/diamond/electrode detector structure required very thin voxels and the use of non-standard DOSXYZnrc Monte Carlo model parameters. The interface phenomena was investigated by simulating a 6 MV beam and detectors with different electrode materials, namely Al, Ag, Cu and Au, with thickens of 0.1 µm for the electrodes and 0.1 mm for the diamond, in both perpendicular and parallel detector orientation with regards to the incident beam. The smallest perturbations were observed for the parallel detector orientation and Al electrodes (Z = 13). In summary, EGSnrc Monte Carlo code is well suited for modelling small detector geometries. The Monte Carlo model developed is a useful tool to investigate the dosimetric effects caused by different electrode materials. To minimise perturbations cause by the detector electrodes, it is recommended that the electrodes should be made from a low-atomic number material and placed parallel to the beam direction.

  14. Perturbative two- and three-loop coefficients from large b Monte Carlo

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    G.P. Lepage; P.B. Mackenzie; N.H. Shakespeare

    1999-10-18

    Perturbative coefficients for Wilson loops and the static quark self-energy are extracted from Monte Carlo simulations at large {beta} on finite volumes, where all the lattice momenta are large. The Monte Carlo results are in excellent agreement with perturbation theory through second order. New results for third order coefficients are reported. Twisted boundary conditions are used to eliminate zero modes and to suppress Z{sub 3} tunneling.

  15. Scalable Domain Decomposed Monte Carlo Particle Transport

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    O'Brien, Matthew Joseph

    2013-12-05

    In this dissertation, we present the parallel algorithms necessary to run domain decomposed Monte Carlo particle transport on large numbers of processors (millions of processors). Previous algorithms were not scalable, and the parallel overhead became more computationally costly than the numerical simulation.

  16. Pushing the limits of Monte Carlo simulations for the three-dimensional Ising model

    NASA Astrophysics Data System (ADS)

    Ferrenberg, Alan M.; Xu, Jiahao; Landau, David P.

    2018-04-01

    While the three-dimensional Ising model has defied analytic solution, various numerical methods like Monte Carlo, Monte Carlo renormalization group, and series expansion have provided precise information about the phase transition. Using Monte Carlo simulation that employs the Wolff cluster flipping algorithm with both 32-bit and 53-bit random number generators and data analysis with histogram reweighting and quadruple precision arithmetic, we have investigated the critical behavior of the simple cubic Ising Model, with lattice sizes ranging from 163 to 10243. By analyzing data with cross correlations between various thermodynamic quantities obtained from the same data pool, e.g., logarithmic derivatives of magnetization and derivatives of magnetization cumulants, we have obtained the critical inverse temperature Kc=0.221 654 626 (5 ) and the critical exponent of the correlation length ν =0.629 912 (86 ) with precision that exceeds all previous Monte Carlo estimates.

  17. Pattern Recognition for a Flight Dynamics Monte Carlo Simulation

    NASA Technical Reports Server (NTRS)

    Restrepo, Carolina; Hurtado, John E.

    2011-01-01

    The design, analysis, and verification and validation of a spacecraft relies heavily on Monte Carlo simulations. Modern computational techniques are able to generate large amounts of Monte Carlo data but flight dynamics engineers lack the time and resources to analyze it all. The growing amounts of data combined with the diminished available time of engineers motivates the need to automate the analysis process. Pattern recognition algorithms are an innovative way of analyzing flight dynamics data efficiently. They can search large data sets for specific patterns and highlight critical variables so analysts can focus their analysis efforts. This work combines a few tractable pattern recognition algorithms with basic flight dynamics concepts to build a practical analysis tool for Monte Carlo simulations. Current results show that this tool can quickly and automatically identify individual design parameters, and most importantly, specific combinations of parameters that should be avoided in order to prevent specific system failures. The current version uses a kernel density estimation algorithm and a sequential feature selection algorithm combined with a k-nearest neighbor classifier to find and rank important design parameters. This provides an increased level of confidence in the analysis and saves a significant amount of time.

  18. Delayed Slater determinant update algorithms for high efficiency quantum Monte Carlo

    DOE PAGES

    McDaniel, Tyler; D’Azevedo, Ed F.; Li, Ying Wai; ...

    2017-11-07

    Within ab initio Quantum Monte Carlo simulations, the leading numerical cost for large systems is the computation of the values of the Slater determinants in the trial wavefunction. Each Monte Carlo step requires finding the determinant of a dense matrix. This is most commonly iteratively evaluated using a rank-1 Sherman-Morrison updating scheme to avoid repeated explicit calculation of the inverse. The overall computational cost is therefore formally cubic in the number of electrons or matrix size. To improve the numerical efficiency of this procedure, we propose a novel multiple rank delayed update scheme. This strategy enables probability evaluation with applicationmore » of accepted moves to the matrices delayed until after a predetermined number of moves, K. The accepted events are then applied to the matrices en bloc with enhanced arithmetic intensity and computational efficiency via matrix-matrix operations instead of matrix-vector operations. Here this procedure does not change the underlying Monte Carlo sampling or its statistical efficiency. For calculations on large systems and algorithms such as diffusion Monte Carlo where the acceptance ratio is high, order of magnitude improvements in the update time can be obtained on both multi- core CPUs and GPUs.« less

  19. Delayed Slater determinant update algorithms for high efficiency quantum Monte Carlo

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    McDaniel, Tyler; D’Azevedo, Ed F.; Li, Ying Wai

    Within ab initio Quantum Monte Carlo simulations, the leading numerical cost for large systems is the computation of the values of the Slater determinants in the trial wavefunction. Each Monte Carlo step requires finding the determinant of a dense matrix. This is most commonly iteratively evaluated using a rank-1 Sherman-Morrison updating scheme to avoid repeated explicit calculation of the inverse. The overall computational cost is therefore formally cubic in the number of electrons or matrix size. To improve the numerical efficiency of this procedure, we propose a novel multiple rank delayed update scheme. This strategy enables probability evaluation with applicationmore » of accepted moves to the matrices delayed until after a predetermined number of moves, K. The accepted events are then applied to the matrices en bloc with enhanced arithmetic intensity and computational efficiency via matrix-matrix operations instead of matrix-vector operations. Here this procedure does not change the underlying Monte Carlo sampling or its statistical efficiency. For calculations on large systems and algorithms such as diffusion Monte Carlo where the acceptance ratio is high, order of magnitude improvements in the update time can be obtained on both multi- core CPUs and GPUs.« less

  20. Applying Monte Carlo Simulation to Launch Vehicle Design and Requirements Analysis

    NASA Technical Reports Server (NTRS)

    Hanson, J. M.; Beard, B. B.

    2010-01-01

    This Technical Publication (TP) is meant to address a number of topics related to the application of Monte Carlo simulation to launch vehicle design and requirements analysis. Although the focus is on a launch vehicle application, the methods may be applied to other complex systems as well. The TP is organized so that all the important topics are covered in the main text, and detailed derivations are in the appendices. The TP first introduces Monte Carlo simulation and the major topics to be discussed, including discussion of the input distributions for Monte Carlo runs, testing the simulation, how many runs are necessary for verification of requirements, what to do if results are desired for events that happen only rarely, and postprocessing, including analyzing any failed runs, examples of useful output products, and statistical information for generating desired results from the output data. Topics in the appendices include some tables for requirements verification, derivation of the number of runs required and generation of output probabilistic data with consumer risk included, derivation of launch vehicle models to include possible variations of assembled vehicles, minimization of a consumable to achieve a two-dimensional statistical result, recontact probability during staging, ensuring duplicated Monte Carlo random variations, and importance sampling.

  1. Bold Diagrammatic Monte Carlo for Fermionic and Fermionized Systems

    NASA Astrophysics Data System (ADS)

    Svistunov, Boris

    2013-03-01

    In three different fermionic cases--repulsive Hubbard model, resonant fermions, and fermionized spins-1/2 (on triangular lattice)--we observe the phenomenon of sign blessing: Feynman diagrammatic series features finite convergence radius despite factorial growth of the number of diagrams with diagram order. Bold diagrammatic Monte Carlo technique allows us to sample millions of skeleton Feynman diagrams. With the universal fermionization trick we can fermionize essentially any (bosonic, spin, mixed, etc.) lattice system. The combination of fermionization and Bold diagrammatic Monte Carlo yields a universal first-principle approach to strongly correlated lattice systems, provided the sign blessing is a generic fermionic phenomenon. Supported by NSF and DARPA

  2. Understanding quantum tunneling using diffusion Monte Carlo simulations

    NASA Astrophysics Data System (ADS)

    Inack, E. M.; Giudici, G.; Parolini, T.; Santoro, G.; Pilati, S.

    2018-03-01

    In simple ferromagnetic quantum Ising models characterized by an effective double-well energy landscape the characteristic tunneling time of path-integral Monte Carlo (PIMC) simulations has been shown to scale as the incoherent quantum-tunneling time, i.e., as 1 /Δ2 , where Δ is the tunneling gap. Since incoherent quantum tunneling is employed by quantum annealers (QAs) to solve optimization problems, this result suggests that there is no quantum advantage in using QAs with respect to quantum Monte Carlo (QMC) simulations. A counterexample is the recently introduced shamrock model (Andriyash and Amin, arXiv:1703.09277), where topological obstructions cause an exponential slowdown of the PIMC tunneling dynamics with respect to incoherent quantum tunneling, leaving open the possibility for potential quantum speedup, even for stoquastic models. In this work we investigate the tunneling time of projective QMC simulations based on the diffusion Monte Carlo (DMC) algorithm without guiding functions, showing that it scales as 1 /Δ , i.e., even more favorably than the incoherent quantum-tunneling time, both in a simple ferromagnetic system and in the more challenging shamrock model. However, a careful comparison between the DMC ground-state energies and the exact solution available for the transverse-field Ising chain indicates an exponential scaling of the computational cost required to keep a fixed relative error as the system size increases.

  3. Advanced Monte Carlo methods for thermal radiation transport

    NASA Astrophysics Data System (ADS)

    Wollaber, Allan B.

    During the past 35 years, the Implicit Monte Carlo (IMC) method proposed by Fleck and Cummings has been the standard Monte Carlo approach to solving the thermal radiative transfer (TRT) equations. However, the IMC equations are known to have accuracy limitations that can produce unphysical solutions. In this thesis, we explicitly provide the IMC equations with a Monte Carlo interpretation by including particle weight as one of its arguments. We also develop and test a stability theory for the 1-D, gray IMC equations applied to a nonlinear problem. We demonstrate that the worst case occurs for 0-D problems, and we extend the results to a stability algorithm that may be used for general linearizations of the TRT equations. We derive gray, Quasidiffusion equations that may be deterministically solved in conjunction with IMC to obtain an inexpensive, accurate estimate of the temperature at the end of the time step. We then define an average temperature T* to evaluate the temperature-dependent problem data in IMC, and we demonstrate that using T* is more accurate than using the (traditional) beginning-of-time-step temperature. We also propose an accuracy enhancement to the IMC equations: the use of a time-dependent "Fleck factor". This Fleck factor can be considered an automatic tuning of the traditionally defined user parameter alpha, which generally provides more accurate solutions at an increased cost relative to traditional IMC. We also introduce a global weight window that is proportional to the forward scalar intensity calculated by the Quasidiffusion method. This weight window improves the efficiency of the IMC calculation while conserving energy. All of the proposed enhancements are tested in 1-D gray and frequency-dependent problems. These enhancements do not unconditionally eliminate the unphysical behavior that can be seen in the IMC calculations. However, for fixed spatial and temporal grids, they suppress them and clearly work to make the solution more

  4. Geologic evolution of the Akna Montes-Atropos Tessera region, Venus

    NASA Astrophysics Data System (ADS)

    Marinangeli, Lucia; Gilmore, Martha S.

    2000-05-01

    The investigated area comprises an arcuate mountain belt, Akna Montes, in Western Ishtar Terra, associated with an outboard plateau, Atropos Tessera, to the west and a volcanic plateau, Lakshmi Planum, to the east. Eight geologic units have been recognized on the basis of their geomorphic and structural characteristics as they appear on Magellan radar images. Our stratigraphic analysis shows that the geological evolution of the study area can be explained by four main steps: (1) formation of the older substrata of Atropos Tessera and Lakshmi, (2) extensive plains emplacement, (3) an orogenic phase including the formation of Akna Montes, and (4) local emplacement of younger plains. The tectonic evolution shows a deformational sequence characterized by contraction, shear, and topographic relaxation. This sequence is interpreted to be a consequence of the variation of crustal stresses and crustal thickening during orogenic events as observed for terrestrial high plateaus associated with a mountain belt (i.e., Himalaya and Tibet, Andes and Altiplano). In order to estimate the amount of crustal shortening associated with the Akna Montes, we considered two end-members for structural style of the mountain belt: a symmetric fold model and fault-bend fold model. The models are theoretical because terrestrial orogenic belts are often formed by a combination of different compressional structures. However, symmetric and fault-bend faults represent the minimum and maximum crustal shortening, respectively, and thus they do place bounds on the amount of strain recorded by Akna Montes. The first model yields a shortening value less than 1%, whereas a range of 17-34% is derived for the second model. The large difference between these values underscores the importance of fold geometries for estimating strain and to place constraints on geodynamic models for mountain belt formation. On the basis of our study we think that a combination of mantle downwelling and horizontal convergence

  5. Fast GPU-based Monte Carlo simulations for LDR prostate brachytherapy.

    PubMed

    Bonenfant, Éric; Magnoux, Vincent; Hissoiny, Sami; Ozell, Benoît; Beaulieu, Luc; Després, Philippe

    2015-07-07

    The aim of this study was to evaluate the potential of bGPUMCD, a Monte Carlo algorithm executed on Graphics Processing Units (GPUs), for fast dose calculations in permanent prostate implant dosimetry. It also aimed to validate a low dose rate brachytherapy source in terms of TG-43 metrics and to use this source to compute dose distributions for permanent prostate implant in very short times. The physics of bGPUMCD was reviewed and extended to include Rayleigh scattering and fluorescence from photoelectric interactions for all materials involved. The radial and anisotropy functions were obtained for the Nucletron SelectSeed in TG-43 conditions. These functions were compared to those found in the MD Anderson Imaging and Radiation Oncology Core brachytherapy source registry which are considered the TG-43 reference values. After appropriate calibration of the source, permanent prostate implant dose distributions were calculated for four patients and compared to an already validated Geant4 algorithm. The radial function calculated from bGPUMCD showed excellent agreement (differences within 1.3%) with TG-43 accepted values. The anisotropy functions at r = 1 cm and r = 4 cm were within 2% of TG-43 values for angles over 17.5°. For permanent prostate implants, Monte Carlo-based dose distributions with a statistical uncertainty of 1% or less for the target volume were obtained in 30 s or less for 1 × 1 × 1 mm(3) calculation grids. Dosimetric indices were very similar (within 2.7%) to those obtained with a validated, independent Monte Carlo code (Geant4) performing the calculations for the same cases in a much longer time (tens of minutes to more than a hour). bGPUMCD is a promising code that lets envision the use of Monte Carlo techniques in a clinical environment, with sub-minute execution times on a standard workstation. Future work will explore the use of this code with an inverse planning method to provide a complete Monte Carlo-based planning solution.

  6. Fast GPU-based Monte Carlo simulations for LDR prostate brachytherapy

    NASA Astrophysics Data System (ADS)

    Bonenfant, Éric; Magnoux, Vincent; Hissoiny, Sami; Ozell, Benoît; Beaulieu, Luc; Després, Philippe

    2015-07-01

    The aim of this study was to evaluate the potential of bGPUMCD, a Monte Carlo algorithm executed on Graphics Processing Units (GPUs), for fast dose calculations in permanent prostate implant dosimetry. It also aimed to validate a low dose rate brachytherapy source in terms of TG-43 metrics and to use this source to compute dose distributions for permanent prostate implant in very short times. The physics of bGPUMCD was reviewed and extended to include Rayleigh scattering and fluorescence from photoelectric interactions for all materials involved. The radial and anisotropy functions were obtained for the Nucletron SelectSeed in TG-43 conditions. These functions were compared to those found in the MD Anderson Imaging and Radiation Oncology Core brachytherapy source registry which are considered the TG-43 reference values. After appropriate calibration of the source, permanent prostate implant dose distributions were calculated for four patients and compared to an already validated Geant4 algorithm. The radial function calculated from bGPUMCD showed excellent agreement (differences within 1.3%) with TG-43 accepted values. The anisotropy functions at r = 1 cm and r = 4 cm were within 2% of TG-43 values for angles over 17.5°. For permanent prostate implants, Monte Carlo-based dose distributions with a statistical uncertainty of 1% or less for the target volume were obtained in 30 s or less for 1 × 1 × 1 mm3 calculation grids. Dosimetric indices were very similar (within 2.7%) to those obtained with a validated, independent Monte Carlo code (Geant4) performing the calculations for the same cases in a much longer time (tens of minutes to more than a hour). bGPUMCD is a promising code that lets envision the use of Monte Carlo techniques in a clinical environment, with sub-minute execution times on a standard workstation. Future work will explore the use of this code with an inverse planning method to provide a complete Monte Carlo-based planning solution.

  7. Multilevel Sequential Monte Carlo Samplers for Normalizing Constants

    DOE PAGES

    Moral, Pierre Del; Jasra, Ajay; Law, Kody J. H.; ...

    2017-08-24

    This article considers the sequential Monte Carlo (SMC) approximation of ratios of normalizing constants associated to posterior distributions which in principle rely on continuum models. Therefore, the Monte Carlo estimation error and the discrete approximation error must be balanced. A multilevel strategy is utilized to substantially reduce the cost to obtain a given error level in the approximation as compared to standard estimators. Two estimators are considered and relative variance bounds are given. The theoretical results are numerically illustrated for two Bayesian inverse problems arising from elliptic partial differential equations (PDEs). The examples involve the inversion of observations of themore » solution of (i) a 1-dimensional Poisson equation to infer the diffusion coefficient, and (ii) a 2-dimensional Poisson equation to infer the external forcing.« less

  8. Oppenheimer's Box of Chocolates: Remediation of the Manhattan Project Landfill at Los Alamos National Laboratory - 12283

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Allen, Donald L.; Ramsey, Susan S.; Finn, Kevin P.

    2012-07-01

    Material Disposal Area B (MDA B) is the oldest radioactive waste disposal facility at Los Alamos National Laboratory. Operated from 1944-48, MDA B was the disposal facility for the Manhattan Project. Recognized as one of the most challenging environmental remediation projects at Los Alamos, the excavation of MDA B received $110 million from the American Recovery and Reinvestment Act of 2009 to accelerate this complex remediation work. Several factors combined to create significant challenges to remediating the landfill known in the 1940's as the 'contaminated dump'. The secrecy surrounding the Manhattan Project meant that no records were kept of radiologicalmore » materials and chemicals disposed or of the landfill design. An extensive review of historical documents and interviews with early laboratory personnel resulted in a list of hundreds of hazardous chemicals that could have been buried in MDA B. Also, historical reports of MDA B spontaneously combusting on three occasions -with 50-foot flames and pink smoke spewing across the mesa during the last incident in 1948-indicated that hazardous materials were likely present in MDA B. To complicate matters further, though MDA B was located on an isolated mesa in the 1940's, the landfill has since been surrounded by a Los Alamos commercial district. The local newspaper, hardware store and a number of other businesses are located directly across the street from MDA B. This close proximity to the public and the potential for hazardous materials in MDA B necessitated conducting remediation work within protective enclosures. Potential chemical hazards and radiological inventory were better defined using a minimally intrusive sampling method called direct push technology (DPT) prior to excavation. Even with extensive sampling and planning the project team encountered many surprises and challenges during the project. The one area where planning did not fail to meet reality was safety. There were no serious worker

  9. Real-time alpha monitoring of a radioactive liquid waste stream at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Johnson, J.D.; Whitley, C.R.; Rawool-Sullivan, M.

    1995-12-31

    This poster display concerns the development, installation, and testing of a real-time radioactive liquid waste monitor at Los Alamos National Laboratory (LANL). The detector system was designed for the LANL Radioactive Liquid Waste Treatment Facility so that influent to the plant could be monitored in real time. By knowing the activity of the influent, plant operators can better monitor treatment, better segregate waste (potentially), and monitor the regulatory compliance of users of the LANL Radioactive Liquid Waste Collection System. The detector system uses long-range alpha detection technology, which is a nonintrusive method of characterization that determines alpha activity on themore » liquid surface by measuring the ionization of ambient air. Extensive testing has been performed to ensure long-term use with a minimal amount of maintenance. The final design was a simple cost-effective alpha monitor that could be modified for monitoring influent waste streams at various points in the LANL Radioactive Liquid Waste Collection System.« less

  10. Low-level radioactive waste management: transitioning to off-site disposal at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Dorries, Alison M

    2010-11-09

    Facing the closure of nearly all on-site management and disposal capability for low-level radioactive waste (LLW), Los Alamos National Laboratory (LANL) is making ready to ship the majority of LLW off-site. In order to ship off-site, waste must meet the Treatment, Storage, and Disposal Facility's (TSDF) Waste Acceptance Criteria (WAC). In preparation, LANL's waste management organization must ensure LANL waste generators characterize and package waste compliantly and waste characterization documentation is complete and accurate. Key challenges that must be addressed to successfully make the shift to off-site disposal of LLW include improving the detail, accuracy, and quality of process knowledgemore » (PK) and acceptable knowledge (AK) documentation, training waste generators and waste management staff on the higher standard of data quality and expectations, improved WAC compliance for off-site facilities, and enhanced quality assurance throughout the process. Certification of LANL generators will allow direct off-site shipping of LLW from their facilities.« less

  11. Continuous-time quantum Monte Carlo impurity solvers

    NASA Astrophysics Data System (ADS)

    Gull, Emanuel; Werner, Philipp; Fuchs, Sebastian; Surer, Brigitte; Pruschke, Thomas; Troyer, Matthias

    2011-04-01

    Continuous-time quantum Monte Carlo impurity solvers are algorithms that sample the partition function of an impurity model using diagrammatic Monte Carlo techniques. The present paper describes codes that implement the interaction expansion algorithm originally developed by Rubtsov, Savkin, and Lichtenstein, as well as the hybridization expansion method developed by Werner, Millis, Troyer, et al. These impurity solvers are part of the ALPS-DMFT application package and are accompanied by an implementation of dynamical mean-field self-consistency equations for (single orbital single site) dynamical mean-field problems with arbitrary densities of states. Program summaryProgram title: dmft Catalogue identifier: AEIL_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEIL_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: ALPS LIBRARY LICENSE version 1.1 No. of lines in distributed program, including test data, etc.: 899 806 No. of bytes in distributed program, including test data, etc.: 32 153 916 Distribution format: tar.gz Programming language: C++ Operating system: The ALPS libraries have been tested on the following platforms and compilers: Linux with GNU Compiler Collection (g++ version 3.1 and higher), and Intel C++ Compiler (icc version 7.0 and higher) MacOS X with GNU Compiler (g++ Apple-version 3.1, 3.3 and 4.0) IBM AIX with Visual Age C++ (xlC version 6.0) and GNU (g++ version 3.1 and higher) compilers Compaq Tru64 UNIX with Compq C++ Compiler (cxx) SGI IRIX with MIPSpro C++ Compiler (CC) HP-UX with HP C++ Compiler (aCC) Windows with Cygwin or coLinux platforms and GNU Compiler Collection (g++ version 3.1 and higher) RAM: 10 MB-1 GB Classification: 7.3 External routines: ALPS [1], BLAS/LAPACK, HDF5 Nature of problem: (See [2].) Quantum impurity models describe an atom or molecule embedded in a host material with which it can exchange electrons. They are basic to nanoscience as

  12. Event-chain Monte Carlo algorithms for three- and many-particle interactions

    NASA Astrophysics Data System (ADS)

    Harland, J.; Michel, M.; Kampmann, T. A.; Kierfeld, J.

    2017-02-01

    We generalize the rejection-free event-chain Monte Carlo algorithm from many-particle systems with pairwise interactions to systems with arbitrary three- or many-particle interactions. We introduce generalized lifting probabilities between particles and obtain a general set of equations for lifting probabilities, the solution of which guarantees maximal global balance. We validate the resulting three-particle event-chain Monte Carlo algorithms on three different systems by comparison with conventional local Monte Carlo simulations: i) a test system of three particles with a three-particle interaction that depends on the enclosed triangle area; ii) a hard-needle system in two dimensions, where needle interactions constitute three-particle interactions of the needle end points; iii) a semiflexible polymer chain with a bending energy, which constitutes a three-particle interaction of neighboring chain beads. The examples demonstrate that the generalization to many-particle interactions broadens the applicability of event-chain algorithms considerably.

  13. A Lattice Kinetic Monte Carlo Solver for First-Principles Microkinetic Trend Studies

    DOE PAGES

    Hoffmann, Max J.; Bligaard, Thomas

    2018-01-22

    Here, mean-field microkinetic models in combination with Brønsted–Evans–Polanyi like scaling relations have proven highly successful in identifying catalyst materials with good or promising reactivity and selectivity. Analysis of the microkinetic model by means of lattice kinetic Monte Carlo promises a faithful description of a range of atomistic features involving short-range ordering of species in the vicinity of an active site. In this paper, we use the “fruit fly” example reaction of CO oxidation on fcc(111) transition and coinage metals to motivate and develop a lattice kinetic Monte Carlo solver suitable for the numerically challenging case of vastly disparate rate constants.more » As a result, we show that for the case of infinitely fast diffusion and absence of adsorbate-adsorbate interaction it is, in fact, possible to match the prediction of the mean-field-theory method and the lattice kinetic Monte Carlo method. As a corollary, we conclude that lattice kinetic Monte Carlo simulations of surface chemical reactions are most likely to provide additional insight over mean-field simulations if diffusion limitations or adsorbate–adsorbate interactions have a significant influence on the mixing of the adsorbates.« less

  14. A Lattice Kinetic Monte Carlo Solver for First-Principles Microkinetic Trend Studies

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hoffmann, Max J.; Bligaard, Thomas

    Here, mean-field microkinetic models in combination with Brønsted–Evans–Polanyi like scaling relations have proven highly successful in identifying catalyst materials with good or promising reactivity and selectivity. Analysis of the microkinetic model by means of lattice kinetic Monte Carlo promises a faithful description of a range of atomistic features involving short-range ordering of species in the vicinity of an active site. In this paper, we use the “fruit fly” example reaction of CO oxidation on fcc(111) transition and coinage metals to motivate and develop a lattice kinetic Monte Carlo solver suitable for the numerically challenging case of vastly disparate rate constants.more » As a result, we show that for the case of infinitely fast diffusion and absence of adsorbate-adsorbate interaction it is, in fact, possible to match the prediction of the mean-field-theory method and the lattice kinetic Monte Carlo method. As a corollary, we conclude that lattice kinetic Monte Carlo simulations of surface chemical reactions are most likely to provide additional insight over mean-field simulations if diffusion limitations or adsorbate–adsorbate interactions have a significant influence on the mixing of the adsorbates.« less

  15. Mosaicing of airborne LiDAR bathymetry strips based on Monte Carlo matching

    NASA Astrophysics Data System (ADS)

    Yang, Fanlin; Su, Dianpeng; Zhang, Kai; Ma, Yue; Wang, Mingwei; Yang, Anxiu

    2017-09-01

    This study proposes a new methodology for mosaicing airborne light detection and ranging (LiDAR) bathymetry (ALB) data based on Monte Carlo matching. Various errors occur in ALB data due to imperfect system integration and other interference factors. To account for these errors, a Monte Carlo matching algorithm based on a nonlinear least-squares adjustment model is proposed. First, the raw data of strip overlap areas were filtered according to their relative drift of depths. Second, a Monte Carlo model and nonlinear least-squares adjustment model were combined to obtain seven transformation parameters. Then, the multibeam bathymetric data were used to correct the initial strip during strip mosaicing. Finally, to evaluate the proposed method, the experimental results were compared with the results of the Iterative Closest Points (ICP) and three-dimensional Normal Distributions Transform (3D-NDT) algorithms. The results demonstrate that the algorithm proposed in this study is more robust and effective. When the quality of the raw data is poor, the Monte Carlo matching algorithm can still achieve centimeter-level accuracy for overlapping areas, which meets the accuracy of bathymetry required by IHO Standards for Hydrographic Surveys Special Publication No.44.

  16. Finite element model updating using the shadow hybrid Monte Carlo technique

    NASA Astrophysics Data System (ADS)

    Boulkaibet, I.; Mthembu, L.; Marwala, T.; Friswell, M. I.; Adhikari, S.

    2015-02-01

    Recent research in the field of finite element model updating (FEM) advocates the adoption of Bayesian analysis techniques to dealing with the uncertainties associated with these models. However, Bayesian formulations require the evaluation of the Posterior Distribution Function which may not be available in analytical form. This is the case in FEM updating. In such cases sampling methods can provide good approximations of the Posterior distribution when implemented in the Bayesian context. Markov Chain Monte Carlo (MCMC) algorithms are the most popular sampling tools used to sample probability distributions. However, the efficiency of these algorithms is affected by the complexity of the systems (the size of the parameter space). The Hybrid Monte Carlo (HMC) offers a very important MCMC approach to dealing with higher-dimensional complex problems. The HMC uses the molecular dynamics (MD) steps as the global Monte Carlo (MC) moves to reach areas of high probability where the gradient of the log-density of the Posterior acts as a guide during the search process. However, the acceptance rate of HMC is sensitive to the system size as well as the time step used to evaluate the MD trajectory. To overcome this limitation we propose the use of the Shadow Hybrid Monte Carlo (SHMC) algorithm. The SHMC algorithm is a modified version of the Hybrid Monte Carlo (HMC) and designed to improve sampling for large-system sizes and time steps. This is done by sampling from a modified Hamiltonian function instead of the normal Hamiltonian function. In this paper, the efficiency and accuracy of the SHMC method is tested on the updating of two real structures; an unsymmetrical H-shaped beam structure and a GARTEUR SM-AG19 structure and is compared to the application of the HMC algorithm on the same structures.

  17. A novel Kinetic Monte Carlo algorithm for Non-Equilibrium Simulations

    NASA Astrophysics Data System (ADS)

    Jha, Prateek; Kuzovkov, Vladimir; Grzybowski, Bartosz; Olvera de La Cruz, Monica

    2012-02-01

    We have developed an off-lattice kinetic Monte Carlo simulation scheme for reaction-diffusion problems in soft matter systems. The definition of transition probabilities in the Monte Carlo scheme are taken identical to the transition rates in a renormalized master equation of the diffusion process and match that of the Glauber dynamics of Ising model. Our scheme provides several advantages over the Brownian dynamics technique for non-equilibrium simulations. Since particle displacements are accepted/rejected in a Monte Carlo fashion as opposed to moving particles following a stochastic equation of motion, nonphysical movements (e.g., violation of a hard core assumption) are not possible (these moves have zero acceptance). Further, the absence of a stochastic ``noise'' term resolves the computational difficulties associated with generating statistically independent trajectories with definitive mean properties. Finally, since the timestep is independent of the magnitude of the interaction forces, much longer time-steps can be employed than Brownian dynamics. We discuss the applications of this scheme for dynamic self-assembly of photo-switchable nanoparticles and dynamical problems in polymeric systems.

  18. ME(SSY)**2: Monte Carlo Code for Star Cluster Simulations

    NASA Astrophysics Data System (ADS)

    Freitag, Marc Dewi

    2013-02-01

    ME(SSY)**2 stands for “Monte-carlo Experiments with Spherically SYmmetric Stellar SYstems." This code simulates the long term evolution of spherical clusters of stars; it was devised specifically to treat dense galactic nuclei. It is based on the pioneering Monte Carlo scheme proposed by Hénon in the 70's and includes all relevant physical ingredients (2-body relaxation, stellar mass spectrum, collisions, tidal disruption, ldots). It is basically a Monte Carlo resolution of the Fokker-Planck equation. It can cope with any stellar mass spectrum or velocity distribution. Being a particle-based method, it also allows one to take stellar collisions into account in a very realistic way. This unique code, featuring most important physical processes, allows million particle simulations, spanning a Hubble time, in a few CPU days on standard personal computers and provides a wealth of data only rivalized by N-body simulations. The current version of the software requires the use of routines from the "Numerical Recipes in Fortran 77" (http://www.nrbook.com/a/bookfpdf.php).

  19. Chemical accuracy from quantum Monte Carlo for the benzene dimer.

    PubMed

    Azadi, Sam; Cohen, R E

    2015-09-14

    We report an accurate study of interactions between benzene molecules using variational quantum Monte Carlo (VMC) and diffusion quantum Monte Carlo (DMC) methods. We compare these results with density functional theory using different van der Waals functionals. In our quantum Monte Carlo (QMC) calculations, we use accurate correlated trial wave functions including three-body Jastrow factors and backflow transformations. We consider two benzene molecules in the parallel displaced geometry, and find that by highly optimizing the wave function and introducing more dynamical correlation into the wave function, we compute the weak chemical binding energy between aromatic rings accurately. We find optimal VMC and DMC binding energies of -2.3(4) and -2.7(3) kcal/mol, respectively. The best estimate of the coupled-cluster theory through perturbative triplets/complete basis set limit is -2.65(2) kcal/mol [Miliordos et al., J. Phys. Chem. A 118, 7568 (2014)]. Our results indicate that QMC methods give chemical accuracy for weakly bound van der Waals molecular interactions, comparable to results from the best quantum chemistry methods.

  20. Performance of quantum Monte Carlo for calculating molecular bond lengths

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cleland, Deidre M., E-mail: deidre.cleland@csiro.au; Per, Manolo C., E-mail: manolo.per@csiro.au

    2016-03-28

    This work investigates the accuracy of real-space quantum Monte Carlo (QMC) methods for calculating molecular geometries. We present the equilibrium bond lengths of a test set of 30 diatomic molecules calculated using variational Monte Carlo (VMC) and diffusion Monte Carlo (DMC) methods. The effect of different trial wavefunctions is investigated using single determinants constructed from Hartree-Fock (HF) and Density Functional Theory (DFT) orbitals with LDA, PBE, and B3LYP functionals, as well as small multi-configurational self-consistent field (MCSCF) multi-determinant expansions. When compared to experimental geometries, all DMC methods exhibit smaller mean-absolute deviations (MADs) than those given by HF, DFT, and MCSCF.more » The most accurate MAD of 3 ± 2 × 10{sup −3} Å is achieved using DMC with a small multi-determinant expansion. However, the more computationally efficient multi-determinant VMC method has a similar MAD of only 4.0 ± 0.9 × 10{sup −3} Å, suggesting that QMC forces calculated from the relatively simple VMC algorithm may often be sufficient for accurate molecular geometries.« less

  1. Studies of Annual and Seasonal Variations in Four Species of Reptiles and Amphibians at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Keller, D.C.; Nelson, E.I.; Mullen, M.A.

    1998-07-01

    Baseline studies of reptiles and amphibians of the Pajarito wetlands at Los Alamos National Laboratory have been conducted by the Ecology group since 1990. With the data gathered from 1990-1997 (excluding 1992), we examined the annual and seasonal population changes of four species of reptiles and amphibians over the past seven years. The four species studied are the Woodhouse toad (Bufo woodhousii), the western chorus frog (Pseudacris triseriata), the many-lined skink (Eunzeces nudtivirgatus), and the plateau striped whiptail lizard (Cnemidophorus velox). Statistical analyses indicate a significant change on a seasonal basis for the western chorus frog and the many-lined skink.more » Results indicate a significant difference in the annual population of the Woodhouse toad.« less

  2. Studies of annual and seasonal variations in four species of reptiles and amphibians at Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Nelson, E.I.; Haarmann, T.; Keller, D.C.

    1998-11-01

    Baseline studies of reptiles and amphibians of the Pajarito wetlands at Los Alamos National Laboratory have been conducted by the Ecology group since 1990. With the gathered data from 1990--1997 (excluding 1992), they plan to determine if patterns can be found in the annual and seasonal population changes of four species of reptiles and amphibians over the past seven years. The four species studied are the Woodhouse toad, the western chorus frog, the many-linked skink, and the plateau striped whiptail lizard. Statistical analysis results show that significant changes occurred on a seasonal basis for the western chorus frog and themore » many-lined skink. Results indicate a significant difference in the annual population of the Woodhouse toad.« less

  3. An In Situ Radiological Survey of Three Canyons at the Los Alamos National Laboratory

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    R.J. Maurer

    1999-06-01

    An in situ radiological survey of Mortandad, Ten Site, and DP Canyons at the Los Alamos National Laboratory was conducted during August 19-30, 1996. The purpose of this survey was to measure the quantities of radionuclides that remain in the canyons from past laboratory operations. A total of 65 in situ measurements were conducted using high-resolution gamma radiation detectors at 1 meter above the ground. The measurements were obtained in the streambeds of the canyons beginning near the water-release points at the laboratories and extending to the ends of the canyons. Three man-made gamma-emitting radionuclides were detected in the canyons:more » americium-241 ({sup 241}Am), cesium-137 ({sup 137}Cs), and cobalt-60 ({sup 60}Co). Estimated contamination levels ranged from 13.3-290.4 picocuries per gram (pCi/g)for {sup 241}Am, 4.4-327.8 pCi/g for {sup 137}Cs, and 0.4-2.6 pCi/g for {sup 60}Co.« less

  4. Monte Carlo simulations within avalanche rescue

    NASA Astrophysics Data System (ADS)

    Reiweger, Ingrid; Genswein, Manuel; Schweizer, Jürg

    2016-04-01

    Refining concepts for avalanche rescue involves calculating suitable settings for rescue strategies such as an adequate probing depth for probe line searches or an optimal time for performing resuscitation for a recovered avalanche victim in case of additional burials. In the latter case, treatment decisions have to be made in the context of triage. However, given the low number of incidents it is rarely possible to derive quantitative criteria based on historical statistics in the context of evidence-based medicine. For these rare, but complex rescue scenarios, most of the associated concepts, theories, and processes involve a number of unknown "random" parameters which have to be estimated in order to calculate anything quantitatively. An obvious approach for incorporating a number of random variables and their distributions into a calculation is to perform a Monte Carlo (MC) simulation. We here present Monte Carlo simulations for calculating the most suitable probing depth for probe line searches depending on search area and an optimal resuscitation time in case of multiple avalanche burials. The MC approach reveals, e.g., new optimized values for the duration of resuscitation that differ from previous, mainly case-based assumptions.

  5. Validation of the Monte Carlo simulator GATE for indium-111 imaging.

    PubMed

    Assié, K; Gardin, I; Véra, P; Buvat, I

    2005-07-07

    Monte Carlo simulations are useful for optimizing and assessing single photon emission computed tomography (SPECT) protocols, especially when aiming at measuring quantitative parameters from SPECT images. Before Monte Carlo simulated data can be trusted, the simulation model must be validated. The purpose of this work was to validate the use of GATE, a new Monte Carlo simulation platform based on GEANT4, for modelling indium-111 SPECT data, the quantification of which is of foremost importance for dosimetric studies. To that end, acquisitions of (111)In line sources in air and in water and of a cylindrical phantom were performed, together with the corresponding simulations. The simulation model included Monte Carlo modelling of the camera collimator and of a back-compartment accounting for photomultiplier tubes and associated electronics. Energy spectra, spatial resolution, sensitivity values, images and count profiles obtained for experimental and simulated data were compared. An excellent agreement was found between experimental and simulated energy spectra. For source-to-collimator distances varying from 0 to 20 cm, simulated and experimental spatial resolution differed by less than 2% in air, while the simulated sensitivity values were within 4% of the experimental values. The simulation of the cylindrical phantom closely reproduced the experimental data. These results suggest that GATE enables accurate simulation of (111)In SPECT acquisitions.

  6. Self-evolving atomistic kinetic Monte Carlo simulations of defects in materials

    DOE PAGES

    Xu, Haixuan; Beland, Laurent K.; Stoller, Roger E.; ...

    2015-01-29

    The recent development of on-the-fly atomistic kinetic Monte Carlo methods has led to an increased amount attention on the methods and their corresponding capabilities and applications. In this review, the framework and current status of Self-Evolving Atomistic Kinetic Monte Carlo (SEAKMC) are discussed. SEAKMC particularly focuses on defect interaction and evolution with atomistic details without assuming potential defect migration/interaction mechanisms and energies. The strength and limitation of using an active volume, the key concept introduced in SEAKMC, are discussed. Potential criteria for characterizing an active volume are discussed and the influence of active volume size on saddle point energies ismore » illustrated. A procedure starting with a small active volume followed by larger active volumes was found to possess higher efficiency. Applications of SEAKMC, ranging from point defect diffusion, to complex interstitial cluster evolution, to helium interaction with tungsten surfaces, are summarized. A comparison of SEAKMC with molecular dynamics and conventional object kinetic Monte Carlo is demonstrated. Overall, SEAKMC is found to be complimentary to conventional molecular dynamics, especially when the harmonic approximation of transition state theory is accurate. However it is capable of reaching longer time scales than molecular dynamics and it can be used to systematically increase the accuracy of other methods such as object kinetic Monte Carlo. Furthermore, the challenges and potential development directions are also outlined.« less

  7. Present Status and Extensions of the Monte Carlo Performance Benchmark

    NASA Astrophysics Data System (ADS)

    Hoogenboom, J. Eduard; Petrovic, Bojan; Martin, William R.

    2014-06-01

    The NEA Monte Carlo Performance benchmark started in 2011 aiming to monitor over the years the abilities to perform a full-size Monte Carlo reactor core calculation with a detailed power production for each fuel pin with axial distribution. This paper gives an overview of the contributed results thus far. It shows that reaching a statistical accuracy of 1 % for most of the small fuel zones requires about 100 billion neutron histories. The efficiency of parallel execution of Monte Carlo codes on a large number of processor cores shows clear limitations for computer clusters with common type computer nodes. However, using true supercomputers the speedup of parallel calculations is increasing up to large numbers of processor cores. More experience is needed from calculations on true supercomputers using large numbers of processors in order to predict if the requested calculations can be done in a short time. As the specifications of the reactor geometry for this benchmark test are well suited for further investigations of full-core Monte Carlo calculations and a need is felt for testing other issues than its computational performance, proposals are presented for extending the benchmark to a suite of benchmark problems for evaluating fission source convergence for a system with a high dominance ratio, for coupling with thermal-hydraulics calculations to evaluate the use of different temperatures and coolant densities and to study the correctness and effectiveness of burnup calculations. Moreover, other contemporary proposals for a full-core calculation with realistic geometry and material composition will be discussed.

  8. Monte Carlo calculations of k{sub Q}, the beam quality conversion factor

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Muir, B. R.; Rogers, D. W. O.

    2010-11-15

    Purpose: To use EGSnrc Monte Carlo simulations to directly calculate beam quality conversion factors, k{sub Q}, for 32 cylindrical ionization chambers over a range of beam qualities and to quantify the effect of systematic uncertainties on Monte Carlo calculations of k{sub Q}. These factors are required to use the TG-51 or TRS-398 clinical dosimetry protocols for calibrating external radiotherapy beams. Methods: Ionization chambers are modeled either from blueprints or manufacturers' user's manuals. The dose-to-air in the chamber is calculated using the EGSnrc user-code egs{sub c}hamber using 11 different tabulated clinical photon spectra for the incident beams. The dose to amore » small volume of water is also calculated in the absence of the chamber at the midpoint of the chamber on its central axis. Using a simple equation, k{sub Q} is calculated from these quantities under the assumption that W/e is constant with energy and compared to TG-51 protocol and measured values. Results: Polynomial fits to the Monte Carlo calculated k{sub Q} factors as a function of beam quality expressed as %dd(10){sub x} and TPR{sub 10}{sup 20} are given for each ionization chamber. Differences are explained between Monte Carlo calculated values and values from the TG-51 protocol or calculated using the computer program used for TG-51 calculations. Systematic uncertainties in calculated k{sub Q} values are analyzed and amount to a maximum of one standard deviation uncertainty of 0.99% if one assumes that photon cross-section uncertainties are uncorrelated and 0.63% if they are assumed correlated. The largest components of the uncertainty are the constancy of W/e and the uncertainty in the cross-section for photons in water. Conclusions: It is now possible to calculate k{sub Q} directly using Monte Carlo simulations. Monte Carlo calculations for most ionization chambers give results which are comparable to TG-51 values. Discrepancies can be explained using individual Monte Carlo

  9. Development and Implementation of Photonuclear Cross-Section Data for Mutually Coupled Neutron-Photon Transport Calculations in the Monte Carlo N-Particle (MCNP) Radiation Transport Code

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    White, Morgan C.

    2000-07-01

    The fundamental motivation for the research presented in this dissertation was the need to development a more accurate prediction method for characterization of mixed radiation fields around medical electron accelerators (MEAs). Specifically, a model is developed for simulation of neutron and other particle production from photonuclear reactions and incorporated in the Monte Carlo N-Particle (MCNP) radiation transport code. This extension of the capability within the MCNP code provides for the more accurate assessment of the mixed radiation fields. The Nuclear Theory and Applications group of the Los Alamos National Laboratory has recently provided first-of-a-kind evaluated photonuclear data for a selectmore » group of isotopes. These data provide the reaction probabilities as functions of incident photon energy with angular and energy distribution information for all reaction products. The availability of these data is the cornerstone of the new methodology for state-of-the-art mutually coupled photon-neutron transport simulations. The dissertation includes details of the model development and implementation necessary to use the new photonuclear data within MCNP simulations. A new data format has been developed to include tabular photonuclear data. Data are processed from the Evaluated Nuclear Data Format (ENDF) to the new class ''u'' A Compact ENDF (ACE) format using a standalone processing code. MCNP modifications have been completed to enable Monte Carlo sampling of photonuclear reactions. Note that both neutron and gamma production are included in the present model. The new capability has been subjected to extensive verification and validation (V&V) testing. Verification testing has established the expected basic functionality. Two validation projects were undertaken. First, comparisons were made to benchmark data from literature. These calculations demonstrate the accuracy of the new data and transport routines to better than 25 percent. Second, the ability

  10. Los Alamos National Laboratory Science Education Program. Annual progress report, October 1, 1995--September 30, 1996

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gill, D.H.

    1997-01-01

    The National Teacher Enhancement program (NTEP) is a three-year, multi-laboratory effort funded by the National Science Foundation and the Department of Energy to improve elementary school science programs. The Los Alamos National Laboratory targets teachers in northern New Mexico. FY96, the third year of the program, involved 11 teams of elementary school teachers (grades 4-6) in a three-week summer session, four two-day workshops during the school year and an on-going planning and implementation process. The teams included twenty-one teachers from 11 schools. Participants earned a possible six semester hours of graduate credit for the summer institute and two hours formore » the academic year workshops from the University of New Mexico. The Laboratory expertise in the earth and environmental science provided the tie between the Laboratory initiatives and program content, and allowed for the design of real world problems.« less

  11. Modifying the Monte Carlo Quiz to Increase Student Motivation, Participation, and Content Retention

    ERIC Educational Resources Information Center

    Simonson, Shawn R.

    2017-01-01

    Fernald developed the Monte Carlo Quiz format to enhance retention, encourage students to prepare for class, read with intention, and organize information in psychology classes. This author modified the Monte Carlo Quiz, combined it with the Minute Paper, and applied it to various courses. Students write quiz questions as part of the Minute Paper…

  12. Data decomposition of Monte Carlo particle transport simulations via tally servers

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Romano, Paul K.; Siegel, Andrew R.; Forget, Benoit

    An algorithm for decomposing large tally data in Monte Carlo particle transport simulations is developed, analyzed, and implemented in a continuous-energy Monte Carlo code, OpenMC. The algorithm is based on a non-overlapping decomposition of compute nodes into tracking processors and tally servers. The former are used to simulate the movement of particles through the domain while the latter continuously receive and update tally data. A performance model for this approach is developed, suggesting that, for a range of parameters relevant to LWR analysis, the tally server algorithm should perform with minimal overhead on contemporary supercomputers. An implementation of the algorithmmore » in OpenMC is then tested on the Intrepid and Titan supercomputers, supporting the key predictions of the model over a wide range of parameters. We thus conclude that the tally server algorithm is a successful approach to circumventing classical on-node memory constraints en route to unprecedentedly detailed Monte Carlo reactor simulations.« less

  13. CONTINUOUS-ENERGY MONTE CARLO METHODS FOR CALCULATING GENERALIZED RESPONSE SENSITIVITIES USING TSUNAMI-3D

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Perfetti, Christopher M; Rearden, Bradley T

    2014-01-01

    This work introduces a new approach for calculating sensitivity coefficients for generalized neutronic responses to nuclear data uncertainties using continuous-energy Monte Carlo methods. The approach presented in this paper, known as the GEAR-MC method, allows for the calculation of generalized sensitivity coefficients for multiple responses in a single Monte Carlo calculation with no nuclear data perturbations or knowledge of nuclear covariance data. The theory behind the GEAR-MC method is presented here, and proof of principle is demonstrated by using the GEAR-MC method to calculate sensitivity coefficients for responses in several 3D, continuous-energy Monte Carlo applications.

  14. Monte Carlo simulation of energy-dispersive x-ray fluorescence and applications

    NASA Astrophysics Data System (ADS)

    Li, Fusheng

    Four key components with regards to Monte Carlo Library Least Squares (MCLLS) have been developed by the author. These include: a comprehensive and accurate Monte Carlo simulation code - CEARXRF5 with Differential Operators (DO) and coincidence sampling, Detector Response Function (DRF), an integrated Monte Carlo - Library Least-Squares (MCLLS) Graphical User Interface (GUI) visualization System (MCLLSPro) and a new reproducible and flexible benchmark experiment setup. All these developments or upgrades enable the MCLLS approach to be a useful and powerful tool for a tremendous variety of elemental analysis applications. CEARXRF, a comprehensive and accurate Monte Carlo code for simulating the total and individual library spectral responses of all elements, has been recently upgraded to version 5 by the author. The new version has several key improvements: input file format fully compatible with MCNP5, a new efficient general geometry tracking code, versatile source definitions, various variance reduction techniques (e.g. weight window mesh and splitting, stratifying sampling, etc.), a new cross section data storage and accessing method which improves the simulation speed by a factor of four and new cross section data, upgraded differential operators (DO) calculation capability, and also an updated coincidence sampling scheme which including K-L and L-L coincidence X-Rays, while keeping all the capabilities of the previous version. The new Differential Operators method is powerful for measurement sensitivity study and system optimization. For our Monte Carlo EDXRF elemental analysis system, it becomes an important technique for quantifying the matrix effect in near real time when combined with the MCLLS approach. An integrated visualization GUI system has been developed by the author to perform elemental analysis using iterated Library Least-Squares method for various samples when an initial guess is provided. This software was built on the Borland C++ Builder

  15. Frequency-resolved Monte Carlo.

    PubMed

    López Carreño, Juan Camilo; Del Valle, Elena; Laussy, Fabrice P

    2018-05-03

    We adapt the Quantum Monte Carlo method to the cascaded formalism of quantum optics, allowing us to simulate the emission of photons of known energy. Statistical processing of the photon clicks thus collected agrees with the theory of frequency-resolved photon correlations, extending the range of applications based on correlations of photons of prescribed energy, in particular those of a photon-counting character. We apply the technique to autocorrelations of photon streams from a two-level system under coherent and incoherent pumping, including the Mollow triplet regime where we demonstrate the direct manifestation of leapfrog processes in producing an increased rate of two-photon emission events.

  16. APPLICATION OF BAYESIAN MONTE CARLO ANALYSIS TO A LAGRANGIAN PHOTOCHEMICAL AIR QUALITY MODEL. (R824792)

    EPA Science Inventory

    Uncertainties in ozone concentrations predicted with a Lagrangian photochemical air quality model have been estimated using Bayesian Monte Carlo (BMC) analysis. Bayesian Monte Carlo analysis provides a means of combining subjective "prior" uncertainty estimates developed ...

  17. Optimization of the Monte Carlo code for modeling of photon migration in tissue.

    PubMed

    Zołek, Norbert S; Liebert, Adam; Maniewski, Roman

    2006-10-01

    The Monte Carlo method is frequently used to simulate light transport in turbid media because of its simplicity and flexibility, allowing to analyze complicated geometrical structures. Monte Carlo simulations are, however, time consuming because of the necessity to track the paths of individual photons. The time consuming computation is mainly associated with the calculation of the logarithmic and trigonometric functions as well as the generation of pseudo-random numbers. In this paper, the Monte Carlo algorithm was developed and optimized, by approximation of the logarithmic and trigonometric functions. The approximations were based on polynomial and rational functions, and the errors of these approximations are less than 1% of the values of the original functions. The proposed algorithm was verified by simulations of the time-resolved reflectance at several source-detector separations. The results of the calculation using the approximated algorithm were compared with those of the Monte Carlo simulations obtained with an exact computation of the logarithm and trigonometric functions as well as with the solution of the diffusion equation. The errors of the moments of the simulated distributions of times of flight of photons (total number of photons, mean time of flight and variance) are less than 2% for a range of optical properties, typical of living tissues. The proposed approximated algorithm allows to speed up the Monte Carlo simulations by a factor of 4. The developed code can be used on parallel machines, allowing for further acceleration.

  18. Monte Carlo Modeling of the Initial Radiation Emitted by a Nuclear Device in the National Capital Region

    DTIC Science & Technology

    2013-07-01

    also simulated in the models. Data was derived from calculations using the three-dimensional Monte Carlo radiation transport code MCNP (Monte Carlo N...32  B.  MCNP PHYSICS OPTIONS ......................................................................................... 33  C.  HAZUS...input deck’) for the MCNP , Monte Carlo N-Particle, radiation transport code. MCNP is a general-purpose code designed to simulate neutron, photon

  19. Quantum Monte Carlo Endstation for Petascale Computing

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lubos Mitas

    2011-01-26

    NCSU research group has been focused on accomplising the key goals of this initiative: establishing new generation of quantum Monte Carlo (QMC) computational tools as a part of Endstation petaflop initiative for use at the DOE ORNL computational facilities and for use by computational electronic structure community at large; carrying out high accuracy quantum Monte Carlo demonstration projects in application of these tools to the forefront electronic structure problems in molecular and solid systems; expanding the impact of QMC methods and approaches; explaining and enhancing the impact of these advanced computational approaches. In particular, we have developed quantum Monte Carlomore » code (QWalk, www.qwalk.org) which was significantly expanded and optimized using funds from this support and at present became an actively used tool in the petascale regime by ORNL researchers and beyond. These developments have been built upon efforts undertaken by the PI's group and collaborators over the period of the last decade. The code was optimized and tested extensively on a number of parallel architectures including petaflop ORNL Jaguar machine. We have developed and redesigned a number of code modules such as evaluation of wave functions and orbitals, calculations of pfaffians and introduction of backflow coordinates together with overall organization of the code and random walker distribution over multicore architectures. We have addressed several bottlenecks such as load balancing and verified efficiency and accuracy of the calculations with the other groups of the Endstation team. The QWalk package contains about 50,000 lines of high quality object-oriented C++ and includes also interfaces to data files from other conventional electronic structure codes such as Gamess, Gaussian, Crystal and others. This grant supported PI for one month during summers, a full-time postdoc and partially three graduate students over the period of the grant duration, it has resulted in

  20. Monte Carlo Methods in Materials Science Based on FLUKA and ROOT

    NASA Technical Reports Server (NTRS)

    Pinsky, Lawrence; Wilson, Thomas; Empl, Anton; Andersen, Victor

    2003-01-01

    A comprehensive understanding of mitigation measures for space radiation protection necessarily involves the relevant fields of nuclear physics and particle transport modeling. One method of modeling the interaction of radiation traversing matter is Monte Carlo analysis, a subject that has been evolving since the very advent of nuclear reactors and particle accelerators in experimental physics. Countermeasures for radiation protection from neutrons near nuclear reactors, for example, were an early application and Monte Carlo methods were quickly adapted to this general field of investigation. The project discussed here is concerned with taking the latest tools and technology in Monte Carlo analysis and adapting them to space applications such as radiation shielding design for spacecraft, as well as investigating how next-generation Monte Carlos can complement the existing analytical methods currently used by NASA. We have chosen to employ the Monte Carlo program known as FLUKA (A legacy acronym based on the German for FLUctuating KAscade) used to simulate all of the particle transport, and the CERN developed graphical-interface object-oriented analysis software called ROOT. One aspect of space radiation analysis for which the Monte Carlo s are particularly suited is the study of secondary radiation produced as albedoes in the vicinity of the structural geometry involved. This broad goal of simulating space radiation transport through the relevant materials employing the FLUKA code necessarily requires the addition of the capability to simulate all heavy-ion interactions from 10 MeV/A up to the highest conceivable energies. For all energies above 3 GeV/A the Dual Parton Model (DPM) is currently used, although the possible improvement of the DPMJET event generator for energies 3-30 GeV/A is being considered. One of the major tasks still facing us is the provision for heavy ion interactions below 3 GeV/A. The ROOT interface is being developed in conjunction with the