Initial conditions and modeling for simulations of shock driven turbulent material mixing
Grinstein, Fernando F.
2016-11-17
Here, we focus on the simulation of shock-driven material mixing driven by flow instabilities and initial conditions (IC). Beyond complex multi-scale resolution issues of shocks and variable density turbulence, me must address the equally difficult problem of predicting flow transition promoted by energy deposited at the material interfacial layer during the shock interface interactions. Transition involves unsteady large-scale coherent-structure dynamics capturable by a large eddy simulation (LES) strategy, but not by an unsteady Reynolds-Averaged Navier–Stokes (URANS) approach based on developed equilibrium turbulence assumptions and single-point-closure modeling. On the engineering end of computations, such URANS with reduced 1D/2D dimensionality and coarsermore » grids, tend to be preferred for faster turnaround in full-scale configurations.« less
Detached Eddy Simulation for the F-16XL Aircraft Configuration
NASA Technical Reports Server (NTRS)
Elmiligui, Alaa; Abdol-Hamid, Khaled; Parlette, Edward B.
2015-01-01
Numerical simulations for the flow around the F-16XL configuration as a contribution to the Cranked Arrow Wing Aerodynamic Project International 2 (CAWAPI-2) have been performed. The NASA Langley Tetrahedral Unstructured Software System (TetrUSS) with its USM3D solver was used to perform the unsteady flow field simulations for the subsonic high angle-of-attack case corresponding to flight condition (FC) 25. Two approaches were utilized to capture the unsteady vortex flow over the wing of the F-16XL. The first approach was to use Unsteady Reynolds-Averaged Navier-Stokes (URANS) coupled with standard turbulence closure models. The second approach was to use Detached Eddy Simulation (DES), which creates a hybrid model that attempts to combine the most favorable elements of URANS models and Large Eddy Simulation (LES). Computed surface static pressure profiles are presented and compared with flight data. Time-averaged and instantaneous results obtained on coarse, medium and fine grids are compared with the flight data. The intent of this study is to demonstrate that the DES module within the USM3D solver can be used to provide valuable data in predicting vortex-flow physics on a complex configuration.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Vijayakumar, Ganesh; Brasseur, James; Lavely, Adam
We describe the response of the NREL 5 MW wind turbine blade boundary layer to the passage of atmospheric turbulence using blade-boundary-layer-resolved computational fluid dynamics with hybrid URANS-LES modeling.
Deng, Qi; Hui, Dafeng; Wang, Junming; Iwuozo, Stephen; Yu, Chih-Li; Jima, Tigist; Smart, David; Reddy, Chandra; Dennis, Sam
2015-01-01
A three-year field experiment was conducted to examine the responses of corn yield and soil nitrous oxide (N2O) emission to various management practices in middle Tennessee. The management practices include no-tillage + regular applications of urea ammonium nitrate (NT-URAN); no-tillage + regular applications of URAN + denitrification inhibitor (NT-inhibitor); no-tillage + regular applications of URAN + biochar (NT-biochar); no-tillage + 20% applications of URAN + chicken litter (NT-litter), no-tillage + split applications of URAN (NT-split); and conventional tillage + regular applications of URAN as a control (CT-URAN). Fertilizer equivalent to 217 kg N ha(-1) was applied to each of the experimental plots. Results showed that no-tillage (NT-URAN) significantly increased corn yield by 28% over the conventional tillage (CT-URAN) due to soil water conservation. The management practices significantly altered soil N2O emission, with the highest in the CT-URAN (0.48 mg N2O m(-2) h(-1)) and the lowest in the NT-inhibitor (0.20 mg N2O m(-2) h(-1)) and NT-biochar (0.16 mg N2O m(-2) h(-1)) treatments. Significant exponential relationships between soil N2O emission and water filled pore space were revealed in all treatments. However, variations in soil N2O emission among the treatments were positively correlated with the moisture sensitivity of soil N2O emission that likely reflects an interactive effect between soil properties and WFPS. Our results indicated that improved fertilizer and soil management have the potential to maintain highly productive corn yield while reducing greenhouse gas emissions.
Deng, Qi; Hui, Dafeng; Wang, Junming; Iwuozo, Stephen; Yu, Chih-Li; Jima, Tigist; Smart, David; Reddy, Chandra; Dennis, Sam
2015-01-01
Background A three-year field experiment was conducted to examine the responses of corn yield and soil nitrous oxide (N2O) emission to various management practices in middle Tennessee. Methodology/Principal Findings The management practices include no-tillage + regular applications of urea ammonium nitrate (NT-URAN); no-tillage + regular applications of URAN + denitrification inhibitor (NT-inhibitor); no-tillage + regular applications of URAN + biochar (NT-biochar); no-tillage + 20% applications of URAN + chicken litter (NT-litter), no-tillage + split applications of URAN (NT-split); and conventional tillage + regular applications of URAN as a control (CT-URAN). Fertilizer equivalent to 217 kg N ha-1 was applied to each of the experimental plots. Results showed that no-tillage (NT-URAN) significantly increased corn yield by 28% over the conventional tillage (CT-URAN) due to soil water conservation. The management practices significantly altered soil N2O emission, with the highest in the CT-URAN (0.48 mg N2O m-2 h-1) and the lowest in the NT-inhibitor (0.20 mg N2O m-2 h-1) and NT-biochar (0.16 mg N2O m-2 h-1) treatments. Significant exponential relationships between soil N2O emission and water filled pore space were revealed in all treatments. However, variations in soil N2O emission among the treatments were positively correlated with the moisture sensitivity of soil N2O emission that likely reflects an interactive effect between soil properties and WFPS. Conclusion/Significance Our results indicated that improved fertilizer and soil management have the potential to maintain highly productive corn yield while reducing greenhouse gas emissions. PMID:25923716
Sahu, M; Gupta, Santosh K; Jain, D; Saxena, M K; Kadam, R M
2018-04-15
An effort was taken to carry our speciation study of uranium ion in technologically important cerate host Sr 2 CeO 4 using time resolved photoluminescence spectroscopy. Such studies are not relevant only to nuclear industry but can give rich insight into fundamentals of 5f electron chemistry in solid state systems. In this work both undoped and varied amount of uranium doped Sr 2 CeO 4 compound is synthesized using complex polymerization method and is characterized systematically using X-ray diffraction (XRD), Raman spectroscopy, impedance spectroscopy and scanning electron microscopy (SEM). Both XRD and Raman spectroscopy confirmed the formation of pure Sr 2 CeO 4 which has tendency to decompose peritectically to SrCeO 3 and SrO at higher temperature. Uranium doping is confirmed by XRD. Uranium exhibits a rich chemistry owing to its variable oxidation state from +3 to +6. Each of them exhibits distinct luminescence properties either due to f-f transitions or ligand to metal charge transfer (LMCT). We have taken Sr 2 CeO 4 as a model host lattice to understand the photophysical characteristics of uranium ion in it. Emission spectroscopy revealed the stabilization of uranium as U (VI) in the form of UO 6 6- (octahedral uranate) in Sr 2 CeO 4 . Emission kinetics study reflects that uranate ions are not homogeneously distributed in Sr 2 CeO 4 and it has two different environments due to its stabilization at both Sr 2+ as well as Ce 4+ site. The lifetime population analysis interestingly pinpointed that majority of uranate ion resided at Ce 4+ site. The critical energy-transfer distance between the uranate ion was determined based on which the concentration quenching mechanism was attributed to electric multipolar interaction. These studies are very important in designing Sr 2 CeO 4 based optoelectronic material as well exploring it for actinides studies. Copyright © 2018 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Sahu, M.; Gupta, Santosh K.; Jain, D.; Saxena, M. K.; Kadam, R. M.
2018-04-01
An effort was taken to carry our speciation study of uranium ion in technologically important cerate host Sr2CeO4 using time resolved photoluminescence spectroscopy. Such studies are not relevant only to nuclear industry but can give rich insight into fundamentals of 5f electron chemistry in solid state systems. In this work both undoped and varied amount of uranium doped Sr2CeO4 compound is synthesized using complex polymerization method and is characterized systematically using X-ray diffraction (XRD), Raman spectroscopy, photoluminescence spectroscopy and scanning electron microscopy (SEM). Both XRD and Raman spectroscopy confirmed the formation of pure Sr2CeO4 which has tendency to decompose peritectically to SrCeO3 and SrO at higher temperature. Uranium doping is confirmed by XRD. Uranium exhibits a rich chemistry owing to its variable oxidation state from +3 to +6. Each of them exhibits distinct luminescence properties either due to f-f transitions or ligand to metal charge transfer (LMCT). We have taken Sr2CeO4 as a model host lattice to understand the photophysical characteristics of uranium ion in it. Emission spectroscopy revealed the stabilization of uranium as U (VI) in the form of UO66- (octahedral uranate) in Sr2CeO4. Emission kinetics study reflects that uranate ions are not homogeneously distributed in Sr2CeO4 and it has two different environments due to its stabilization at both Sr2+ as well as Ce4+ site. The lifetime population analysis interestingly pinpointed that majority of uranate ion resided at Ce4+ site. The critical energy-transfer distance between the uranate ion was determined based on which the concentration quenching mechanism was attributed to electric multipolar interaction. These studies are very important in designing Sr2CeO4 based optoelectronic material as well exploring it for actinides studies.
NASA Astrophysics Data System (ADS)
Dawood, Yehia H.; Harbi, Hesham M.; Abd El-Naby, Hamdy H.
2010-01-01
In this study, we report kasolite Pb(UO 2)SiO 4·(H 2O) for the first time as a main uranyl mineral in the mineralized aplite-pegmatite of Jabal Sayid, Hijaz region. It commonly forms clusters of yellow acicular crystals in the voids and fractures. The mineral chemistry and mineralogical characteristics of kasolite were investigated using different techniques. Calcium, iron and phosphorus are detected in kasolite in addition to its major constituents; uranium, lead and silicon. Lead does not exist as a radiogenic product and not even as a substitute for uranium in the mineral structure. Alternatively, galena mineralization could be considered as a source for lead. The fluoride and carbonate complexes played a significant role in the formation of kasolite. High temperature hydrothermal solutions reacted with pre-existing uranium-bearing metamictized accessory minerals such as pyrochlore, U-rich thorite and zircon to form uranous fluoride complexes. These complexes are predominant in reducing environment and at pH 4. When the fluids approached the surface passing through fracture system, the oxygen fugacity ( fO 2) and the pH increased because of the loss of volatile components. At these conditions, uranous fluorides would convert to uranyl fluoride complexes UO 2F 3-. Further decrease in temperature was associated with the decay of the activity of fluorine ion by the dilution of hydrothermal solutions and precipitation of fluorite. At this condition, uranyl-carbonate complexes are favoured. These complexes were combined later with silica and lead to form kasolite.
Sensitivity of LES results from turbine rim seals to changes in grid resolution and sector size
NASA Astrophysics Data System (ADS)
O'Mahoney, T.; Hills, N.; Chew, J.
2012-07-01
Large-Eddy Simulations (LES) were carried out for a turbine rim seal and the sensitivity of the results to changes in grid resolution and the size of the computational domain are investigated. Ingestion of hot annulus gas into the rotor-stator cavity is compared between LES results and against experiments and Unsteady Reynolds-Averaged Navier-Stokes (URANS) calculations. The LES calculations show greater ingestion than the URANS calculation and show better agreement with experiments. Increased grid resolution shows a small improvement in ingestion predictions whereas increasing the sector model size has little effect on the results. The contrast between the different CFD models is most stark in the inner cavity, where the URANS shows almost no ingestion. Particular attention is also paid to the presence of low frequency oscillations in the disc cavity. URANS calculations show such low frequency oscillations at different frequencies than the LES. The oscillations also take a very long time to develop in the LES. The results show that the difficult problem of estimating ingestion through rim seals could be overcome by using LES but that the computational requirements were still restrictive.
NASA Astrophysics Data System (ADS)
Griffiths, Trevor R.; Volkovich, Vladimir A.
An extensive review of the literature on the high temperature reactions (both in melts and in the solid state) of uranium oxides (UO 2, U 3O 8 and UO 3) resulting in the formation of insoluble alkali metal (Li to Cs) uranates is presented. Their uranate(VI) and uranate(V) compounds are examined, together with mixed and oxygen-deficient uranates. The reactions of uranium oxides with carbonates, oxides, per- and superoxides, chlorides, sulfates, nitrates and nitrites under both oxidising and non-oxidising conditions are critically examined and systematised, and the established compositions of a range of uranate(VI) and (V) compounds formed are discussed. Alkali metal uranates(VI) are examined in detail and their structural, physical, thermodynamic and spectroscopic properties considered. Chemical properties of alkali metal uranates(VI), including various methods for their reduction, are also reported. Errors in the current theoretical treatment of uranate(VI) spectra are identified and the need to develop routes for the preparation of single crystals is stressed.
NASA Astrophysics Data System (ADS)
Imai, M.; Lecacheux, A.; Higgins, C. A.; Clarke, T.; Panchenko, M.; Brazhenko, A. I.; Frantsuzenko, A. V.; Konovalenko, A. A.; Imai, K.
2015-12-01
From December 2014 to March 2015, Jupiter's decametric (DAM) radio observations were carried out by using simultaneously three powerful low-frequency radio telescopes: Long Wavelength Array One (LWA1), Socorro, USA; Nançay Decameter Array (NDA), Nançay, France; and URAN2 telescope, Poltava, Ukraine. Baselines are 10000, 8600, and 2400 kilometers for LWA1-URAN2, NDA-LWA1, and URAN2-NDA, respectively. One Io-B and two Io-A emissions were simultaneously observed. Using cross-correlation analysis of obtained spectrograms, it was found that, as a function of lag time in a pair of two stations, Io-B (mainly S-bursts) and Io-A (L-bursts) show different kinds of cross-correlation coefficients, with sharp and broad peaks, respectively. By measuring lag times between LWA1-URAN2, NDA-LWA1, and URAN2-NDA pairs, it can be tested if either flashlight- or beacon-like beaming is emanated from Jupiter. Measurements of beaming width are also analyzed. Most probable beaming scenarios for Io-B and -A events are suggested.
Detailed Validation Assessment of Turbine Stage Disc Cavity Rotating Flows
NASA Astrophysics Data System (ADS)
Kanjiyani, Shezan
The subject of this thesis is concerned with the amount of cooling air assigned to seal high pressure turbine rim cavities which is critical for performance as well as component life. Insufficient air leads to excessive hot annulus gas ingestion and its penetration deep into the cavity compromising disc life. Excessive purge air, adversely affects performance. Experiments on a rotating turbine stage rig which included a rotor-stator forward disc cavity were performed at Arizona State University. The turbine rig has 22 vanes and 28 blades, while the rim cavity is composed of a single-tooth rim lab seal and a rim platform overlap seal. Time-averaged static pressures were measured in the gas path and the cavity, while mainstream gas ingestion into the cavity was determined by measuring the concentration distribution of tracer gas (carbon dioxide). Additionally, particle image velocimetry (PIV) was used to measure fluid velocity inside the rim cavity between the lab seal and the overlap. The data from the experiments were compared to an 360-degree unsteady RANS (URANS) CFD simulations. Although not able to match the time-averaged test data satisfactorily, the CFD simulations brought to light the unsteadiness present in the flow during the experiment which the slower response data did not fully capture. To interrogate the validity of URANS simulations in capturing complex rotating flow physics, the scope of this work also included to validating the CFD tool by comparing its predictions against experimental LDV data in a closed rotor-stator cavity. The enclosed cavity has a stationary shroud, a rotating hub, and mass flow does not enter or exit the system. A full 360 degree numerical simulation was performed comparing Fluent LES, with URANS turbulence models. Results from these investigations point to URANS state of art under-predicting closed cavity tangential velocity by 32% to 43%, and open rim cavity effectiveness by 50% compared to test data. The goal of this thesis is to assess the validity of URANS turbulence models in more complex rotating flows, compare accuracy with LES simulations, suggest CFD settings to better simulate turbine stage mainstream/disc cavity interaction with ingestion, and recommend experimentation techniques.
Michelini, Maria Del Carmen; Marçalo, Joaquim; Russo, Nino; Gibson, John K
2010-04-19
Bimolecular reactions of uranium oxide molecular anions with methanol have been studied experimentally, by Fourier transform ion cyclotron resonance mass spectrometry, and computationally, by density functional theory (DFT). The primary goals were to provide fundamental insights into mechanistic and structural details of model reactions of uranium oxides with organics, and to examine the validity of theoretical modeling of these types of reactions. The ions UO(3)(-), UO(4)(-), and UO(4)H(-) each reacted with methanol to give a singular product; the primary products each exhibited sequential reactions with two additional methanol molecules to again give singular products. The observed reactions were elimination of water, formaldehyde, or hydrogen, and in one case addition of a methanol molecule. The potential energy profiles were computed for each reaction, and isotopic labeling experiments were performed to probe the validity of the computed mechanisms and structures-in each case where the experiments could be compared with the theory there was concurrence, clearly establishing the efficacy of the employed DFT methodologies for these and related reaction systems. The DFT results were furthermore in accord with the surprisingly inert nature of UO(2)(-). The results provide a basis to understand mechanisms of key reactions of uranium oxides with organics, and a foundation to extend DFT methodologies to more complex actinide systems which are not amenable to such direct experimental studies.
Unsteady Thick Airfoil Aerodynamics: Experiments, Computation, and Theory
NASA Technical Reports Server (NTRS)
Strangfeld, C.; Rumsey, C. L.; Mueller-Vahl, H.; Greenblatt, D.; Nayeri, C. N.; Paschereit, C. O.
2015-01-01
An experimental, computational and theoretical investigation was carried out to study the aerodynamic loads acting on a relatively thick NACA 0018 airfoil when subjected to pitching and surging, individually and synchronously. Both pre-stall and post-stall angles of attack were considered. Experiments were carried out in a dedicated unsteady wind tunnel, with large surge amplitudes, and airfoil loads were estimated by means of unsteady surface mounted pressure measurements. Theoretical predictions were based on Theodorsen's and Isaacs' results as well as on the relatively recent generalizations of van der Wall. Both two- and three-dimensional computations were performed on structured grids employing unsteady Reynolds-averaged Navier-Stokes (URANS). For pure surging at pre-stall angles of attack, the correspondence between experiments and theory was satisfactory; this served as a validation of Isaacs theory. Discrepancies were traced to dynamic trailing-edge separation, even at low angles of attack. Excellent correspondence was found between experiments and theory for airfoil pitching as well as combined pitching and surging; the latter appears to be the first clear validation of van der Wall's theoretical results. Although qualitatively similar to experiment at low angles of attack, two-dimensional URANS computations yielded notable errors in the unsteady load effects of pitching, surging and their synchronous combination. The main reason is believed to be that the URANS equations do not resolve wake vorticity (explicitly modeled in the theory) or the resulting rolled-up un- steady flow structures because high values of eddy viscosity tend to \\smear" the wake. At post-stall angles, three-dimensional computations illustrated the importance of modeling the tunnel side walls.
Flow Control Under Low-Pressure Turbine Conditions Using Pulsed Jets
NASA Technical Reports Server (NTRS)
Volino, Ralph J.; Ibrahim, Mounir B.
2012-01-01
This publication is the final report of research performed under an NRA/Cooperative Interagency Agreement, and includes a supplemental CD-ROM with detailed data. It is complemented by NASA/CR-2012-217416 and NASA/CR-2012-217417 which include a Ph.D. Dissertation and an M.S. thesis respectively, performed under this contract. In this study the effects of unsteady wakes and flow control using vortex generator jets (VGJs) were studied experimentally and computationally on the flow over the L1A low pressure turbine (LPT) airfoil. The experimental facility was a six passage linear cascade in a low speed wind tunnel at the U.S. Naval Academy. In parallel, computational work using the commercial code FLUENT (ANSYS, Inc.) was performed at Cleveland State University, using Unsteady Reynolds Averaged Navier Stokes (URANS) and Large Eddy Simulations (LES) methods. In the first phase of the work, the baseline flow was documented under steady inflow conditions without flow control. URANS calculations were done using a variety of turbulence models. In the second phase of the work, flow control was added using steady and pulsed vortex generator jets. The VGJs successfully suppressed separation and reduced aerodynamic losses. Pulsed operation was more effective and mass flow requirements are very low. Numerical simulations of the VGJs cases showed that URANS failed to capture the effect of the jets. LES results were generally better. In the third phase, effects of unsteady wakes were studied. Computations with URANS and LES captured the wake effect and generally predicted separation and reattachment to match the experiments. Quantitatively the results were mixed. In the final phase of the study, wakes and VGJs were combined and synchronized using various timing schemes. The timing of the jets with respect to the wakes had some effect, but in general once the disturbance frequency was high enough to control separation, the timing was not very important.
Flow Control Under Low-Pressure Turbine Conditions Using Pulsed Jets: Experimental Data Archive
NASA Technical Reports Server (NTRS)
Volino, Ralph J.; Ibrahim, Mounir B.
2012-01-01
This publication is the final report of research performed under an NRA/Cooperative Interagency Agreement, and includes a supplemental CD-ROM with detailed data. It is complemented by NASA/CR-2012-217416 and NASA/CR-2012-217417 which include a Ph.D. Dissertation and an M.S. thesis respectively, performed under this contract. In this study the effects of unsteady wakes and flow control using vortex generator jets (VGJs) were studied experimentally and computationally on the flow over the L1A low pressure turbine (LPT) airfoil. The experimental facility was a six passage linear cascade in a low speed wind tunnel at the U.S. Naval Academy. In parallel, computational work using the commercial code FLUENT (ANSYS, Inc.) was performed at Cleveland State University, using Unsteady Reynolds Averaged Navier Stokes (URANS) and Large Eddy Simulations (LES) methods. In the first phase of the work, the baseline flow was documented under steady inflow conditions without flow control. URANS calculations were done using a variety of turbulence models. In the second phase of the work, flow control was added using steady and pulsed vortex generator jets. The VGJs successfully suppressed separation and reduced aerodynamic losses. Pulsed operation was more effective and mass flow requirements are very low. Numerical simulations of the VGJs cases showed that URANS failed to capture the effect of the jets. LES results were generally better. In the third phase, effects of unsteady wakes were studied. Computations with URANS and LES captured the wake effect and generally predicted separation and reattachment to match the experiments. Quantitatively the results were mixed. In the final phase of the study, wakes and VGJs were combined and synchronized using various timing schemes. The timing of the jets with respect to the wakes had some effect, but in general once the disturbance frequency was high enough to control separation, the timing was not very important. This is the supplemental CD-ROM
Computational Simulations of Convergent Nozzles for the AIAA 1st Propulsion Aerodynamics Workshop
NASA Technical Reports Server (NTRS)
Dippold, Vance F., III
2014-01-01
Computational Fluid Dynamics (CFD) simulations were completed for a series of convergent nozzles in participation of the American Institute of Aeronautics and Astronautics (AIAA) 1st Propulsion Aerodynamics Workshop. The simulations were performed using the Wind-US flow solver. Discharge and thrust coefficients were computed for four axisymmetric nozzles with nozzle pressure ratios (NPR) ranging from 1.4 to 7.0. The computed discharge coefficients showed excellent agreement with available experimental data; the computed thrust coefficients captured trends observed in the experimental data, but over-predicted the thrust coefficient by 0.25 to 1.0 percent. Sonic lines were computed for cases with NPR >= 2.0 and agreed well with experimental data for NPR >= 2.5. Simulations were also performed for a 25 deg. conic nozzle bifurcated by a flat plate at NPR = 4.0. The jet plume shock structure was compared with and without the splitter plate to the experimental data. The Wind-US simulations predicted the shock structure well, though lack of grid resolution in the plume reduced the sharpness of the shock waves. Unsteady Reynolds-Averaged Navier-Stokes (URANS) simulations and Detached Eddy Simulations (DES) were performed at NPR = 1.6 for the 25 deg conic nozzle with splitter plate. The simulations predicted vortex shedding from the trailing edge of the splitter plate. However, the vortices of URANS and DES solutions appeared to dissipate earlier than observed experimentally. It is believed that a lack of grid resolution in the region of the vortex shedding may have caused the vortices to break down too soon
NASA Astrophysics Data System (ADS)
Melnik, V.; Konovalenko, A.; Brazhenko, A.; Briand, C.; Dorovskyy, V.; Zarka, P.; Denis, L.; Bulatzen, V.; Frantzusenko, A.; Rucker, H.; Stanislavskyy, A.
2012-09-01
From 25 June till 12 August 2011 sporadic solar radio emission was observed simultaneously by three separate radio telescopes: UTR-2 (Kharkov, Ukraine), URAN-2 (Poltava, Ukraine) and NDA (Nancay, France). During these observations some interesting phenomena were observed. Some of them are discussed in this paper.
PROCESS FOR RECOVERY OF URANIUM VALUES FROM IMPURE SOLUTIONS THEREOF
Kilner, S.B.
1959-11-01
A process is presented for the recovery of uraninm values from impure solutions which are obtained, for example, by washing residual uranium salt or uranium metal deposits from stainless steel surfaces using an aqueous or certain acidic aqueous solutions. The solutions include uranyl and oxidized iron, chromium, nickel, and copper ions and may contain manganese, zinc, and silver ions. In accordance with one procedure. the uranyl ions are reduced to the uranous state, and the impurity ions are complexed with cyanide under acidic conditions. The solution is then treated with ammonium hydroxide or alkali metal hydroxide to precipitate uranous hydroxide away from the complexed impurity ions in the solution. Alternatively, an excess of alkali metal cyanide is added to the reduced solution until the solution becomes sufficiently alkaline for the uranons hydroxide to precipitate. An essential feature in operating the process is in maintaining the pH of the solution sufficiently acid during the complexing operation to prevent the precipitation of the impurity metal hydroxides.
Kilner, S.B.
1959-12-29
A method is presented for separating and recovering uranium from a complex mixure of impurities. The uranium is dissolved to produce an aqueous acidic solution including various impurities. In accordance with one method, with the uranium in the uranyl state, hydrogen cyanide is introduced into the solution to complex the impurities. Subsequently, ammonia is added to the solution to precipitate the uraniunn as ammonium diuranate away from the impurities in the solution. Alternatively, the uranium is precipitated by adding an alkaline metal hydroxide. In accordance with the second method, the uranium is reduced to the uranous state in the solution. The reduced solution is then treated with solid alkali metal cyanide sufficient to render the solution about 0.1 to 1.0 N in cyanide ions whereat cyanide complex ions of the metal impurities are produced and the uranium is simultaneously precipituted as uranous hydroxide. Alternatively, hydrogen cyanide may be added to the reduced solution and the uranium precipitated subsequently by adding ammonium hydroxide or an alkali metal hydroxide. Other refinements of the method are also disclosed.
NASA Astrophysics Data System (ADS)
Melnik, V. N.; Konovalenko, A. A.; Rucker, H. O.; Brazhenko, A. I.; Briand, C.; Dorovskyy, V. V.; Zarka, P.; Denis, L.; Bulatzen, V. G.; Frantzusenko, A. V.; Stanislavskyy, A. A.
2012-04-01
From 25 June till 12 August 2011 sporadic solar radio emission was observed simultaneously by three separate radio telescopes: UTR-2 (Kharkov, Ukraine), URAN-2 (Poltava, Ukraine) and NDA (Nancay, France). During these observations several type II bursts with double and triple harmonics were registered, as well as type II bursts with complex herringbone structure. The events of particular interest were type II bursts registered on 9 and 11 August 2011. These bursts had opposite sign of circular polarization at different parts of their dynamic spectra. In our opinion we registered the emissions, which came from the different parts of the shock propagating through the solar corona. We have observed also groups of type III bursts merged into one burst, type III bursts with triple harmonics and type III bursts with "split" polarization. In addition some unusual solar bursts were registered: storms of strange narrow-band (up to 500kHz) bursts with high polarization degree (about 80%), decameter spikes of extremely short durations (200-300ms), "tadpole-like" bursts with durations of 1-2s and polarization degree up to 60%.
Computational fluid dynamics simulation of sound propagation through a blade row.
Zhao, Lei; Qiao, Weiyang; Ji, Liang
2012-10-01
The propagation of sound waves through a blade row is investigated numerically. A wave splitting method in a two-dimensional duct with arbitrary mean flow is presented, based on which pressure amplitude of different wave mode can be extracted at an axial plane. The propagation of sound wave through a flat plate blade row has been simulated by solving the unsteady Reynolds average Navier-Stokes equations (URANS). The transmission and reflection coefficients obtained by Computational Fluid Dynamics (CFD) are compared with semi-analytical results. It indicates that the low order URANS scheme will cause large errors if the sound pressure level is lower than -100 dB (with as reference pressure the product of density, main flow velocity, and speed of sound). The CFD code has sufficient precision when solving the interaction of sound wave and blade row providing the boundary reflections have no substantial influence. Finally, the effects of flow Mach number, blade thickness, and blade turning angle on sound propagation are studied.
AFRL’s ALREST Physics-Based Combustion Stability Program
2012-11-08
enduring challenge because of the inherent complexities in the physics of multiphase turbulent flames. The present paper provides the Air Force...Combustor F i d e l i t y URANS LES Steady RANS HLES Current SOA Capability with 2000 cores Capability at Program End in 2015 (2,000 cores+GPUs) Capability...Unlimited ALREST Validation Cases “Final Exam ” Hydrogen Stable Single Element (PSU) Stable Single Element Methane (Singla) Supercritical Non
On the Harmonic Coupling of Components in Pairs of IIIb-III Bursts at Decameter Wavelengths
NASA Astrophysics Data System (ADS)
Brazhenko, A. I.; Melnik, V. N.; Frantsuzenko, A. V.; Dorovskyy, V. V.; Rucker, H. O.; Panchenko, M.
2015-06-01
The properties of IIIb-III pairs observed by the URAN-2 radioThe properties of IIIb-III pairs observed by the URAN-2 radiotelescope at frequencies 16-32 MHz are analyzed. Observations of these bursts were hold in April, June and September 2011. Durations, frequency drift rates, simultaneous frequency ratio of pairs components and their polarizations are analyzed. Pro and contra of IIIb-III harmonic connection are discussed.
Pawar, Prabhakar R
2013-10-15
Surface water samples were collected from substations along Sheva creek and Dharamtar creek mangrove ecosystems of Uran (Raigad), Navi Mumbai, west coast of India. Water samples were collected fortnightly from April 2009 to March 2011 during spring low and high tides and were analyzed for pH, Temperature, Turbidity, Total solids (TS), Total dissolved solids (TDS), Total suspended solids (TSS), Dissolved oxygen (DO), Biochemical oxygen demand (BOD), Carbon dioxide (CO2), Chemical oxygen demand (COD), Salinity, Orthophosphate (O-PO4), Nitrite-nitrogen (NO2-N), Nitrate-nitrogen (NO3-N), and Silicates. Variables like pH, turbidity, TDS, salinity, DO, and BOD show seasonal variations. Higher content of O-PO4, NO3-N, and silicates is recorded due to discharge of domestic wastes and sewage, effluents from industries, oil tanking depots and also from maritime activities of Jawaharlal Nehru Port Trust (JNPT), hectic activities of Container Freight Stations (CFS), and other port wastes. This study reveals that water quality from mangrove ecosystems of Uran is deteriorating due to industrial pollution and that mangrove from Uran is facing the threat due to anthropogenic stress. Copyright © 2013 Elsevier Ltd. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yuan, Zhongwei; Yan, Taihong; Zheng, Weifang
2013-07-01
The electrochemical reduction of uranyl nitrate is a green, mild way to make uranous ions. Undivided electrolyzers whose maintenance is less but their conversion ratio and current efficiency are low, have been chosen. However, at the beginning of undivided electrolysis, high current efficiency can also be maintained. Divided electrolyzers' conversion ratio and current efficiency is much higher because the re-oxidation of uranous on anode is avoided, but their maintenance costs are more, because in radioactive environment the membrane has to be changed after several operations. In this paper, a combined method of uranous production is proposed which consists of 2more » stages: undivided electrolysis (early stage) and divided electrolysis (late stage) to benefit from the advantages of both electrolysis modes. The performance of the combined method was tested. The results show that in combined mode, after 200 min long electrolysis (80 min undivided electrolysis and 120 min divided electrolysis), U(IV) yield can achieve 92.3% (500 ml feed, U 199 g/l, 72 cm{sup 2} cathode, 120 mA/cm{sup 2}). Compared with divided mode, about 1/3 working time in divided electrolyzer is reduced to achieve the same U(IV) yield. If 120 min long undivided electrolysis was taken, more than 1/2 working time can be reduced in divided electrolyzer, which means that about half of the maintenance cost can also be reduced. (authors)« less
Guidelines for Computing Longitudinal Dynamic Stability Characteristics of a Subsonic Transport
NASA Technical Reports Server (NTRS)
Thompson, Joseph R.; Frank, Neal T.; Murphy, Patrick C.
2010-01-01
A systematic study is presented to guide the selection of a numerical solution strategy for URANS computation of a subsonic transport configuration undergoing simulated forced oscillation about its pitch axis. Forced oscillation is central to the prevalent wind tunnel methodology for quantifying aircraft dynamic stability derivatives from force and moment coefficients, which is the ultimate goal for the computational simulations. Extensive computations are performed that lead in key insights of the critical numerical parameters affecting solution convergence. A preliminary linear harmonic analysis is included to demonstrate the potential of extracting dynamic stability derivatives from computational solutions.
PROCESS OF SEPARATING PLUTONIUM FROM URANIUM
Brown, H.S.; Hill, O.F.
1958-09-01
A process is presented for recovering plutonium values from aqueous solutions. It comprises forming a uranous hydroxide precipitate in such a plutonium bearing solution, at a pH of at least 5. The plutonium values are precipitated with and carried by the uranium hydroxide. The carrier precipitate is then redissolved in acid solution and the pH is adjusted to about 2.5, causing precipitation of the uranous hydroxide but leaving the still soluble plutonium values in solution.
NASA Astrophysics Data System (ADS)
Verma, Sudeep; Dewan, Anupam
2018-01-01
The Partially-Averaged Navier-Stokes (PANS) approach has been applied for the first time to model turbulent flow and heat transfer in an ideal Czochralski set up with the realistic boundary conditions. This method provides variable level of resolution ranging from the Reynolds-Averaged Navier-Stokes (RANS) modelling to Direct Numerical Simulation (DNS) based on the filter control parameter. For the present case, a low-Re PANS model has been developed for Czochralski melt flow, which includes the effect of coriolis, centrifugal, buoyant and surface tension induced forces. The aim of the present study is to assess improvement in results on switching to PANS modelling from unsteady RANS (URANS) approach on the same computational mesh. The PANS computed results were found to be in good agreement with the reported experimental, DNS and Large Eddy Simulation (LES) data. A clear improvement in computational accuracy is observed in switching from the URANS approach to the PANS methodology. The computed results further improved with a reduction in the PANS filter width. Further the capability of the PANS model to capture key characteristics of the Czochralski crystal growth is also highlighted. It was observed that the PANS model was able to resolve the three-dimensional turbulent nature of the melt, characteristic flow structures arising due to flow instabilities and generation of thermal plumes and vortices in the Czochralski melt.
Effect of Turbulence Modeling on an Excited Jet
NASA Technical Reports Server (NTRS)
Brown, Clifford A.; Hixon, Ray
2010-01-01
The flow dynamics in a high-speed jet are dominated by unsteady turbulent flow structures in the plume. Jet excitation seeks to control these flow structures through the natural instabilities present in the initial shear layer of the jet. Understanding and optimizing the excitation input, for jet noise reduction or plume mixing enhancement, requires many trials that may be done experimentally or computationally at a significant cost savings. Numerical simulations, which model various parts of the unsteady dynamics to reduce the computational expense of the simulation, must adequately capture the unsteady flow dynamics in the excited jet for the results are to be used. Four CFD methods are considered for use in an excited jet problem, including two turbulence models with an Unsteady Reynolds Averaged Navier-Stokes (URANS) solver, one Large Eddy Simulation (LES) solver, and one URANS/LES hybrid method. Each method is used to simulate a simplified excited jet and the results are evaluated based on the flow data, computation time, and numerical stability. The knowledge gained about the effect of turbulence modeling and CFD methods from these basic simulations will guide and assist future three-dimensional (3-D) simulations that will be used to understand and optimize a realistic excited jet for a particular application.
I. S. Shklovsky and Low-Frequency Radio Astronomy
NASA Astrophysics Data System (ADS)
Konovalenko, A. A.
2017-03-01
Purpose: Proving of the high astrophysical significance of the low-frequency radio astronomy (decameter and adjacent hectometer and meter wavelengths), demonstration of the priority results of the Ukrainian low-frequency radio astronomy as well as significant contribution of I. S. Shklovsky to its development. Design/methodology/approach: The requirements to characteristics of high efficiency radio telescopes UTR-2, URAN, GURT and to sensitive and interference immune observational methods at low frequencies are formulated by using the theoretical analysis and astrophysical predictions including those I. S. Shklovsky’s. Findings: New generation radio telescopes UTR-2, URAN, GURT are created and modernized. New observational methods at low frequencies are introduced. Large-scale investigations of the Solar system, Galaxy and Methagalaxy are carried out. They have allowed to detect new objects and phenomena for the continuum, monochromatic, pulse and sporadic cosmic radio emission. The role of I. S. Shklovsky in the development of many low-frequency radio astronomy directions is noted, too. Conclusions: The unique possibilities of the low-frequency radio astronomy which gives new information about the Universe, inaccessible with the other astrophysical methods, are shown. The progress of the low-frequency radio astronomy opens the impressive possibilities for the future. It includes modernization of the largest radio telescopes UTR-2, URAN, NDA and creation of new instruments GURT, NenuFAR, LOFAR, LWA, MWA, SKA as well as making multi-antenna and ground-space experiments. The contribution of outstanding astrophysicist of the XX century I. S. Shklovsky to this part of actual astronomical science is evident, claiming for attention and will never be forgotten.
An Investigation of Transonic Resonance in a Mach 2.2 Round Convergent-Divergent Nozzle
NASA Technical Reports Server (NTRS)
Dippold, Vance F., III; Zaman, Khairul B. M. Q.
2015-01-01
Hot-wire and acoustic measurements were taken for a round convergent nozzle and a round convergent-divergent (C-D) nozzle at a jet Mach number of 0.61. The C-D nozzle had a design Mach number of 2.2. Compared to the convergent nozzle jet flow, the Mach 2.2 nozzle jet flow produced excess broadband noise (EBBN). It also produced a transonic resonance tone at 1200 Herz. Computational simulations were performed for both nozzle flows. A steady Reynolds-Averaged Navier-Stokes simulation was performed for the convergent nozzle jet flow. For the Mach 2.2 nozzle flow, a steady RANS simulation, an unsteady RANS (URANS) simulation, and an unsteady Detached Eddy Simulation (DES) were performed. The RANS simulation of the convergent nozzle showed good agreement with the hot-wire velocity and turbulence measurements, though the decay of the potential core was over-predicted. The RANS simulation of the Mach 2.2 nozzle showed poor agreement with the experimental data, and more closely resembled an ideally-expanded jet. The URANS simulation also showed qualitative agreement with the hot-wire data, but predicted a transonic resonance at 1145 Herz. The DES showed good agreement with the hot-wire velocity and turbulence data. The DES also produced a transonic tone at 1135 Herz. The DES solution showed that the destabilization of the shock-induced separation region inside the nozzle produced increased levels of turbulence intensity. This is likely the source of the EBBN.
Selective recovery of uranium from Ca-Mg uranates by chlorination
NASA Astrophysics Data System (ADS)
Pomiro, Federico J.; Gaviría, Juan P.; Quinteros, Raúl D.; Bohé, Ana E.
2017-07-01
A chlorination process is proposed for the uranium extraction and separation using Calciumsbnd Magnesium uranates such as starting reactants which were obtained by precipitation from uranyl nitrate solutions with calcium hydroxide. The study is based on thermodynamic and reaction analysis using chlorine gas as chlorination agent. The results showed that the chlorination reaction of Ca uranate is more feasible to occur than the Mg uranate. The products obtained after chlorination reactions were washed with deionized water to remove the chlorides produced and analyzed. The XRD patterns of the washed products indicated that the chlorination between 400 and 500 °C result in a single phase of calcium uranate (CaUO4) as reaction product. The formation of U3O8 and MgU3O10 was observed at temperatures between 600 °C and 700 °C for 8 hs. The optimal conditions to recover uranium were 3 l h-1 of chlorine and 10 hs of reaction at 700 °C being U3O8 the single uranium product obtained.
NASA Astrophysics Data System (ADS)
Kumar, Mithlesh; Mohapatra, M.
2016-04-01
Zinc aluminate (ZAO), a member of spinel class of inorganic compounds has been of much interest of late due to its wide range of use in catalysis, optical, electronic and ceramic industries. When doped with several lanthanides, this material has proved to be a potential host matrix for phosphors. As lanthanides suffer from poor (direct) excitation and emission cross sections, the use of a co-dopant ion can help to circumvent this and extract better emission from a lanthanide doped ZAO system. In this connection, energy transfer mechanism from uranium to europium in the ZAO host was investigated by photoluminescence spectroscopic technique. It was seen that uranium gets stabilized in the hexavalent state as UO66 - (octahedral uranate) where as the lanthanide ion, Eu is stabilized in its trivalent state in the ZAO host. In the co-doped system, an efficient energy transfer pathway from the uranate to europium ion was observed. Based upon emission and life time data a suitable mechanism was proposed for the energy transfer (quenching) process. It was proposed that after excitation by photons, the uranate ions transfer their energy to nearby 5D1 level of Eu3 + ions which non-radiatively de-excites to the corresponding lower levels of 5D0. Further this 5D0 level decays in a radiative mode to the 7F manifold giving the characteristic emission profile of trivalent Eu. It was proposed that both static and dynamic types of energy transfer mechanism were responsible for this process.
NASA Technical Reports Server (NTRS)
Jenkins, Luther N.; Khorrami, Mehdi R.; Choudhari, Meelan M.; McGinley, Catherine B.
2005-01-01
A joint computational and experimental study has been performed at NASA Langley Research Center to investigate the unsteady flow generated by the components of an aircraft landing gear system. Because the flow field surrounding a full landing gear is so complex, the study was conducted on a simplified geometry consisting of two cylinders in tandem arrangement to isolate and characterize the pertinent flow phenomena. This paper focuses on the experimental effort where surface pressures, 2-D Particle Image Velocimetry, and hot-wire anemometry were used to document the flow interaction around the two cylinders at a Reynolds Number of 1.66 x 10(exp 5), based on cylinder diameter, and cylinder spacing-todiameter ratios, L/D, of 1.435 and 3.70. Transition strips were applied to the forward cylinder to produce a turbulent boundary layer upstream of the flow separation. For these flow conditions and L/D ratios, surface pressures on both the forward and rear cylinders show the effects of L/D on flow symmetry, base pressure, and the location of flow separation and attachment. Mean velocities and instantaneous vorticity obtained from the PIV data are used to examine the flow structure between and aft of the cylinders. Shedding frequencies and spectra obtained using hot-wire anemometry are presented. These results are compared with unsteady, Reynolds-Averaged Navier-Stokes (URANS) computations for the same configuration in a companion paper by Khorrami, Choudhari, Jenkins, and McGinley (2005). The experimental dataset produced in this study provides information to better understand the mechanisms associated with component interaction noise, develop and validate time-accurate computer methods used to calculate the unsteady flow field, and assist in modeling of the radiated noise from landing gears.
Room temperature electrodeposition of actinides from ionic solutions
Hatchett, David W.; Czerwinski, Kenneth R.; Droessler, Janelle; Kinyanjui, John
2017-04-25
Uranic and transuranic metals and metal oxides are first dissolved in ozone compositions. The resulting solution in ozone can be further dissolved in ionic liquids to form a second solution. The metals in the second solution are then electrochemically deposited from the second solutions as room temperature ionic liquid (RTIL), tri-methyl-n-butyl ammonium n-bis(trifluoromethansulfonylimide) [Me.sub.3N.sup.nBu][TFSI] providing an alternative non-aqueous system for the extraction and reclamation of actinides from reprocessed fuel materials. Deposition of U metal is achieved using TFSI complexes of U(III) and U(IV) containing the anion common to the RTIL. TFSI complexes of uranium were produced to ensure solubility of the species in the ionic liquid. The methods provide a first measure of the thermodynamic properties of U metal deposition using Uranium complexes with different oxidation states from RTIL solution at room temperature.
NASA Technical Reports Server (NTRS)
Biedron, Robert T.; Vatsa, Veer N.; Atkins, Harold L.
2005-01-01
We apply an unsteady Reynolds-averaged Navier-Stokes (URANS) solver for unstructured grids to unsteady flows on moving and stationary grids. Example problems considered are relevant to active flow control and stability and control. Computational results are presented using the Spalart-Allmaras turbulence model and are compared to experimental data. The effect of grid and time-step refinement are examined.
Large-eddy and unsteady RANS simulations of a shock-accelerated heavy gas cylinder
Morgan, B. E.; Greenough, J. A.
2015-04-08
Two-dimensional numerical simulations of the Richtmyer–Meshkov unstable “shock-jet” problem are conducted using both large-eddy simulation (LES) and unsteady Reynolds-averaged Navier–Stokes (URANS) approaches in an arbitrary Lagrangian–Eulerian hydrodynamics code. Turbulence statistics are extracted from LES by running an ensemble of simulations with multimode perturbations to the initial conditions. Detailed grid convergence studies are conducted, and LES results are found to agree well with both experiment and high-order simulations conducted by Shankar et al. (Phys Fluids 23, 024102, 2011). URANS results using a k–L approach are found to be highly sensitive to initialization of the turbulence lengthscale L and to the timemore » at which L becomes resolved on the computational mesh. As a result, it is observed that a gradient diffusion closure for turbulent species flux is a poor approximation at early times, and a new closure based on the mass-flux velocity is proposed for low-Reynolds-number mixing.« less
CONCENTRATION OF Pu USING OXALATE TYPE CARRIER
Ritter, D.M.; Black, R.P.S.
1960-04-19
A method is given for dissolving and reprecipitating an oxalate carrier precipitate in a carrier precipitation process for separating and recovering plutonium from an aqueous solution. Uranous oxalate, together with plutonium being carried thereby, is dissolved in an aqueous alkaline solution. Suitable alkaline reagents are the carbonates and oxulates of the alkali metals and ammonium. An oxidizing agent selected from hydroxylamine and hydrogen peroxide is then added to the alkaline solution, thereby oxidizing uranium to the hexavalent state. The resulting solution is then acidified and a source of uranous ions provided in the acidified solution, thereby forming a second plutoniumcarrying uranous oxalate precipitate.
Unsteady Aerodynamic Modeling of A Maneuvering Aircraft Using Indicial Functions
2016-03-30
indicial functions are directly calculated using the results of unsteady Reynolds-averaged Navier - Stokes simulation and a grid-movement tool. Results are...but meanwhile, the full-order model based on Unsteady Reynolds-averaged Navier - Stokes (URANS) equation is too computationally expensive to be used...The flow solver used in this study solves the unsteady, three-dimensional and compressible Navier - Stokes equations. The equations in terms of
Reduced Order Modeling of Combustion Instability in a Gas Turbine Model Combustor
NASA Astrophysics Data System (ADS)
Arnold-Medabalimi, Nicholas; Huang, Cheng; Duraisamy, Karthik
2017-11-01
Hydrocarbon fuel based propulsion systems are expected to remain relevant in aerospace vehicles for the foreseeable future. Design of these devices is complicated by combustion instabilities. The capability to model and predict these effects at reduced computational cost is a requirement for both design and control of these devices. This work focuses on computational studies on a dual swirl model gas turbine combustor in the context of reduced order model development. Full fidelity simulations are performed utilizing URANS and Hybrid RANS-LES with finite rate chemistry. Following this, data decomposition techniques are used to extract a reduced basis representation of the unsteady flow field. These bases are first used to identify sensor locations to guide experimental interrogations and controller feedback. Following this, initial results on developing a control-oriented reduced order model (ROM) will be presented. The capability of the ROM will be further assessed based on different operating conditions and geometric configurations.
Method for the recovery of uranium values from uranium tetrafluoride
Kreuzmann, Alvin B.
1983-01-01
The invention is a novel method for the recovery of uranium from dry, particulate uranium tetrafluoride. In one aspect, the invention comprises reacting particulate uranium tetrafluoride and calcium oxide in the presence of gaseous oxygen to effect formation of the corresponding alkaline earth metal uranate and alkaline earth metal fluoride. The product uranate is highly soluble in various acidic solutions wherein the product fluoride is virtually insoluble therein. The product mixture of uranate and alkaline earth metal fluoride is contacted with a suitable acid to provide a uranium-containing solution, from which the uranium is recovered. The invention can achieve quantitative recovery of uranium in highly pure form.
Method for the recovery of uranium values from uranium tetrafluoride
Kreuzmann, A.B.
1982-10-27
The invention is a novel method for the recovery of uranium from dry, particulate uranium tetrafluoride. In one aspect, the invention comprises reacting particulate uranium tetrafluoride and calcium oxide in the presence of gaseous oxygen to effect formation of the corresponding alkaline earth metal uranate and alkaline earth metal fluoride. The product uranate is highly soluble in various acidic solutions whereas the product fluoride is virtually insoluble therein. The product mixture of uranate and alkaline earth metal fluoride is contacted with a suitable acid to provide a uranium-containing solution, from which the uranium is recovered. The invention can achieve quantitative recovery of uranium in highly pure form.
NASA Technical Reports Server (NTRS)
Westra, Doug G.; West, Jeffrey S.; Richardson, Brian R.
2015-01-01
Historically, the analysis and design of liquid rocket engines (LREs) has relied on full-scale testing and one-dimensional empirical tools. The testing is extremely expensive and the one-dimensional tools are not designed to capture the highly complex, and multi-dimensional features that are inherent to LREs. Recent advances in computational fluid dynamics (CFD) tools have made it possible to predict liquid rocket engine performance, stability, to assess the effect of complex flow features, and to evaluate injector-driven thermal environments, to mitigate the cost of testing. Extensive efforts to verify and validate these CFD tools have been conducted, to provide confidence for using them during the design cycle. Previous validation efforts have documented comparisons of predicted heat flux thermal environments with test data for a single element gaseous oxygen (GO2) and gaseous hydrogen (GH2) injector. The most notable validation effort was a comprehensive validation effort conducted by Tucker et al. [1], in which a number of different groups modeled a GO2/GH2 single element configuration by Pal et al [2]. The tools used for this validation comparison employed a range of algorithms, from both steady and unsteady Reynolds Averaged Navier-Stokes (U/RANS) calculations, large-eddy simulations (LES), detached eddy simulations (DES), and various combinations. A more recent effort by Thakur et al. [3] focused on using a state-of-the-art CFD simulation tool, Loci/STREAM, on a two-dimensional grid. Loci/STREAM was chosen because it has a unique, very efficient flamelet parameterization of combustion reactions that are too computationally expensive to simulate with conventional finite-rate chemistry calculations. The current effort focuses on further advancement of validation efforts, again using the Loci/STREAM tool with the flamelet parameterization, but this time with a three-dimensional grid. Comparisons to the Pal et al. heat flux data will be made for both RANS and Hybrid RANSLES/ Detached Eddy simulations (DES). Computation costs will be reported, along with comparison of accuracy and cost to much less expensive two-dimensional RANS simulations of the same geometry.
Unsteady Aero Computation of a 1 1/2 Stage Large Scale Rotating Turbine
NASA Technical Reports Server (NTRS)
To, Wai-Ming
2012-01-01
This report is the documentation of the work performed for the Subsonic Rotary Wing Project under the NASA s Fundamental Aeronautics Program. It was funded through Task Number NNC10E420T under GESS-2 Contract NNC06BA07B in the period of 10/1/2010 to 8/31/2011. The objective of the task is to provide support for the development of variable speed power turbine technology through application of computational fluid dynamics analyses. This includes work elements in mesh generation, multistage URANS simulations, and post-processing of the simulation results for comparison with the experimental data. The unsteady CFD calculations were performed with the TURBO code running in multistage single passage (phase lag) mode. Meshes for the blade rows were generated with the NASA developed TCGRID code. The CFD performance is assessed and improvements are recommended for future research in this area. For that, the United Technologies Research Center's 1 1/2 stage Large Scale Rotating Turbine was selected to be the candidate engine configuration for this computational effort because of the completeness and availability of the data.
NASA Astrophysics Data System (ADS)
Gan, Jiaye
The purpose of this research is to develop high fidelity numerical methods to investigate the complex aeroelasticity fluid-structural problems of aircraft and aircraft engine turbomachinery. Unsteady 3D compressible Navier-Stokes equations in generalized coordinates are solved to simulate the complex fluid dynamic problems in aeroelasticity. An efficient and low diffusion E-CUSP (LDE) scheme designed to minimize numerical dissipation is used as a Riemann solver to capture shock waves in transonic and supersonic flows. An improved hybrid turbulence modeling, delayed detached eddy simulation (DDES), is implemented to simulate shock induced separation and rotating stall flows. High order accuracy (3rd and 5th order) weighted essentially non-oscillatory (WENO) schemes for inviscid flux and a conservative 2nd and 4th order viscous flux differencing are employed. To resolve the nonlinear interaction between flow and vibrating blade structures, a fully coupled fluid-structure interaction (FSI) procedure that solves the structural modal equations and time accurate Navier-Stokes equations simultaneously is adopted. A rotor/stator sliding interpolation technique is developed to accurately capture the blade rows interaction at the interface with general grid distribution. Phase lag boundary conditions (BC) based on the time shift (direct store) method and the Fourier series phase lag BC are applied to consider the effect of phase difference for a sector of annulus simulation. Extensive validations are conducted to demonstrate high accuracy and robustness of the high fidelity FSI methodology. The accuracy and robustness of RANS, URANS and DDES turbulence models with high order schemes for predicting the lift and drag of the DLR-F6 configuration are verified. The DDES predicts the drag very well whereas the URANS model significantly over predicts the drag. DDES of a finned projectile base flows is conducted to further validate the high fidelity methods with vortical flow. The DDES is demonstrated to be superior to the URANS for the projectile flow prediction. DDES of a 3D transonic wing flutter is validated with AGARD Wing 445.6 aeroelasticity experiment at free stream Mach number varied from subsonic to supersonic. The predicted flutter boundary at different free stream Mach number including the sonic dip achieves very good agreement with the experiment. In particular, the predicted flutter boundaries at the supersonic conditions match the experiment accurately. The mechanism of sonic dip is investigated. Simulation of supersonic fluid-structural interaction of a flat panel is performed by using DDES with high order shock capturing scheme. The panel vibration induced by the shock boundary layer interaction is well resolved by the high fidelity method. The dominant panel response agrees well with the experiment in terms of the mean panel displacement and frequency. The DDES methodology is used to investigate the stall inception of NASA Stage 35 compressor. The process of rotating stall is compared between the results using both URANS and DDES with full annulus. The stall process begins with spike inception and develops to full stall. The numbers of stall cell, and the size and propagating speed of the stall cells are well captured by both URANS and DDES. Two stall cells with 42% rotor rotating speed are resolved by DDES and one stall cell with 90% rotor rotating speed by URANS. It is not conclusive which method is more accurate since there is no experimental data, but the DDES does show more realistic vortical turbulence with more small scale structures. The non-synchronous vibration (NSV) of a high speed 1-1/2 stage axial compressor is investigated by using rigid blade and vibrating blade with fluid-structural interaction. An interpolation sliding boundary condition is used for the rotor-stator interaction. The URANS simulation with rigid blades shows that the leading edge(LE) circumferentially traveling vortices, roughly above 80% rotor span, travel backwards relative to the rotor rotation and cause an excitation with the frequency agreeing with the measured NSV frequency. The predicted excitation frequency of the traveling vortices in the rigid blade simulation is a non-engine order frequency of 2603 Hz, which agrees very well with the rig measured frequency of 2600 Hz. For the FSI simulation, the results show that there exist two dominant frequencies in the spectrum of the blade vibration. The lower dominant frequency is close to the first bending mode. The higher dominant frequency close to the first torsional mode agrees very well with the measured NSV frequency. To investigate whether the NSV is caused by flow excitation or by flow-structure locked-in phenomenon, the rotating speed is varied within a small RPM range, in which the rig test detected the NSV. The unsteady flows with rigid blades are simulated first at several RPMs. A dominant excitation NSV frequency caused by the circumferentially traveling tip vortices are captured. The simulation then switches to fluid structure interaction that allows the blades to vibrate freely. (Abstract shortened by ProQuest.).
Dakota Uncertainty Quantification Methods Applied to the CFD code Nek5000
DOE Office of Scientific and Technical Information (OSTI.GOV)
Delchini, Marc-Olivier; Popov, Emilian L.; Pointer, William David
This report presents the state of advancement of a Nuclear Energy Advanced Modeling and Simulation (NEAMS) project to characterize the uncertainty of the computational fluid dynamics (CFD) code Nek5000 using the Dakota package for flows encountered in the nuclear engineering industry. Nek5000 is a high-order spectral element CFD code developed at Argonne National Laboratory for high-resolution spectral-filtered large eddy simulations (LESs) and unsteady Reynolds-averaged Navier-Stokes (URANS) simulations.
URANOUS IODATE AS A CARRIER FOR PLUTONIUM
Miller, D.R.; Seaborg, G.T.; Thompson, S.G.
1959-12-15
A process is described for precipitating plutonium on a uranous iodate carrier from an aqueous acid solution conA plutonium solution more concentrated than the original solution can then be obtained by oxidizing the uranium to the hexavalent state and dissolving the precipitate, after separating the latter from the original solution, by means of warm nitric acid.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Clement, J.M.; Luina, A.P.; Jodra, L.G.
1957-01-01
In the recovery of uraniuma from leach solutions, the pilot plant of the J.E.N, does not clarify the solution and the sodium uranate carries with it a high proportion of impurities. Therefore, a study was made to determine the optimum conditions for the filtration of sodium uranate from nitric acid solution and to establish modifications in the dissolution processes at present in use for the concentrates. The effects of pressure, addition of CaSO/sub 4/ and Kieselgur, pH, and temperature were investigated The modifications made to the pilot plant as a result of these studies are briefly described. (J.S.R.)
Study of Near-Stall Flow Behavior in a Modern Transonic Fan with Composite Sweep
NASA Technical Reports Server (NTRS)
Hah, Chunill; Shin, Hyoun-Woo
2011-01-01
Detailed flow behavior in a modern transonic fan with a composite sweep is investigated in this paper. Both unsteady Reynolds-averaged Navier-Stokes (URANS) and Large Eddy Simulation (LES) methods are applied to investigate the flow field over a wide operating range. The calculated flow fields are compared with the data from an array of high-frequency response pressure transducers embedded in the fan casing. The current study shows that a relatively fine computational grid is required to resolve the flow field adequately and to calculate the pressure rise across the fan correctly. The calculated flow field shows detailed flow structure near the fan rotor tip region. Due to the introduction of composite sweep toward the rotor tip, the flow structure at the rotor tip is much more stable compared to that of the conventional blade design. The passage shock stays very close to the leading edge at the rotor tip even at the throttle limit. On the other hand, the passage shock becomes stronger and detaches earlier from the blade passage at the radius where the blade sweep is in the opposite direction. The interaction between the tip clearance vortex and the passage shock becomes intense as the fan operates toward the stall limit, and tip clearance vortex breakdown occurs at near-stall operation. URANS calculates the time-averaged flow field fairly well. Details of measured RMS static pressure are not calculated with sufficient accuracy with URANS. On the other hand, LES calculates details of the measured unsteady flow features in the current transonic fan with composite sweep fairly well and reveals the flow mechanism behind the measured unsteady flow field.
NASA Technical Reports Server (NTRS)
Booth, David; Flegel, Ashlie
2015-01-01
A computational assessment of the aerodynamic performance of the midspan section of a variable-speed power-turbine blade is described. The computation comprises a periodic single blade that represents the 2-D Midspan section VSPT blade that was tested in the NASA Glenn Research Center Transonic Turbine Blade Cascade Facility. A commercial, off-the-shelf (COTS) software package, Pointwise and CFD++, was used for the grid generation and RANS and URANS computations. The CFD code, which offers flexibility in terms of turbulence and transition modeling options, was assessed in terms of blade loading, loss, and turning against test data from the transonic tunnel. Simulations were assessed at positive and negative incidence angles that represent the turbine cruise and take-off design conditions. The results indicate that the secondary flow induced at the positive incidence cruise condition results in a highly loaded case and transitional flow on the blade is observed. The negative incidence take-off condition is unloaded and the flow is very two-dimensional. The computational results demonstrate the predictive capability of the gridding technique and COTS software for a linear transonic turbine blade cascade with large incidence angle variation.
NASA Technical Reports Server (NTRS)
Booth, David T.; Flegel, Ashlie B.
2015-01-01
A computational assessment of the aerodynamic performance of the midspan section of a variable-speed power-turbine blade is described. The computation comprises a periodic single blade that represents the 2-D Midspan section VSPT blade that was tested in the NASA Glenn Research Center Transonic Turbine Blade Cascade Facility. A commercial, off-the-shelf (COTS) software package, Pointwise and CFD++, was used for the grid generation and RANS and URANS computations. The CFD code, which offers flexibility in terms of turbulence and transition modeling options, was assessed in terms of blade loading, loss, and turning against test data from the transonic tunnel. Simulations were assessed at positive and negative incidence angles that represent the turbine cruise and take-off design conditions. The results indicate that the secondary flow induced at the positive incidence cruise condition results in a highly loaded case and transitional flow on the blade is observed. The negative incidence take-off condition is unloaded and the flow is very two-dimensional. The computational results demonstrate the predictive capability of the gridding technique and COTS software for a linear transonic turbine blade cascade with large incidence angle variation.
An assessment of CFD-based wall heat transfer models in piston engines
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sircar, Arpan; Paul, Chandan; Ferreyro-Fernandez, Sebastian
The lack of accurate submodels for in-cylinder heat transfer has been identified as a key shortcoming in developing truly predictive, physics-based computational fluid dynamics (CFD) models that can be used to develop combustion systems for advanced high-efficiency, low-emissions engines. Only recently have experimental methods become available that enable accurate near-wall measurements to enhance simulation capability via advancing models. Initial results show crank-angle dependent discrepancies with respect to previously used boundary-layer models of up to 100%. However, available experimental data is quite sparse (only few data points on engine walls) and limited (available measurements are those of heat flux only). Predictivemore » submodels are needed for medium-resolution ("engineering") LES and for unsteady Reynolds-averaged simulations (URANS). Recently, some research groups have performed DNS studies on engine-relevant conditions using simple geometries. These provide very useful data for benchmarking wall heat transfer models under such conditions. Further, a number of new and more sophisticated models have also become available in the literature which account for these engine-like conditions. Some of these have been incorporated while others of a more complex nature, which include solving additional partial differential equations (PDEs) within the thin boundary layer near the wall, are underway. These models will then be tested against the available DNS/experimental data in both SI (spark-ignition) and CI (compression-ignition) engines.« less
Slat Cove Noise Modeling: A Posteriori Analysis of Unsteady RANS Simulations
NASA Technical Reports Server (NTRS)
Choudhari, Meelan; Khorrami, Mehdi R.; Lockard, David P.; Atkins, Harold L.; Lilley, Geoffrey M.
2002-01-01
A companion paper by Khorrami et al demonstrates the feasibility of simulating the (nominally) self-sustained, large-scale unsteadiness within the leading-edge slat-cove region of multi-element airfoils using unsteady Reynolds-Averaged Navier-Stokes (URANS) equations, provided that the turbulence production term in the underlying two-equation turbulence model is switched off within the cove region. In conjunction with a FfowesWilliams-Hawkings solver, the URANS computations were shown to capture the dominant portion of the acoustic spectrum attributed to slat noise, as well as reproducing the increased intensity of slat cove motions (and, correspondingly, far-field noise as well) at the lower angles of attack. This paper examines that simulation database, augmented by additional simulations, with the objective of transitioning this apparent success to aeroacoustic predictions in an engineering context. As a first step towards this goal, the simulated flow and acoustic fields are compared with experiment and simplified analytical model. Rather intense near-field fluctuations in the simulated flow are found to be associated with unsteady separation along the slat bottom surface, relatively close to the slat cusp. Accuracy of the laminar-cove simulations in this near-wall region is raised to be an open issue. The adjoint Green's function approach is also explored in an attempt to identify the most efficient noise source locations.
NASA Astrophysics Data System (ADS)
Wilhelm, S.; Balarac, G.; Métais, O.; Ségoufin, C.
2016-11-01
Flow prediction in a bulb turbine draft tube is conducted for two operating points using Unsteady RANS (URANS) simulations and Large Eddy Simulations (LES). The inlet boundary condition of the draft tube calculation is a rotating two dimensional velocity profile exported from a RANS guide vane- runner calculation. Numerical results are compared with experimental data in order to validate the flow field and head losses prediction. Velocity profiles prediction is improved with LES in the center of the draft tube compared to URANS results. Moreover, more complex flow structures are obtained with LES. A local analysis of the predicted flow field using the energy balance in the draft tube is then introduced in order to detect the hydrodynamic instabilities responsible for head losses in the draft tube. In particular, the production of turbulent kinetic energy next to the draft tube wall and in the central vortex structure is found to be responsible for a large part of the mean kinetic energy dissipation in the draft tube and thus for head losses. This analysis is used in order to understand the differences in head losses for different operating points. The numerical methodology could then be improved thanks to an in-depth understanding of the local flow topology.
Reductive stripping process for the recovery of uranium from wet-process phosphoric acid
Hurst, Fred J.; Crouse, David J.
1984-01-01
A reductive stripping flow sheet for recovery of uranium from wet-process phosphoric acid is described. Uranium is stripped from a uranium-loaded organic phase by a redox reaction converting the uranyl to uranous ion. The uranous ion is reoxidized to the uranyl oxidation state to form an aqueous feed solution highly concentrated in uranium. Processing of this feed through a second solvent extraction cycle requires far less stripping reagent as compared to a flow sheet which does not include the reductive stripping reaction.
Trap level spectroscopic investigations of U: ZnAl2O4: Role of defect centres in the TSL process
NASA Astrophysics Data System (ADS)
Mohapatra, M.; Kumar, Mithlesh; Kadam, R. M.
2018-03-01
In order to evaluate the trap level spectroscopic properties of Uranium in ZnAl2O4 spinel host, undoped and Uranium doped ZnAl2O4 samples were synthesized. From photoluminescence (PL) data it was confirmed that uranium gets stabilized in the system as UO66- (octahedral uranate). Electron spin resonance (ESR) studies for the gamma irradiated sample suggested the formation of O2-, F+ and V centres. From the TSL (thermally stimulated luminescence) data, the trap parameters such as frequency factor and activation energy etc. were evaluated. From ESR-TSL correlation it was confirmed that the destruction of O2- ion coincides with TSL glow peak appeared at 332 K.
Validation of High-Fidelity CFD Simulations for Rocket Injector Design
NASA Technical Reports Server (NTRS)
Tucker, P. Kevin; Menon, Suresh; Merkle, Charles L.; Oefelein, Joseph C.; Yang, Vigor
2008-01-01
Computational fluid dynamics (CFD) has the potential to improve the historical rocket injector design process by evaluating the sensitivity of performance and injector-driven thermal environments to the details of the injector geometry and key operational parameters. Methodical verification and validation efforts on a range of coaxial injector elements have shown the current production CFD capability must be improved in order to quantitatively impact the injector design process. This paper documents the status of a focused effort to compare and understand the predictive capabilities and computational requirements of a range of CFD methodologies on a set of single element injector model problems. The steady Reynolds-Average Navier-Stokes (RANS), unsteady Reynolds-Average Navier-Stokes (URANS) and three different approaches using the Large Eddy Simulation (LES) technique were used to simulate the initial model problem, a single element coaxial injector using gaseous oxygen and gaseous hydrogen propellants. While one high-fidelity LES result matches the experimental combustion chamber wall heat flux very well, there is no monotonic convergence to the data with increasing computational tool fidelity. Systematic evaluation of key flow field regions such as the flame zone, the head end recirculation zone and the downstream near wall zone has shed significant, though as of yet incomplete, light on the complex, underlying causes for the performance level of each technique. 1 Aerospace Engineer and Combustion CFD Team Leader, MS ER42, NASA MSFC, AL 35812, Senior Member, AIAA. 2 Professor and Director, Computational Combustion Laboratory, School of Aerospace Engineering, 270 Ferst Dr., Atlanta, GA 30332, Associate Fellow, AIAA. 3 Reilly Professor of Engineering, School of Mechanical Engineering, 585 Purdue Mall, West Lafayette, IN 47907, Fellow, AIAA. 4 Principal Member of Technical Staff, Combustion Research Facility, 7011 East Avenue, MS9051, Livermore, CA 94550, Associate Fellow, AIAA. 5 J. L. and G. H. McCain Endowed Chair, Mechanical Engineering, 104 Research Building East, University Park, PA 16802, Fellow, AIAA. American Institute of Aeronautics and Astronautics 1
NASA Astrophysics Data System (ADS)
Imai, Masafumi; Lecacheux, Alain; Clarke, Tracy E.; Higgins, Charles A.; Panchenko, Mykhaylo; Dowell, Jayce; Imai, Kazumasa; Brazhenko, Anatolii I.; Frantsuzenko, Anatolii V.; Konovalenko, Alexandr A.
2016-08-01
On 2015 February 21, simultaneous observations of Jupiter's decametric radio emission between 10 and 33 MHz were carried out using three powerful low-frequency radio telescopes: the Long Wavelength Array Station One in the USA, the Nançay Decameter Array in France, and the URAN2 telescope in Ukraine. We measured the lag times of short-bursts (S-bursts) for 105 minutes of data over effective baselines of up to 8460 km by using cross-correlation analysis of the spectrograms from each instrument. Of particular interest is the measurement of the beaming thickness of S-bursts, testing if either flashlight- or beacon-like beaming is emanating from Jupiter. We find that the lag times for all pairs drift slightly as time elapses, in agreement with expectations from the flashlight-like beaming model. This leads to a new constraint of the minimum beaming thickness of 2.″66. Also, we find that most of the analyzed data abound with S-bursts, whose occurrence probability peaks at 17-18 MHz.
Computationally efficient simulation of unsteady aerodynamics using POD on the fly
NASA Astrophysics Data System (ADS)
Moreno-Ramos, Ruben; Vega, José M.; Varas, Fernando
2016-12-01
Modern industrial aircraft design requires a large amount of sufficiently accurate aerodynamic and aeroelastic simulations. Current computational fluid dynamics (CFD) solvers with aeroelastic capabilities, such as the NASA URANS unstructured solver FUN3D, require very large computational resources. Since a very large amount of simulation is necessary, the CFD cost is just unaffordable in an industrial production environment and must be significantly reduced. Thus, a more inexpensive, yet sufficiently precise solver is strongly needed. An opportunity to approach this goal could follow some recent results (Terragni and Vega 2014 SIAM J. Appl. Dyn. Syst. 13 330-65 Rapun et al 2015 Int. J. Numer. Meth. Eng. 104 844-68) on an adaptive reduced order model that combines ‘on the fly’ a standard numerical solver (to compute some representative snapshots), proper orthogonal decomposition (POD) (to extract modes from the snapshots), Galerkin projection (onto the set of POD modes), and several additional ingredients such as projecting the equations using a limited amount of points and fairly generic mode libraries. When applied to the complex Ginzburg-Landau equation, the method produces acceleration factors (comparing with standard numerical solvers) of the order of 20 and 300 in one and two space dimensions, respectively. Unfortunately, the extension of the method to unsteady, compressible flows around deformable geometries requires new approaches to deal with deformable meshes, high-Reynolds numbers, and compressibility. A first step in this direction is presented considering the unsteady compressible, two-dimensional flow around an oscillating airfoil using a CFD solver in a rigidly moving mesh. POD on the Fly gives results whose accuracy is comparable to that of the CFD solver used to compute the snapshots.
Numerical Study of Outlet Boundary Conditions for Unsteady Turbulent Internal Flows Using the NCC
NASA Technical Reports Server (NTRS)
Liu, Nan-Suey; Shih, Tsan-Hsing
2009-01-01
This paper presents the results of studies on the outlet boundary conditions for turbulent internal flow simulations. Several outlet boundary conditions have been investigated by applying the National Combustion Code (NCC) to the configuration of a LM6000 single injector flame tube. First of all, very large eddy simulations (VLES) have been performed using the partially resolved numerical simulation (PRNS) approach, in which both the nonlinear and linear dynamic subscale models were employed. Secondly, unsteady Reynolds averaged Navier- Stokes (URANS) simulations have also been performed for the same configuration to investigate the effects of different outlet boundary conditions in the context of URANS. Thirdly, the possible role of the initial condition is inspected by using three different initial flow fields for both the PRNS/VLES simulation and the URANS simulation. The same grid is used for all the simulations and the number of mesh element is about 0.5 million. The main purpose of this study is to examine the long-time behavior of the solution as determined by the imposed outlet boundary conditions. For a particular simulation to be considered as successful under the given initial and boundary conditions, the solution must be sustainable in a physically meaningful manner over a sufficiently long period of time. The commonly used outlet boundary condition for steady Reynolds averaged Navier-Stokes (RANS) simulation is a fixed pressure at the outlet with all the other dependent variables being extrapolated from the interior. The results of the present study suggest that this is also workable for the URANS simulation of the LM6000 injector flame tube. However, it does not work for the PRNS/VLES simulation due to the unphysical reflections of the pressure disturbances at the outlet boundary. This undesirable situation can be practically alleviated by applying a simple unsteady convection equation for the pressure disturbances at the outlet boundary. The numerical results presented in this paper suggest that this unsteady convection of pressure disturbances at the outlet works very well for all the unsteady simulations (both PRNS/VLES and URANS) of the LM6000 single injector flame tube.
CFD investigations of the aerodynamics of vehicle overtaking maneuvers
NASA Astrophysics Data System (ADS)
Uddin, Mesbah; Chellaram, Arune Dhiren; Robinson, Austin Clay
2017-06-01
When two vehicle bodies are involved in a passing maneuver, interesting and intricate aerodynamic interactions occur between them. Such passing maneuvers are very important in racing and have been an area of active interest in motorsports for quite some time. The existing literature shows only a few studies in this area, and, as such, very little is known about the complex aerodynamics of racing in proximity. This paper presents a Computational Fluid Dynamics (CFD) methodology capable of describing the transient effects that occur in this scenario. This is achieved by simulating two tandem simplified vehicle bodies, the Ahmed body, which were placed in a virtual wind tunnel. One Ahmed body was kept stationary, while the other was allowed to move in the longitudinal direction with a relatively low velocity. In order to achieve reliable CFD results when one of the solid objects is moving, a new meshing methodology, called the overset mesh model, was implemented in the CFD process. The simulations were run using Star CCM+, a commercial finite-volume CFD program, in which the unsteady Reynolds Averaged Navier-Stokes (URANS) solver was applied. The CFD results are compared against fully transient and quasi-steady-state experimental results where encouraging correlations between the CFD and experiments are observed. The veracity of the CFD work presented in this paper provides significant insight into the complex aerodynamics of a passing maneuver, and lays the foundation for further analysis in this area using more complex vehicle shapes and more complex tandem racing or passing maneuvers at a yaw angle.
Development of a High-Order Space-Time Matrix-Free Adjoint Solver
NASA Technical Reports Server (NTRS)
Ceze, Marco A.; Diosady, Laslo T.; Murman, Scott M.
2016-01-01
The growth in computational power and algorithm development in the past few decades has granted the science and engineering community the ability to simulate flows over complex geometries, thus making Computational Fluid Dynamics (CFD) tools indispensable in analysis and design. Currently, one of the pacing items limiting the utility of CFD for general problems is the prediction of unsteady turbulent ows.1{3 Reynolds-averaged Navier-Stokes (RANS) methods, which predict a time-invariant mean flowfield, struggle to provide consistent predictions when encountering even mild separation, such as the side-of-body separation at a wing-body junction. NASA's Transformative Tools and Technologies project is developing both numerical methods and physical modeling approaches to improve the prediction of separated flows. A major focus of this e ort is efficient methods for resolving the unsteady fluctuations occurring in these flows to provide valuable engineering data of the time-accurate flow field for buffet analysis, vortex shedding, etc. This approach encompasses unsteady RANS (URANS), large-eddy simulations (LES), and hybrid LES-RANS approaches such as Detached Eddy Simulations (DES). These unsteady approaches are inherently more expensive than traditional engineering RANS approaches, hence every e ort to mitigate this cost must be leveraged. Arguably, the most cost-effective approach to improve the efficiency of unsteady methods is the optimal placement of the spatial and temporal degrees of freedom (DOF) using solution-adaptive methods.
NASA Technical Reports Server (NTRS)
Vatsa, Veer N.; Turkel, Eli
2006-01-01
We apply an unsteady Reynolds-averaged Navier-Stokes (URANS) solver for the simulation of a synthetic jet created by a single diaphragm piezoelectric actuator in quiescent air. This configuration was designated as Case 1 for the CFDVAL2004 workshop held at Williamsburg, Virginia, in March 2004. Time-averaged and instantaneous data for this case were obtained at NASA Langley Research Center, using multiple measurement techniques. Computational results for this case using one-equation Spalart-Allmaras and two-equation Menter's turbulence models are presented along with the experimental data. The effect of grid refinement, preconditioning and time-step variation are also examined in this paper.
Numerical study on wake characteristics of high-speed trains
NASA Astrophysics Data System (ADS)
Yao, Shuan-Bao; Sun, Zhen-Xu; Guo, Di-Long; Chen, Da-Wei; Yang, Guo-Wei
2013-12-01
Intensive turbulence exists in the wakes of high speed trains, and the aerodynamic performance of the trailing car could deteriorate rapidly due to complicated features of the vortices in the wake zone. As a result, the safety and amenity of high speed trains would face a great challenge. This paper considers mainly the mechanism of vortex formation and evolution in the train flow field. A real CRH2 model is studied, with a leading car, a middle car and a trailing car included. Different running speeds and cross wind conditions are considered, and the approaches of unsteady Reynold-averaged Navier-Stokes (URANS) and detached eddy simulation (DES) are utilized, respectively. Results reveal that DES has better capability of capturing small eddies compared to URANS. However, for large eddies, the effects of two approaches are almost the same. In conditions without cross winds, two large vortex streets stretch from the train nose and interact strongly with each other in the wake zone. With the reinforcement of the ground, a complicated wake vortex system generates and becomes strengthened as the running speed increases. However, the locations of flow separations on the train surface and the separation mechanism keep unchanged. In conditions with cross winds, three large vortices develop along the leeward side of the train, among which the weakest one has no obvious influence on the wake flow while the other two stretch to the tail of the train and combine with the helical vortices in the train wake. Thus, optimization of the aerodynamic performance of the trailing car should be aiming at reducing the intensity of the wake vortex system.
Numerical simulation of vessel dynamics in manoeuvrability and seakeeping problems
NASA Astrophysics Data System (ADS)
Blishchik, A. E.; Taranov, A. E.
2018-05-01
This paper deals with some examples of numerical modelling for ship's dynamics problems and data comparison with corresponding experimental results. It was considered two kinds of simulation: self-propelled turning motion of crude carrier KVLCC2 and changing position of container carrier S 175 due to wave loadings. Mesh generation and calculation were made in STAR-CCM+ package. URANS equations were used as system of equations closed by k-w SST turbulence model. The vessel had several degrees of freedom, which depend on task. Based on the results of this research, the conclusion was made concerning the applicability of used numerical methods.
U(v) in metal uranates: A combined experimental and theoretical study of MgUO 4, CrUO 4, and FeUO 4
Guo, Xiaofeng; Tiferet, Eitan; Qi, Liang; ...
2016-01-01
Although pentavalent uranium can exist in aqueous solution, its presence in the solid state is uncommon. Metal monouranates, MgUO 4, CrUO 4 and FeUO 4 were synthesized for detailed structural and energetic investigations. Structural characteristics of these uranates used powder X-ray diffraction, synchrotron X-ray absorption spectroscopy, X-ray photoelectron spectroscopy, and 57Fe-Mossbauer spectroscopy. Enthalpies of formation were measured by high temperature oxide melt solution calorimetry. Density functional theory (DFT) calculations provided both structural and energetic information. The measured structural and thermodynamic properties show good consistency with those predicted from DFT. The presence of U 5+ has been solidly confirmed in CrUOmore » 4 and FeUO 4, which are thermodynamically stable compounds, and the origin and stability of U 5+ in the system was elaborated by DFT. Lastly, the structural and thermodynamic behaviour of U 5+ elucidated in this work is relevant to fundamental actinide redox chemistry and to applications in the nuclear industry and radioactive waste disposal.« less
Aerodynamic performance of a small vertical axis wind turbine using an overset grid method
NASA Astrophysics Data System (ADS)
Bangga, Galih; Solichin, Mochammad; Daman, Aida; Sa'adiyah, Devy; Dessoky, Amgad; Lutz, Thorsten
2017-08-01
The present paper aims to asses the aerodynamic performance of a small vertical axis wind turbine operating at a small wind speed of 5 m/s for 6 different tip speed ratios (λ=2-7). The turbine consists of two blades constructed using the NACA 0015 airfoil. The study is carried out using computational fluid dynamics (CFD) methods employing an overset grid approach. The (URANS) SST k - ω is used as the turbulence model. For the preliminary study, simulations of the NACA 0015 under static conditions for a broad range of angle of attack and a rotating two-bladed VAWT are carried out. The results are compared with available measurement data and a good agreement is obtained. The simulations demonstrate that the maximum power coefficient attained is 0.45 for λ=4. The aerodynamic loads hysteresis are presented showing that the dynamic stall effect decreases with λ.
Uranium luminescence in La2 Zr2 O7 : effect of concentration and annealing temperature.
Mohapatra, M; Rajeswari, B; Hon, N S; Kadam, R M
2016-12-01
The speciation of a particular element in any given matrix is a prerequisite to understanding its solubility and leaching properties. In this context, speciation of uranium in lanthanum zirconate pyrochlore (La 2 Zr 2 O 7 = LZO), prepared by a low-temperature combustion route, was carried out using a simple photoluminescence lifetime technique. The LZO matrix is considered to be a potential ceramic host for fixing nuclear and actinide waste products generated during the nuclear fuel cycle. Special emphasis has been given to understanding the dynamics of the uranium species in the host as a function of annealing temperature and concentration. It was found that, in the LZO host, uranium is stabilized as the commonly encountered uranyl species (UO 2 2+ ) up to a heat treatment of 500 °C at the surface. Above 500 °C, the uranyl ion is diffused into the matrix as the more symmetric octahedral uranate species (UO 6 6- ). The uranate ions thus formed replace the six-coordinated 'Zr' atoms at regular lattice positions. Further, it was observed that concentration quenching takes place beyond 5 mol% of uranium doping. The mechanism of the quenching was found to be a multipolar interaction. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.
An Approach to Improved Credibility of CFD Simulations for Rocket Injector Design
NASA Technical Reports Server (NTRS)
Tucker, Paul K.; Menon, Suresh; Merkle, Charles L.; Oefelein, Joseph C.; Yang, Vigor
2007-01-01
Computational fluid dynamics (CFD) has the potential to improve the historical rocket injector design process by simulating the sensitivity of performance and injector-driven thermal environments to. the details of the injector geometry and key operational parameters. Methodical verification and validation efforts on a range of coaxial injector elements have shown the current production CFD capability must be improved in order to quantitatively impact the injector design process.. This paper documents the status of an effort to understand and compare the predictive capabilities and resource requirements of a range of CFD methodologies on a set of model problem injectors. Preliminary results from a steady Reynolds-Average Navier-Stokes (RANS), an unsteady Reynolds-Average Navier Stokes (URANS) and three different Large Eddy Simulation (LES) techniques used to model a single element coaxial injector using gaseous oxygen and gaseous hydrogen propellants are presented. Initial observations are made comparing instantaneous results, corresponding time-averaged and steady-state solutions in the near -injector flow field. Significant differences in the flow fields exist, as expected, and are discussed. An important preliminary result is the identification of a fundamental mixing mechanism, accounted for by URANS and LES, but missing in the steady BANS methodology. Since propellant mixing is the core injector function, this mixing process may prove to have a profound effect on the ability to more correctly simulate injector performance and resulting thermal environments. Issues important to unifying the basis for future comparison such as solution initialization, required run time and grid resolution are addressed.
Study of the Unsteady Aerodynamics associated with a Cycloidally Rotating Blade
NASA Astrophysics Data System (ADS)
Agarwal, Nishant
Cycloidal Rotors have been studied for over 100 years, with a focus on applications for vertical axis wind turbines (VAWTs) for energy production and vertical-take-off-and-landing (VTOL) vehicles. Although, numerous experimental and analytical studies have demonstrated their potential competency compared to conventional horizontal-axis rotors, it is not until recently that the focus of these studies has shifted towards understanding the fundamental science behind how these complex systems function. The present study extends the existing fundamental knowledge about cycloidal rotors by particularly focusing on the unsteady aerodynamic phenomena associated with a single-fixed NACA 0012 blade cycloidal rotor as the system translates across an advance ratio (mu = Uinfinity/oR ) of 1. This phenomena was studied both experimentally, making use of particle image velocimetry (PIV) measurements on the system, and computationally, making use of both simple analytical tools and two-dimensional Unsteady Reynolds-Averaged Navier Stokes computational fluid dynamics (URANS-CFD) simulations. It is important to study the transition of the system through mu = 1 in order to better understand the incapability of VAWTs to self-start, and also the progression of VTOL vehicles into forward flight. When the advance ratio is less than one the blade cuts through its own wake. As it approaches one the local airspeed of the flow over the airfoil approaches zero during the retreating portion of the cycle. Finally, as the advance ratio increases beyond one the airfoil will experience reversed flow relative to its direction of rotation. The analysis of the PIV results show that the flow just downstream of the rotor is similar for cases at the same advance ratios, and that the wake structures do not depend upon the Reynolds number, within the range investigated. The phase-history velocity contour plots of the wake structure show a distinct cycloidal pattern for the advance ratio of mu = 1.25, a more stationary wake pattern for mu = 1, and a retarding wake pattern for mu = 0.75. CFD analysis using three different turbulence models showed that an asymmetric wake was generated behind the rotor with a more complex structure (both inside and outside the rotor diameter). This asymmetric wake generation is attributed to the difference in flow conditions at the advancing and retreating sides of the cycle. The complex structures account for the occurrence of dynamic stall, shedding of wake from the trailing edge, flow reversal on the airfoil in the cycle, and the wake-blade interaction. Also, it is observed that a region of high velocity is generated by the airfoil as it sweeps through the flow, which interacts with the airfoil at a later point in the cycle and affects the net force on the airfoil. It was seen that the blade-vortex interaction is not a characteristic property of the cycloidal rotor system. Rather, it depends on the advance ratio at which the system operates.
LES and URANS predictions of the hydrodynamic loads on a tension-leg platform
NASA Astrophysics Data System (ADS)
Abrishamchi, A.; Younis, B. A.
2012-01-01
This paper reports on the prediction of the unsteady hydrodynamic forces that act on a floating Tension-Leg Platform (TLP) due to the action of a steady current. The results were obtained by solving the three-dimensional, time-dependent form of the equations governing conservation of mass and momentum. Movement of the free surface was tracked using the volume of fluid algorithm. The effects of turbulence were accounted for using two very different approaches: Large-Eddy Simulations (LES) and Unsteady Reynolds-Averaged Navier-Stokes (URANS). The latter approach utilized a two-equation turbulence closure that has been extended to capture the occurrence and consequences of vortex shedding from bluff bodies. The primary objective of the work was to explore the merits and de-merits of each modeling approach when applied to a large-scale structure of the type frequently encountered in practice. The test case chosen for this purposes of this assessment was the case of a conventional TLP in steady current at Reynolds numbers (based on column diameter) of 7.5×106 and 7.5×107. These values are representative of those encountered in deep-sea operations. Experiments and field observations have indicated that the resulting flows exhibit a number of complicated features due to the interactions between the shed vortices and the various structural components of the TLP. Many but not all of these features were captured by the present computations. In addition to a critical assessment of the two modeling approaches, the paper reports on a number of practical experiences gained in the course of conducting this study, including an assessment of the importance of allowing for the movement of the free surface (as opposed to adopting the usual solid-lid approximation) and an illustration of the effects of the current's angle of incidence on the computed hydrodynamic loads.
NASA Astrophysics Data System (ADS)
Amiraux, Mathieu
Rotorcraft Blade-Vortex Interaction (BVI) remains one of the most challenging flow phenomenon to simulate numerically. Over the past decade, the HART-II rotor test and its extensive experimental dataset has been a major database for validation of CFD codes. Its strong BVI signature, with high levels of intrusive noise and vibrations, makes it a difficult test for computational methods. The main challenge is to accurately capture and preserve the vortices which interact with the rotor, while predicting correct blade deformations and loading. This doctoral dissertation presents the application of a coupled CFD/CSD methodology to the problem of helicopter BVI and compares three levels of fidelity for aerodynamic modeling: a hybrid lifting-line/free-wake (wake coupling) method, with modified compressible unsteady model; a hybrid URANS/free-wake method; and a URANS-based wake capturing method, using multiple overset meshes to capture the entire flow field. To further increase numerical correlation, three helicopter fuselage models are implemented in the framework. The first is a high resolution 3D GPU panel code; the second is an immersed boundary based method, with 3D elliptic grid adaption; the last one uses a body-fitted, curvilinear fuselage mesh. The main contribution of this work is the implementation and systematic comparison of multiple numerical methods to perform BVI modeling. The trade-offs between solution accuracy and computational cost are highlighted for the different approaches. Various improvements have been made to each code to enhance physical fidelity, while advanced technologies, such as GPU computing, have been employed to increase efficiency. The resulting numerical setup covers all aspects of the simulation creating a truly multi-fidelity and multi-physics framework. Overall, the wake capturing approach showed the best BVI phasing correlation and good blade deflection predictions, with slightly under-predicted aerodynamic loading magnitudes. However, it proved to be much more expensive than the other two methods. Wake coupling with RANS solver had very good loading magnitude predictions, and therefore good acoustic intensities, with acceptable computational cost. The lifting-line based technique often had over-predicted aerodynamic levels, due to the degree of empiricism of the model, but its very short run-times, thanks to GPU technology, makes it a very attractive approach.
Aerodynamic investigations of a disc-wing
NASA Astrophysics Data System (ADS)
Dumitrache, Alexandru; Frunzulica, Florin; Grigorescu, Sorin
2017-01-01
The purpose of this paper is to evaluate the aerodynamic characteristics of a wing-disc, for a civil application in the fire-fighting system. The aerodynamic analysis is performed using a CFD code, named ANSYS Fluent, in the flow speed range up to 25 m/s, at lower and higher angle of attack. The simulation is three-dimensional, using URANS completed by a SST turbulence model. The results are used to examine the flow around the disc with increasing angle of attack and the structure of the wake.
NASA Astrophysics Data System (ADS)
Berland, Matthew W.
As scientists use the tools of computational and complex systems theory to broaden science perspectives (e.g., Bar-Yam, 1997; Holland, 1995; Wolfram, 2002), so can middle-school students broaden their perspectives using appropriate tools. The goals of this dissertation project are to build, study, evaluate, and compare activities designed to foster both computational and complex systems fluencies through collaborative constructionist virtual and physical robotics. In these activities, each student builds an agent (e.g., a robot-bird) that must interact with fellow students' agents to generate a complex aggregate (e.g., a flock of robot-birds) in a participatory simulation environment (Wilensky & Stroup, 1999a). In a participatory simulation, students collaborate by acting in a common space, teaching each other, and discussing content with one another. As a result, the students improve both their computational fluency and their complex systems fluency, where fluency is defined as the ability to both consume and produce relevant content (DiSessa, 2000). To date, several systems have been designed to foster computational and complex systems fluencies through computer programming and collaborative play (e.g., Hancock, 2003; Wilensky & Stroup, 1999b); this study suggests that, by supporting the relevant fluencies through collaborative play, they become mutually reinforcing. In this work, I will present both the design of the VBOT virtual/physical constructionist robotics learning environment and a comparative study of student interaction with the virtual and physical environments across four middle-school classrooms, focusing on the contrast in systems perspectives differently afforded by the two environments. In particular, I found that while performance gains were similar overall, the physical environment supported agent perspectives on aggregate behavior, and the virtual environment supported aggregate perspectives on agent behavior. The primary research questions are: (1) What are the relative affordances of virtual and physical constructionist robotics systems towards computational and complex systems fluencies? (2) What can middle school students learn using computational/complex systems learning environments in a collaborative setting? (3) In what ways are these environments and activities effective in teaching students computational and complex systems fluencies?
NASA Astrophysics Data System (ADS)
Piro, M. H. A.; Banfield, J.; Clarno, K. T.; Simunovic, S.; Besmann, T. M.; Lewis, B. J.; Thompson, W. T.
2013-10-01
Predictive capabilities for simulating irradiated nuclear fuel behavior are enhanced in the current work by coupling thermochemistry, isotopic evolution and heat transfer. Thermodynamic models that are incorporated into this framework not only predict the departure from stoichiometry of UO2, but also consider dissolved fission and activation products in the fluorite oxide phase, noble metal inclusions, secondary oxides including uranates, zirconates, molybdates and the gas phase. Thermochemical computations utilize the spatial and temporal evolution of the fission and activation product inventory in the pellet, which is typically neglected in nuclear fuel performance simulations. Isotopic computations encompass the depletion, decay and transmutation of more than 2000 isotopes that are calculated at every point in space and time. These computations take into consideration neutron flux depression and the increased production of fissile plutonium near the fuel pellet periphery (i.e., the so-called “rim effect”). Thermochemical and isotopic predictions are in very good agreement with reported experimental measurements of highly irradiated UO2 fuel with an average burnup of 102 GW d t(U)-1. Simulation results demonstrate that predictions are considerably enhanced when coupling thermochemical and isotopic computations in comparison to empirical correlations. Notice: This manuscript has been authored by UT-Battelle, LLC, under Contract No. DE-AC05-00OR22725 with the U.S. Department of Energy. The United States Government retains and the publisher, by accepting the article for publication, acknowledges that the United States Government retains a non-exclusive, paid-up, irrevocable, world-wide license to publish or reproduce the published form of this manuscript, or allow others to do so, for United States Government purposes.
PREPARATION OF HIGH PURITY UF$sub 4$
Magner, J.E.; Long, R.S.; Ellis, D.A.; Grinstead, R.R.
1962-04-17
S>A process for preparing very highly pure uranous tetrafluoride from impure uranium laden solvent extraction strip solutions, ion exchange process and resin-inpulp process eluate solutions which are at least 8M in hydrochloric acid is described. The process first comprises treating any of the above-mentioned solutions with a reducing agent to reduce the uranium to the + 4 oxidation state, and then contacting the reduced solution with an extractant phase comprising about 10 to 70% of tri-butyl phosphate in an organic solvent-diluent selected from benzene, ethyl-benzene, chlorobenzene, xylene, kerosene, or the like. The uranium is extracted into the extractant phase and is subsequently precipitated by treating the extractant with an aqueous fluoride solution. The highly pure uranous tetrafluoride precipitate is separated from the phases and recovered for subsequent utilization. (AEC)
Felipe-Sotelo, M; Hinchliff, J; Field, L P; Milodowski, A E; Preedy, O; Read, D
2017-07-01
The solubility of uranium and thorium has been measured under the conditions anticipated in a cementitious, geological disposal facility for low and intermediate level radioactive waste. Similar solubilities were obtained for thorium in all media, comprising NaOH, Ca(OH) 2 and water equilibrated with a cement designed as repository backfill (NRVB, Nirex Reference Vault Backfill). In contrast, the solubility of U(VI) was one order of magnitude higher in NaOH than in the remaining solutions. The presence of cellulose degradation products (CDP) results in a comparable solubility increase for both elements. Extended X-ray Absorption Fine Structure (EXAFS) data suggest that the solubility-limiting phase for uranium corresponds to a becquerelite-type solid whereas thermodynamic modelling predicts a poorly crystalline, hydrated calcium uranate phase. The solubility-limiting phase for thorium was ThO 2 of intermediate crystallinity. No breakthrough of either uranium or thorium was observed in diffusion experiments involving NRVB after three years. Nevertheless, backscattering electron microscopy and microfocus X-ray fluorescence confirmed that uranium had penetrated about 40 μm into the cement, implying active diffusion governed by slow dissolution-precipitation kinetics. Precise identification of the uranium solid proved difficult, displaying characteristics of both calcium uranate and becquerelite. Copyright © 2017 Elsevier Ltd. All rights reserved.
OCCUPATIONAL DERMATITIS DUE TO SENSITIZATION TO SODIUM AND CALCIUM URANYL CARBONATES (in French)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Thiers, H.; Chanial, G.; Rivoire, J.
1961-01-01
BS>A report on a rare type of occupational dermatitis from sensitization to Na and Ca uranates observed in two workers in the Guegnon uranium-processing plandt is presented. In both patients there was no history of personal or family allergies. The dermatitis was especially evident on the fingers and presented characteristics not reported previously: the skin was dry and wrinkled with small periangualar fissures. Small nodules 2--3 mm, appearing as centers of hyperkeratosis encased in the skin, appeared on the sides and under surface of fingers. The nails were grooved and peeling at the edges. Also, at the base of themore » fingers and over the hand hyperkeritosic eczema of a common type was seen. These lessons were filled with pus. Skin tests confirmed their sensitization to 2% solutions of Na and Ca uranates; otherwise these patients appeared to be completely normal. A survey of working conditions showed that there was no possibility of their dermatitis resulting from radioactivity. That excessive radiation was not involved in the pathogenesis of these lesions was attested to by their normal blood pictures. The manifestations of the lesions indicated their definite allergic origin. Both patients recovered eventually and, after occupying other positions in the plant, have had no recrudescence of the lesions. (BBB)« less
Characterization of Unsteady Flow Structures Near Leading-Edge Slat. Part 1; PIV Measurements
NASA Technical Reports Server (NTRS)
Jenkins, Luther N.; Khorrami, Mehdi R.; Choudhari, Meelan
2004-01-01
A comprehensive computational and experimental study has been performed at the NASA Langley Research Center as part of the Quiet Aircraft Technology (QAT) Program to investigate the unsteady flow near a leading-edge slat of a two-dimensional, high-lift system. This paper focuses on the experimental effort conducted in the NASA Langley Basic Aerodynamics Research Tunnel (BART) where Particle Image Velocimetry (PIV) data was acquired in the slat cove and at the slat trailing edge of a three-element, high-lift model at 4, 6, and 8 degrees angle of attack and a freestream Mach Number of 0.17. Instantaneous velocities obtained from PIV images are used to obtain mean and fluctuating components of velocity and vorticity. The data show the recirculation in the cove, reattachment of the shear layer on the slat lower surface, and discrete vortical structures within the shear layer emanating from the slat cusp and slat trailing edge. Detailed measurements are used to examine the shear layer formation at the slat cusp, vortex shedding at the slat trailing edge, and convection of vortical structures through the slat gap. Selected results are discussed and compared with unsteady, Reynolds-Averaged Navier-Stokes (URANS) computations for the same configuration in a companion paper by Khorrami, Choudhari, and Jenkins (2004). The experimental dataset provides essential flow-field information for the validation of near-field inputs to noise prediction tools.
Computation of unsteady turbomachinery flows: Part 2—LES and hybrids
NASA Astrophysics Data System (ADS)
Tucker, P. G.
2011-10-01
The choice of turbulence model can have a strong impact on results for many turbomachinery zones. Palliative corrections to them and also transition modeling can have a further profound solution impact. The spectral gaps necessary for theoretically valid URANS solutions are also lacking in certain turbomachinery zones. Large Eddy Simulation (LES) alleviates the serious area of turbulence modeling uncertainty but with an extreme increase in computational cost. However, there seems a lack of validation data to explore in depth the performance of LES and thus strategies to refine it. LES best practices are needed. Although LES is, obviously, much less model dependent than RANS, grids currently used for more practical simulations are clearly insufficiently fine for the LES model and numerical schemes not to be playing an excessively strong role. Very few turbomachinery simulations make use of properly constructed, correlated turbulence inflow. Even if this is attempted, most measurement sets are incomplete and lack an adequate basis for modeling this inflow. Gas turbines are highly complex coupled systems and hence inflow and outflow boundary condition specification needs to go beyond just synthesizing turbulent structures and preventing their reflection. Despite the strong limitations of the dissipative Smagorinsky model, it still sees the most wide spread use, generally, in excessively dissipative flow solvers. Monotone Integrated LES (MILES) related approaches, hybrid LES-RANS and more advanced LES models seem to have an equal but subservient frequency of use in turbomachinery applications. Clearly the introduction of a RANS layer can have a substantial accuracy penalty. However, it does allow LES to be rationally used, albeit in a diluted sense for industrial applications. The Reynolds numbers found in turbomachinery are substantial. However, in certain areas evidence suggests they will not be enough to ensure a long inertial subrange and hence the use of standard LES modeling practices. Despite the excessively coarse grids used in much of the LES work reviewed, with essentially RANS based codes, meaningful results are often gained. This can perhaps be attributed to the choice of cases, these being ones for which RANS modeling gives extremely poor performance. It is a concern that for practical turbomachinery LES studies grid densities used tend to have an Reynolds number scaling to a strong negative power.
GESA--a two-dimensional processing system using knowledge base techniques.
Rowlands, D G; Flook, A; Payne, P I; van Hoff, A; Niblett, T; McKee, S
1988-12-01
The successful analysis of two-dimensional (2-D) polyacrylamide electrophoresis gels demands considerable experience and understanding of the protein system under investigation as well as knowledge of the separation technique itself. The present work concerns the development of a computer system for analysing 2-D electrophoretic separations which incorporates concepts derived from artificial intelligence research such that non-experts can use the technique as a diagnostic or identification tool. Automatic analysis of 2-D gel separations has proved to be extremely difficult using statistical methods. Non-reproducibility of gel separations is also difficult to overcome using automatic systems. However, the human eye is extremely good at recognising patterns in images, and human intervention in semi-automatic computer systems can reduce the computational complexities of fully automatic systems. Moreover, the expertise and understanding of an "expert" is invaluable in reducing system complexity if it can be encapsulated satisfactorily in an expert system. The combination of user-intervention in the computer system together with the encapsulation of expert knowledge characterises the present system. The domain within which the system has been developed is that of wheat grain storage proteins (gliadins) which exhibit polymorphism to such an extent that cultivars can be uniquely identified by their gliadin patterns. The system can be adapted to other domains where a range of polymorpic protein sub-units exist. In its generalised form, the system can also be used for comparing more complex 2-D gel electrophoretic separations.
Development of simulation computer complex specification
NASA Technical Reports Server (NTRS)
1973-01-01
The Training Simulation Computer Complex Study was one of three studies contracted in support of preparations for procurement of a shuttle mission simulator for shuttle crew training. The subject study was concerned with definition of the software loads to be imposed on the computer complex to be associated with the shuttle mission simulator and the development of procurement specifications based on the resulting computer requirements. These procurement specifications cover the computer hardware and system software as well as the data conversion equipment required to interface the computer to the simulator hardware. The development of the necessary hardware and software specifications required the execution of a number of related tasks which included, (1) simulation software sizing, (2) computer requirements definition, (3) data conversion equipment requirements definition, (4) system software requirements definition, (5) a simulation management plan, (6) a background survey, and (7) preparation of the specifications.
Tangential synthetic jets for separation control
NASA Astrophysics Data System (ADS)
Esmaeili Monir, H.; Tadjfar, M.; Bakhtian, A.
2014-02-01
A numerical study of separation control has been made to investigate aerodynamic characteristics of a NACA23012 airfoil with a tangential synthetic jet. Simulations are carried out at the chord Reynolds number of Re=2.19×106. The present approach relies on solving the Unsteady Reynolds-Averaged Navier-Stokes (URANS) equations. The turbulence model used in the present computation is the Spalart-Allmaras one-equation model. All computations are performed with a finite volume based code. Stall characteristics are significantly improved by controlling the formation of separation vortices in the flow. We placed the synthetic jet at the 12% chord, xj=0.12c, where we expected the separation to occur. Two distinct jet oscillating frequencies: Fj+=0.159 and Fj+=1 were considered. We studied the effect of blowing ratio, Vj/U∞, where it was varied from 0 to 5. The inclined angle of the synthetic jet was varied from αj=0° up to αj=83°. For the non-zero inclined angles, the local maximum in the aerodynamic performance, Cl/Cd, of 6.89 was found for the inclined angle of about 43°. In the present method, by means of creating a dent on the airfoil, linear momentum is transferred to the flow system in tangential direction to the airfoil surface. Thus the absolute maximum of 11.19 was found for the tangential synthetic jet at the inclined angle of the jet of 0°. The mechanisms involved for a tangential jet appear to behave linearly, as by multiplying the activation frequency of the jet by a factor produces the same multiplication factor in the resulting frequency in the flow. However, the mechanisms involved in the non-zero inclined angle cases behave nonlinearly when the activation frequency is multiplied.
1983-09-01
6ENFRAL. ELECTROMAGNETIC MODEL FOR THE ANALYSIS OF COMPLEX SYSTEMS **%(GEMA CS) Computer Code Documentation ii( Version 3 ). A the BDM Corporation Dr...ANALYSIS FnlTcnclRpr F COMPLEX SYSTEM (GmCS) February 81 - July 83- I TR CODE DOCUMENTATION (Version 3 ) 6.PROMN N.REPORT NUMBER 5. CONTRACT ORGAT97...the ti and t2 directions on the source patch. 3 . METHOD: The electric field at a segment observation point due to the source patch j is given by 1-- lnA
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sanchez, L.G.; Cellini, R.F.
1959-01-01
The thermal decomposition of some intermediate compounds in the metallurgy of uranium such as uranium peroxide, ammonium uranate, ammonium uranium pentafluoride, uranium tetrafluoride, and UO/sub 2/, were studied using Chevenard's thermobalance. Some data on the pyrolysis of synthetic mixtures of intermediate compounds which may appear during the industrial processing are given. Thermogravimetric methods of control are suggested for use in uranium metallurgy. (tr-auth)
NASA Astrophysics Data System (ADS)
Iqbal, S.; Benim, A. C.; Fischer, S.; Joos, F.; Kluβ, D.; Wiedermann, A.
2016-10-01
Turbulent reacting flows in a generic swirl gas turbine combustor model are investigated both numerically and experimentally. In the investigation, an emphasis is placed upon the external flue gas recirculation, which is a promising technology for increasing the efficiency of the carbon capture and storage process, which, however, can change the combustion behaviour significantly. A further emphasis is placed upon the investigation of alternative fuels such as biogas and syngas in comparison to the conventional natural gas. Flames are also investigated numerically using the open source CFD software OpenFOAM. In the numerical simulations, a laminar flamelet model based on mixture fraction and reaction progress variable is adopted. As turbulence model, the SST model is used within a URANS concept. Computational results are compared with the experimental data, where a fair agreement is observed.
Small Universal Bacteria and Plasmid Computing Systems.
Wang, Xun; Zheng, Pan; Ma, Tongmao; Song, Tao
2018-05-29
Bacterial computing is a known candidate in natural computing, the aim being to construct "bacterial computers" for solving complex problems. In this paper, a new kind of bacterial computing system, named the bacteria and plasmid computing system (BP system), is proposed. We investigate the computational power of BP systems with finite numbers of bacteria and plasmids. Specifically, it is obtained in a constructive way that a BP system with 2 bacteria and 34 plasmids is Turing universal. The results provide a theoretical cornerstone to construct powerful bacterial computers and demonstrate a concept of paradigms using a "reasonable" number of bacteria and plasmids for such devices.
Undecidability and Irreducibility Conditions for Open-Ended Evolution and Emergence.
Hernández-Orozco, Santiago; Hernández-Quiroz, Francisco; Zenil, Hector
2018-01-01
Is undecidability a requirement for open-ended evolution (OEE)? Using methods derived from algorithmic complexity theory, we propose robust computational definitions of open-ended evolution and the adaptability of computable dynamical systems. Within this framework, we show that decidability imposes absolute limits on the stable growth of complexity in computable dynamical systems. Conversely, systems that exhibit (strong) open-ended evolution must be undecidable, establishing undecidability as a requirement for such systems. Complexity is assessed in terms of three measures: sophistication, coarse sophistication, and busy beaver logical depth. These three complexity measures assign low complexity values to random (incompressible) objects. As time grows, the stated complexity measures allow for the existence of complex states during the evolution of a computable dynamical system. We show, however, that finding these states involves undecidable computations. We conjecture that for similar complexity measures that assign low complexity values, decidability imposes comparable limits on the stable growth of complexity, and that such behavior is necessary for nontrivial evolutionary systems. We show that the undecidability of adapted states imposes novel and unpredictable behavior on the individuals or populations being modeled. Such behavior is irreducible. Finally, we offer an example of a system, first proposed by Chaitin, that exhibits strong OEE.
NASA Astrophysics Data System (ADS)
Barkin, Yu. V.
2003-04-01
BELT-HIERARCHIC STRUCTURE OF THE RING, SATELLITE AND PLANET SYSTEMS: PREDICTION S/2001 U1 AND OTHERS OBJECTS IN SOLAR SYSTEM Yu.V.Barkin Sternberg Astronomical Institute, Moscow, Russia, barkin@sai.msu.ru Structure regularities of the planet and satellite systems have been studied. Statistic analysis of the distribution of the major semi-axes of the orbits of the planets, comets and centaurs of the Solar system, satellite and ring systems of Jupiter, Saturn, Neptune and Uran, exoplanet systems of the pulsars PSR 1257+12, PSR 1828-11 and of the main consequence star Ups And was fulfilled. The following empirical regularities were described [1]: 1) the bodies of systems are combined into hierarchic groups and main from them combine 5 companions; 2) differences of the major semi-axes of the neighboring orbits for bodies of every group are constant; 4) for main neighboring hierarchic group these distances are distinguished in 6 times increasing to external grope; 5) the filling of the gropes and some present changes in their structure are caused by the past catastrophes in corresponding systems. The special method of reconstruction of the catastrophes which had place in the life of the Solar system (SS) was developed. Suggested method has let us to explain uniformly observed values of the major semi-axes and average values of eccentricities of the planets. In particular the Pancul’s hypothesis about Jupiter formation from two giant protoplanets (Jupiter I and Jupiter II) was confirmed. The new empirical law of the filling of the orbits of the regular groups of the planets or satellites (or rings structures) of the hierarchic ordered systems of celestial bodies was established. It was shown that sum number of bodies is proportional to the value of catastrophic value of the eccentricities which are same for first, second ,.... and fifth orbits of all gropes. The theoretical numbers of bodies for pointed orbits practically coincide with their observed numbers in main gropes of the all considered systems of celestial bodies (in Solar system and also in exoplanets systems of the pulsars PSR 1257+12, PSR 1828-11 and Ups And). Established regularities of the orbit structures let us to predict some new objects in the Solar system and in exoplanet systems. Some from them have been predicted in last years. So the new satellite of Uran (S/2001 U 1) is characterized by major semi-axis in 8 570 000 km (Minor Planet Electronic Circular, Issued 2002 Sept. 30). This satellite was predicted earlier as satellite E1 (8 640 000 km) [1]. [1] Yu.V.Barkin (2001) Electronic journal «Studied in Russia», 161, pp.1821-1830. http: // zhurnal. ape. relarn.ru/articles/2001/161.pdf.
NASA Astrophysics Data System (ADS)
Kelly, Ryan T.
Aero-optical disturbances produced from turbulent compressible flow-fields can seriously degrade the performance of an optical signal. At compressible flight speeds these disturbances stem from the density variations present in turbulent boundary layers and free shear layers; however helicopters typically operate at incompressible speeds, which nearly eliminates the aberrating effect of these flows. For helicopter platforms the sources of aberration originate from the high subsonic flow-field near the rotor blade tips in the form of rotor-tip vortices and from the high temperatures of the engine effluence. During hover the shed rotor-tip vortices and engine effluence convect with the rotor wake encircling the airframe and subsequently a helicopter mounted optical system. The aero-optical effects of the wake beneath a hovering helicopter were analyzed using a combination of Unsteady RANS (URANS) and Large-Eddy Simulations (LES). The spatial and temporal characteristics of the numerical optical wavefronts were compared to full-scale aero-optic experimental measurements. The results indicate that the turbulence of the rotor-tip vortices contributes to the higher order aberrations measured experimentally and that the thermal exhaust plumes effectively limit the optical field-of-regard to forward- and side-looking beam directions. This information along with the computed optical aberrations of the wake can be used to guide the development of adaptive-optic systems or other beam-control approaches.
Automated validation of a computer operating system
NASA Technical Reports Server (NTRS)
Dervage, M. M.; Milberg, B. A.
1970-01-01
Programs apply selected input/output loads to complex computer operating system and measure performance of that system under such loads. Technique lends itself to checkout of computer software designed to monitor automated complex industrial systems.
Reliability of numerical wind tunnels for VAWT simulation
NASA Astrophysics Data System (ADS)
Raciti Castelli, M.; Masi, M.; Battisti, L.; Benini, E.; Brighenti, A.; Dossena, V.; Persico, G.
2016-09-01
Computational Fluid Dynamics (CFD) based on the Unsteady Reynolds Averaged Navier Stokes (URANS) equations have long been widely used to study vertical axis wind turbines (VAWTs). Following a comprehensive experimental survey on the wakes downwind of a troposkien-shaped rotor, a campaign of bi-dimensional simulations is presented here, with the aim of assessing its reliability in reproducing the main features of the flow, also identifying areas needing additional research. Starting from both a well consolidated turbulence model (k-ω SST) and an unstructured grid typology, the main simulation settings are here manipulated in a convenient form to tackle rotating grids reproducing a VAWT operating in an open jet wind tunnel. The dependence of the numerical predictions from the selected grid spacing is investigated, thus establishing the less refined grid size that is still capable of capturing some relevant flow features such as integral quantities (rotor torque) and local ones (wake velocities).
Simulation of Cold Flow in a Truncated Ideal Nozzle with Film Cooling
NASA Technical Reports Server (NTRS)
Braman, K. E.; Ruf, J. H.
2015-01-01
Flow transients during rocket start-up and shut-down can lead to significant side loads on rocket nozzles. The capability to estimate these side loads computationally can streamline the nozzle design process. Towards this goal, the flow in a truncated ideal contour (TIC) nozzle has been simulated using RANS and URANS for a range of nozzle pressure ratios (NPRs) aimed to match a series of cold flow experiments performed at the NASA MSFC Nozzle Test Facility. These simulations were performed with varying turbulence model choices and for four approximations of the supersonic film injection geometry, each of which was created with a different simplification of the test article geometry. The results show that although a reasonable match to experiment can be obtained with varying levels of geometric fidelity, the modeling choices made do not fully represent the physics of flow separation in a TIC nozzle with film cooling.
Energy conservation and analysis and evaluation. [specifically at Slidell Computer Complex
NASA Technical Reports Server (NTRS)
1976-01-01
The survey assembled and made recommendations directed at conserving utilities and reducing the use of energy at the Slidell Computer Complex. Specific items included were: (1) scheduling and controlling the use of gas and electricity, (2) building modifications to reduce energy, (3) replacement of old, inefficient equipment, (4) modifications to control systems, (5) evaluations of economizer cycles in HVAC systems, and (6) corrective settings for thermostats, ductstats, and other temperature and pressure control devices.
NASA Technical Reports Server (NTRS)
Mitchell, Christine M.
1993-01-01
This chapter examines a class of human-computer interaction applications, specifically the design of human-computer interaction for the operators of complex systems. Such systems include space systems (e.g., manned systems such as the Shuttle or space station, and unmanned systems such as NASA scientific satellites), aviation systems (e.g., the flight deck of 'glass cockpit' airplanes or air traffic control) and industrial systems (e.g., power plants, telephone networks, and sophisticated, e.g., 'lights out,' manufacturing facilities). The main body of human-computer interaction (HCI) research complements but does not directly address the primary issues involved in human-computer interaction design for operators of complex systems. Interfaces to complex systems are somewhat special. The 'user' in such systems - i.e., the human operator responsible for safe and effective system operation - is highly skilled, someone who in human-machine systems engineering is sometimes characterized as 'well trained, well motivated'. The 'job' or task context is paramount and, thus, human-computer interaction is subordinate to human job interaction. The design of human interaction with complex systems, i.e., the design of human job interaction, is sometimes called cognitive engineering.
You've Come a Long Way, Baby, but...
ERIC Educational Resources Information Center
Fayen, Emily Gallup
An online library system is an example of a complex computer system in that it supports a variety of users, both patrons and staff, and is made up of many intricate programs with complex relationships among them. Certain features are essential to a user friendly system: (1) users cannot get lost in the system; (2) users cannot enter illegal…
ERIC Educational Resources Information Center
Berland, Matthew; Wilensky, Uri
2015-01-01
Both complex systems methods (such as agent-based modeling) and computational methods (such as programming) provide powerful ways for students to understand new phenomena. To understand how to effectively teach complex systems and computational content to younger students, we conducted a study in four urban middle school classrooms comparing…
NASA Technical Reports Server (NTRS)
Wolpert, David H.; Koga, Dennis (Technical Monitor)
2000-01-01
In the first of this pair of papers, it was proven that there cannot be a physical computer to which one can properly pose any and all computational tasks concerning the physical universe. It was then further proven that no physical computer C can correctly carry out all computational tasks that can be posed to C. As a particular example, this result means that no physical computer that can, for any physical system external to that computer, take the specification of that external system's state as input and then correctly predict its future state before that future state actually occurs; one cannot build a physical computer that can be assured of correctly "processing information faster than the universe does". These results do not rely on systems that are infinite, and/or non-classical, and/or obey chaotic dynamics. They also hold even if one uses an infinitely fast, infinitely dense computer, with computational powers greater than that of a Turing Machine. This generality is a direct consequence of the fact that a novel definition of computation - "physical computation" - is needed to address the issues considered in these papers, which concern real physical computers. While this novel definition does not fit into the traditional Chomsky hierarchy, the mathematical structure and impossibility results associated with it have parallels in the mathematics of the Chomsky hierarchy. This second paper of the pair presents a preliminary exploration of some of this mathematical structure. Analogues of Chomskian results concerning universal Turing Machines and the Halting theorem are derived, as are results concerning the (im)possibility of certain kinds of error-correcting codes. In addition, an analogue of algorithmic information complexity, "prediction complexity", is elaborated. A task-independent bound is derived on how much the prediction complexity of a computational task can differ for two different reference universal physical computers used to solve that task, a bound similar to the "encoding" bound governing how much the algorithm information complexity of a Turing machine calculation can differ for two reference universal Turing machines. Finally, it is proven that either the Hamiltonian of our universe proscribes a certain type of computation, or prediction complexity is unique (unlike algorithmic information complexity), in that there is one and only version of it that can be applicable throughout our universe.
Large Eddy Simulation of Flow in Turbine Cascades Using LESTool and UNCLE Codes
NASA Technical Reports Server (NTRS)
Huang, P. G.
2004-01-01
During the period December 23,1997 and December August 31,2004, we accomplished the development of 2 CFD codes for DNS/LES/RANS simulation of turbine cascade flows, namely LESTool and UNCLE. LESTool is a structured code making use of 5th order upwind differencing scheme and UNCLE is a second-order-accuracy unstructured code. LESTool has both Dynamic SGS and Spalart's DES models and UNCLE makes use of URANS and DES models. The current report provides a description of methodologies used in the codes.
Large Eddy Simulation of Flow in Turbine Cascades Using LEST and UNCLE Codes
NASA Technical Reports Server (NTRS)
Ashpis, David (Technical Monitor); Huang, P. G.
2004-01-01
During the period December 23, 1997 and December August 31, 2004, we accomplished the development of 2 CFD codes for DNS/LES/RANS simulation of turbine cascade flows, namely LESTool and UNCLE. LESTool is a structured code making use of 5th order upwind differencing scheme and UNCLE is a second-order-accuracy unstructured code. LESTool has both Dynamic SGS and Sparlart's DES models and UNCLE makes use of URANS and DES models. The current report provides a description of methodologies used in the codes.
NASA Astrophysics Data System (ADS)
Puzyrkov, Dmitry; Polyakov, Sergey; Podryga, Viktoriia; Markizov, Sergey
2018-02-01
At the present stage of computer technology development it is possible to study the properties and processes in complex systems at molecular and even atomic levels, for example, by means of molecular dynamics methods. The most interesting are problems related with the study of complex processes under real physical conditions. Solving such problems requires the use of high performance computing systems of various types, for example, GRID systems and HPC clusters. Considering the time consuming computational tasks, the need arises of software for automatic and unified monitoring of such computations. A complex computational task can be performed over different HPC systems. It requires output data synchronization between the storage chosen by a scientist and the HPC system used for computations. The design of the computational domain is also quite a problem. It requires complex software tools and algorithms for proper atomistic data generation on HPC systems. The paper describes the prototype of a cloud service, intended for design of atomistic systems of large volume for further detailed molecular dynamic calculations and computational management for this calculations, and presents the part of its concept aimed at initial data generation on the HPC systems.
Quantum Gauss-Jordan Elimination and Simulation of Accounting Principles on Quantum Computers
NASA Astrophysics Data System (ADS)
Diep, Do Ngoc; Giang, Do Hoang; Van Minh, Nguyen
2017-06-01
The paper is devoted to a version of Quantum Gauss-Jordan Elimination and its applications. In the first part, we construct the Quantum Gauss-Jordan Elimination (QGJE) Algorithm and estimate the complexity of computation of Reduced Row Echelon Form (RREF) of N × N matrices. The main result asserts that QGJE has computation time is of order 2 N/2. The second part is devoted to a new idea of simulation of accounting by quantum computing. We first expose the actual accounting principles in a pure mathematics language. Then, we simulate the accounting principles on quantum computers. We show that, all accounting actions are exhousted by the described basic actions. The main problems of accounting are reduced to some system of linear equations in the economic model of Leontief. In this simulation, we use our constructed Quantum Gauss-Jordan Elimination to solve the problems and the complexity of quantum computing is a square root order faster than the complexity in classical computing.
NASA Technical Reports Server (NTRS)
Kavi, K. M.
1984-01-01
There have been a number of simulation packages developed for the purpose of designing, testing and validating computer systems, digital systems and software systems. Complex analytical tools based on Markov and semi-Markov processes have been designed to estimate the reliability and performance of simulated systems. Petri nets have received wide acceptance for modeling complex and highly parallel computers. In this research data flow models for computer systems are investigated. Data flow models can be used to simulate both software and hardware in a uniform manner. Data flow simulation techniques provide the computer systems designer with a CAD environment which enables highly parallel complex systems to be defined, evaluated at all levels and finally implemented in either hardware or software. Inherent in data flow concept is the hierarchical handling of complex systems. In this paper we will describe how data flow can be used to model computer system.
Study of fission-product segregation in used CANDU fuel by X-ray photoelectron spectroscopy (XPS) II
NASA Astrophysics Data System (ADS)
Hocking, William H.; Duclos, A. Michael; Johnson, Lawrence H.
1994-03-01
A thorough investigation of the grain-boundary chemistry of used CANDU fuel from one intact element has been conducted by X-ray photoelectron spectroscopy (XPS). Selected findings from more extensive XPS measurements on other used CANDU fuels exposed to storage conditions are included for comparison. Cesium, rubidium, tellurium and barium have been commonly observed, often reaching high degrees of surface enrichment, although their relative abundances can vary widely with a complex dependence on the fuel irradiation history. Lower concentrations of cadmium, molybdenum, strontium and iodine have also been occasionally detected. Except for iodine, chemical-shift data are indicative of oxidized species, possibly uranates. Segregation at monolayer-level coverages has been demonstrated by sequential XPS analysis and argon-ion sputtering. Calculations based on an idealized thin-film model are consistent with the depth profiles. The interpretation of these results is discussed in the context of previous studies, especially on LWR fuels.
The Transfer of Abstract Principles Governing Complex Adaptive Systems
ERIC Educational Resources Information Center
Goldstone, Robert L.; Sakamoto, Yasuaki
2003-01-01
Four experiments explored participants' understanding of the abstract principles governing computer simulations of complex adaptive systems. Experiments 1, 2, and 3 showed better transfer of abstract principles across simulations that were relatively dissimilar, and that this effect was due to participants who performed relatively poorly on the…
Large-scale structural analysis: The structural analyst, the CSM Testbed and the NAS System
NASA Technical Reports Server (NTRS)
Knight, Norman F., Jr.; Mccleary, Susan L.; Macy, Steven C.; Aminpour, Mohammad A.
1989-01-01
The Computational Structural Mechanics (CSM) activity is developing advanced structural analysis and computational methods that exploit high-performance computers. Methods are developed in the framework of the CSM testbed software system and applied to representative complex structural analysis problems from the aerospace industry. An overview of the CSM testbed methods development environment is presented and some numerical methods developed on a CRAY-2 are described. Selected application studies performed on the NAS CRAY-2 are also summarized.
1978-09-12
the population. Only a socialist, planned economy can cope with such problems. However, the in- creasing complexity of the tasks faced’ by...the development of systems allowing man-machine dialogue does not decrease, but rather increase the complexity of the systems involved, simply...shifting the complexity to another sphere, where it is invisible to the human utilizing the system. Figures 5; refer- ences 3: 2 Russian, 1 Western
Hively, Lee M [Philadelphia, TN
2011-07-12
The invention relates to a method and apparatus for simultaneously processing different sources of test data into informational data and then processing different categories of informational data into knowledge-based data. The knowledge-based data can then be communicated between nodes in a system of multiple computers according to rules for a type of complex, hierarchical computer system modeled on a human brain.
A Model-based Framework for Risk Assessment in Human-Computer Controlled Systems
NASA Technical Reports Server (NTRS)
Hatanaka, Iwao
2000-01-01
The rapid growth of computer technology and innovation has played a significant role in the rise of computer automation of human tasks in modem production systems across all industries. Although the rationale for automation has been to eliminate "human error" or to relieve humans from manual repetitive tasks, various computer-related hazards and accidents have emerged as a direct result of increased system complexity attributed to computer automation. The risk assessment techniques utilized for electromechanical systems are not suitable for today's software-intensive systems or complex human-computer controlled systems. This thesis will propose a new systemic model-based framework for analyzing risk in safety-critical systems where both computers and humans are controlling safety-critical functions. A new systems accident model will be developed based upon modem systems theory and human cognitive processes to better characterize system accidents, the role of human operators, and the influence of software in its direct control of significant system functions. Better risk assessments will then be achievable through the application of this new framework to complex human-computer controlled systems.
Safety Metrics for Human-Computer Controlled Systems
NASA Technical Reports Server (NTRS)
Leveson, Nancy G; Hatanaka, Iwao
2000-01-01
The rapid growth of computer technology and innovation has played a significant role in the rise of computer automation of human tasks in modem production systems across all industries. Although the rationale for automation has been to eliminate "human error" or to relieve humans from manual repetitive tasks, various computer-related hazards and accidents have emerged as a direct result of increased system complexity attributed to computer automation. The risk assessment techniques utilized for electromechanical systems are not suitable for today's software-intensive systems or complex human-computer controlled systems.This thesis will propose a new systemic model-based framework for analyzing risk in safety-critical systems where both computers and humans are controlling safety-critical functions. A new systems accident model will be developed based upon modem systems theory and human cognitive processes to better characterize system accidents, the role of human operators, and the influence of software in its direct control of significant system functions Better risk assessments will then be achievable through the application of this new framework to complex human-computer controlled systems.
Incorporating Flexibility in the Design of Repairable Systems - Design of Microgrids
2014-01-01
MICROGRIDS Vijitashwa Pandey1 Annette Skowronska1,2...optimization of complex systems such as a microgrid is however, computationally intensive. The problem is exacerbated if we must incorporate...flexibility in terms of allowing the microgrid architecture and its running protocol to change with time. To reduce the computational effort, this paper
Systems and methods for rapid processing and storage of data
Stalzer, Mark A.
2017-01-24
Systems and methods of building massively parallel computing systems using low power computing complexes in accordance with embodiments of the invention are disclosed. A massively parallel computing system in accordance with one embodiment of the invention includes at least one Solid State Blade configured to communicate via a high performance network fabric. In addition, each Solid State Blade includes a processor configured to communicate with a plurality of low power computing complexes interconnected by a router, and each low power computing complex includes at least one general processing core, an accelerator, an I/O interface, and cache memory and is configured to communicate with non-volatile solid state memory.
Atomic switch networks-nanoarchitectonic design of a complex system for natural computing.
Demis, E C; Aguilera, R; Sillin, H O; Scharnhorst, K; Sandouk, E J; Aono, M; Stieg, A Z; Gimzewski, J K
2015-05-22
Self-organized complex systems are ubiquitous in nature, and the structural complexity of these natural systems can be used as a model to design new classes of functional nanotechnology based on highly interconnected networks of interacting units. Conventional fabrication methods for electronic computing devices are subject to known scaling limits, confining the diversity of possible architectures. This work explores methods of fabricating a self-organized complex device known as an atomic switch network and discusses its potential utility in computing. Through a merger of top-down and bottom-up techniques guided by mathematical and nanoarchitectonic design principles, we have produced functional devices comprising nanoscale elements whose intrinsic nonlinear dynamics and memorization capabilities produce robust patterns of distributed activity and a capacity for nonlinear transformation of input signals when configured in the appropriate network architecture. Their operational characteristics represent a unique potential for hardware implementation of natural computation, specifically in the area of reservoir computing-a burgeoning field that investigates the computational aptitude of complex biologically inspired systems.
Simulation of Supersonic Base Flows: Numerical Investigations Using DNS, LES, and URANS
2006-10-01
global instabilities were found for a two-dimensional bluff body with a blunt base by Hannemann & Oertel (1989). Oertel (1990) found that the... Hannemann , K. & Oertel, H. 1989 Numerical simulation of the absolutely and convectively unstable wake. J. Fluid Mech. 199, 55–88. Harris, P. J. 1997
Karavitis, G.A.
1984-01-01
The SIMSYS2D two-dimensional water-quality simulation system is a large-scale digital modeling software system used to simulate flow and transport of solutes in freshwater and estuarine environments. Due to the size, processing requirements, and complexity of the system, there is a need to easily move the system and its associated files between computer sites when required. A series of job control language (JCL) procedures was written to allow transferability between IBM and IBM-compatible computers. (USGS)
NASA Technical Reports Server (NTRS)
Townsend, James C.; Weston, Robert P.; Eidson, Thomas M.
1993-01-01
The Framework for Interdisciplinary Design Optimization (FIDO) is a general programming environment for automating the distribution of complex computing tasks over a networked system of heterogeneous computers. For example, instead of manually passing a complex design problem between its diverse specialty disciplines, the FIDO system provides for automatic interactions between the discipline tasks and facilitates their communications. The FIDO system networks all the computers involved into a distributed heterogeneous computing system, so they have access to centralized data and can work on their parts of the total computation simultaneously in parallel whenever possible. Thus, each computational task can be done by the most appropriate computer. Results can be viewed as they are produced and variables changed manually for steering the process. The software is modular in order to ease migration to new problems: different codes can be substituted for each of the current code modules with little or no effect on the others. The potential for commercial use of FIDO rests in the capability it provides for automatically coordinating diverse computations on a networked system of workstations and computers. For example, FIDO could provide the coordination required for the design of vehicles or electronics or for modeling complex systems.
Prediction of helicopter rotor noise in hover
NASA Astrophysics Data System (ADS)
Kusyumov, A. N.; Mikhailov, S. A.; Garipova, L. I.; Batrakov, A. S.; Barakos, G.
2015-05-01
Two mathematical models are used in this work to estimate the acoustics of a hovering main rotor. The first model is based on the Ffowcs Williams-Howkings equations using the formulation of Farassat. An analytical approach is followed for this model, to determine the thickness and load noise contributions of the rotor blade in hover. The second approach allows using URANS and RANS CFD solutions and based on numerical solution of the Ffowcs Williams-Howkings equations. The employed test cases correspond to a model rotor available at the KNRTUKAI aerodynamics laboratory. The laboratory is equipped with a system of acoustic measurements, and comparisons between predictions and measurements are to be attempted as part of this work.
Grid Computing in K-12 Schools. Soapbox Digest. Volume 3, Number 2, Fall 2004
ERIC Educational Resources Information Center
AEL, 2004
2004-01-01
Grid computing allows large groups of computers (either in a lab, or remote and connected only by the Internet) to extend extra processing power to each individual computer to work on components of a complex request. Grid middleware, recognizing priorities set by systems administrators, allows the grid to identify and use this power without…
NASA Technical Reports Server (NTRS)
Mckay, Charles W.; Feagin, Terry; Bishop, Peter C.; Hallum, Cecil R.; Freedman, Glenn B.
1987-01-01
The principle focus of one of the RICIS (Research Institute for Computing and Information Systems) components is computer systems and software engineering in-the-large of the lifecycle of large, complex, distributed systems which: (1) evolve incrementally over a long time; (2) contain non-stop components; and (3) must simultaneously satisfy a prioritized balance of mission and safety critical requirements at run time. This focus is extremely important because of the contribution of the scaling direction problem to the current software crisis. The Computer Systems and Software Engineering (CSSE) component addresses the lifestyle issues of three environments: host, integration, and target.
Engineering and Design: Control Stations and Control Systems for Navigation Locks and Dams
1997-05-30
of human intelli- hypothetical lock and dam configurations. Finally, b. Terminology. (1) PLC system. The computer- based systems utilize special...electrical industry for industrial use. There- fore, for purposes of this document, a computer- based system is referred to as a PLC system. (2) Relay- based ...be custom made, because most of today’s control systems of any complexity are PLC - based , the standard size of a given motor starter cubicle is not
1988-05-01
ifforiable manpower investement. On the basis of our current experience it seems that the basic design principles are valid. The system developed will... system is operational on various computer networks, and in both industrial and in research environments. The design pri,lciples for the construction of...to a useful numerical simulation and design system for very complex configurations and flows. 7. REFERENCES 1. Bartlett G. W. , "An experimental
The FuturICT education accelerator
NASA Astrophysics Data System (ADS)
Johnson, J.; Buckingham Shum, S.; Willis, A.; Bishop, S.; Zamenopoulos, T.; Swithenby, S.; MacKay, R.; Merali, Y.; Lorincz, A.; Costea, C.; Bourgine, P.; Louçã, J.; Kapenieks, A.; Kelley, P.; Caird, S.; Bromley, J.; Deakin Crick, R.; Goldspink, C.; Collet, P.; Carbone, A.; Helbing, D.
2012-11-01
Education is a major force for economic and social wellbeing. Despite high aspirations, education at all levels can be expensive and ineffective. Three Grand Challenges are identified: (1) enable people to learn orders of magnitude more effectively, (2) enable people to learn at orders of magnitude less cost, and (3) demonstrate success by exemplary interdisciplinary education in complex systems science. A ten year `man-on-the-moon' project is proposed in which FuturICT's unique combination of Complexity, Social and Computing Sciences could provide an urgently needed transdisciplinary language for making sense of educational systems. In close dialogue with educational theory and practice, and grounded in the emerging data science and learning analytics paradigms, this will translate into practical tools (both analytical and computational) for researchers, practitioners and leaders; generative principles for resilient educational ecosystems; and innovation for radically scalable, yet personalised, learner engagement and assessment. The proposed Education Accelerator will serve as a `wind tunnel' for testing these ideas in the context of real educational programmes, with an international virtual campus delivering complex systems education exploiting the new understanding of complex, social, computationally enhanced organisational structure developed within FuturICT.
NASA Technical Reports Server (NTRS)
1990-01-01
NASA's Space Station Freedom Program (SSFP) planning efforts have identified a need for a payload training simulator system to serve as both a training facility and as a demonstrator to validate operational concepts. The envisioned MSFC Payload Training Complex (PTC) required to meet this need will train the Space Station payload scientists, station scientists, and ground controllers to operate the wide variety of experiments that will be onboard the Space Station Freedom. The Simulation Computer System (SCS) is the computer hardware, software, and workstations that will support the Payload Training Complex at MSFC. The purpose of this SCS Study is to investigate issues related to the SCS, alternative requirements, simulator approaches, and state-of-the-art technologies to develop candidate concepts and designs.
Making classical ground-state spin computing fault-tolerant.
Crosson, I J; Bacon, D; Brown, K R
2010-09-01
We examine a model of classical deterministic computing in which the ground state of the classical system is a spatial history of the computation. This model is relevant to quantum dot cellular automata as well as to recent universal adiabatic quantum computing constructions. In its most primitive form, systems constructed in this model cannot compute in an error-free manner when working at nonzero temperature. However, by exploiting a mapping between the partition function for this model and probabilistic classical circuits we are able to show that it is possible to make this model effectively error-free. We achieve this by using techniques in fault-tolerant classical computing and the result is that the system can compute effectively error-free if the temperature is below a critical temperature. We further link this model to computational complexity and show that a certain problem concerning finite temperature classical spin systems is complete for the complexity class Merlin-Arthur. This provides an interesting connection between the physical behavior of certain many-body spin systems and computational complexity.
1977-01-26
Sisteme Matematicheskogo Obespecheniya YeS EVM [ Applied Programs in the Software System for the Unified System of Computers], by A. Ye. Fateyev, A. I...computerized systems are most effective in large production complexes , in which the level of utilization of computers can be as high as 500,000...performance of these tasks could be furthered by the complex introduction of electronic computers in automated control systems. The creation of ASU
Atomic switch networks—nanoarchitectonic design of a complex system for natural computing
NASA Astrophysics Data System (ADS)
Demis, E. C.; Aguilera, R.; Sillin, H. O.; Scharnhorst, K.; Sandouk, E. J.; Aono, M.; Stieg, A. Z.; Gimzewski, J. K.
2015-05-01
Self-organized complex systems are ubiquitous in nature, and the structural complexity of these natural systems can be used as a model to design new classes of functional nanotechnology based on highly interconnected networks of interacting units. Conventional fabrication methods for electronic computing devices are subject to known scaling limits, confining the diversity of possible architectures. This work explores methods of fabricating a self-organized complex device known as an atomic switch network and discusses its potential utility in computing. Through a merger of top-down and bottom-up techniques guided by mathematical and nanoarchitectonic design principles, we have produced functional devices comprising nanoscale elements whose intrinsic nonlinear dynamics and memorization capabilities produce robust patterns of distributed activity and a capacity for nonlinear transformation of input signals when configured in the appropriate network architecture. Their operational characteristics represent a unique potential for hardware implementation of natural computation, specifically in the area of reservoir computing—a burgeoning field that investigates the computational aptitude of complex biologically inspired systems.
NASA Astrophysics Data System (ADS)
Abu-Dalo, M. A.; Al-Rawashdeh, N. A. F.; Al-Mheidat, I. R.; Nassory, N. S.
2015-10-01
In the present study uranyl selective electrodes in polyvinyl chloride (PVC) matrix membrane were prepared based on a complex of uranyl ion (UO2) with carboxybenzotriazole (CBT) as ligand. The effect of the nature of plasticizer in PVC matrix were evaluated using three different plasticizers, these are dibutyl phthalate (DBP), dioctyl phthalate (DOP) and bis(2-ethylhexyl) sebacate (BHS). The results of this study indicated that the best plasticizer could be used is the DBP, which may be attributed to its lowest viscosity value compared to DOP and BHS. The electrodes with DBP as plasticizer exhibits a Nernstian response with a slope of 28.0 mV/ decade, over a wide range of concentration from 3.0×10-5-6.0×10-2 M and a detection limit of 4.0×10-6 M. It can be used in the pH range of 4.0-10.0 with a response time of less than 10 s for DBP and 25 s for both DOP and BHS. The effects of ions interferences on the electrode response were evaluated. The di- and tri-valent cations were found to interfere less than univalent cations, which was attributed to the high diffusion and the exchange rate between the univalent ions and the uranyl ion solution. The electrodes were characterized by Fourier Transform Infrared Spectroscopy (FTIR) and Scanning Electron microscopy (SEM). The results of the standard addition method were satisfactory with errors less than 7%. The developed electrode was found to be fast, sensitive and reliable indicated its potential use in measuring the uranly ion concentration in the field.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nandi, Taraj; Brasseur, James; Vijayakumar, Ganesh
2016-01-04
This study is aimed at gaining insight into the nonsteady transitional boundary layer dynamics of wind turbine blades and the predictive capabilities of URANS based transition and turbulence models for similar physics through the analysis of a controlled flow with similar nonsteady parameters.
Computing Mass Properties From AutoCAD
NASA Technical Reports Server (NTRS)
Jones, A.
1990-01-01
Mass properties of structures computed from data in drawings. AutoCAD to Mass Properties (ACTOMP) computer program developed to facilitate quick calculations of mass properties of structures containing many simple elements in such complex configurations as trusses or sheet-metal containers. Mathematically modeled in AutoCAD or compatible computer-aided design (CAD) system in minutes by use of three-dimensional elements. Written in Microsoft Quick-Basic (Version 2.0).
Decameter Type IV Burst Associated with a Behind-the-limb CME Observed on 7 November 2013
NASA Astrophysics Data System (ADS)
Melnik, V. N.; Brazhenko, A. I.; Konovalenko, A. A.; Dorovskyy, V. V.; Rucker, H. O.; Panchenko, M.; Frantsuzenko, A. V.; Shevchuk, M. V.
2018-03-01
We report on the results of observations of a type IV burst made by the Ukrainian Radio interferometer of the Academy of Sciences (URAN-2) in the frequency range 22 - 33 MHz. The burst is associated with a coronal mass ejection (CME) initiated by a behind-the-limb active region (N05E151) and was also observed by the Nançay Decameter Array (NDA) radio telescope in the frequency band 30 - 60 MHz. The purpose of the article is the determination of the source of this type IV burst. After analysis of the observational data obtained with the URAN-2, the NDA, the Solar-Terrestrial Relations Observatory (STEREO) A and B spacecraft, and the Solar and Heliospheric Observatory (SOHO) spacecraft, we come to the conclusion that the source of the burst is the core of a behind-the-limb CME. We conclude that the radio emission can escape the center of the CME core at a frequency of 60 MHz and originates from the periphery of the core at a frequency of 30 MHz that is due to occultation by the solar corona at the corresponding frequencies. We find plasma densities in these regions assuming the plasma mechanism of radio emission. We show that the frequency drift of the start of the type IV burst is governed by an expansion of the CME core. The type III bursts that were observed against this type IV burst are shown to be generated by fast electrons propagating through the CME core plasma. A type II burst was registered at frequencies of 44 - 64 MHz and 3 - 16 MHz and was radiated by a shock with velocities of about 1000 km s^{-1} and 800 km s^{-1}, respectively.
Accomplishment Summary 1968-1969. Biological Computer Laboratory.
ERIC Educational Resources Information Center
Von Foerster, Heinz; And Others
This report summarizes theoretical, applied, and experimental studies in the areas of computational principles in complex intelligent systems, cybernetics, multivalued logic, and the mechanization of cognitive processes. This work is summarized under the following topic headings: properties of complex dynamic systems; computers and the language…
Computer-Assisted Monitoring Of A Complex System
NASA Technical Reports Server (NTRS)
Beil, Bob J.; Mickelson, Eric M.; Sterritt, John M.; Costantino, Rob W.; Houvener, Bob C.; Super, Mike A.
1995-01-01
Propulsion System Advisor (PSA) computer-based system assists engineers and technicians in analyzing masses of sensory data indicative of operating conditions of space shuttle propulsion system during pre-launch and launch activities. Designed solely for monitoring; does not perform any control functions. Although PSA developed for highly specialized application, serves as prototype of noncontrolling, computer-based subsystems for monitoring other complex systems like electric-power-distribution networks and factories.
Robot, computer problem solving system
NASA Technical Reports Server (NTRS)
Becker, J. D.; Merriam, E. W.
1973-01-01
The TENEX computer system, the ARPA network, and computer language design technology was applied to support the complex system programs. By combining the pragmatic and theoretical aspects of robot development, an approach is created which is grounded in realism, but which also has at its disposal the power that comes from looking at complex problems from an abstract analytical point of view.
40-Gb/s PAM4 with low-complexity equalizers for next-generation PON systems
NASA Astrophysics Data System (ADS)
Tang, Xizi; Zhou, Ji; Guo, Mengqi; Qi, Jia; Hu, Fan; Qiao, Yaojun; Lu, Yueming
2018-01-01
In this paper, we demonstrate 40-Gb/s four-level pulse amplitude modulation (PAM4) transmission with 10 GHz devices and low-complexity equalizers for next-generation passive optical network (PON) systems. Simple feed-forward equalizer (FFE) and decision feedback equalizer (DFE) enable 20 km fiber transmission while high-complexity Volterra algorithm in combination with FFE and DFE can extend the transmission distance to 40 km. A simplified Volterra algorithm is proposed for reducing computational complexity. Simulation results show that the simplified Volterra algorithm reduces up to ∼75% computational complexity at a relatively low cost of only 0.4 dB power budget. At a forward error correction (FEC) threshold of 10-3 , we achieve 31.2 dB and 30.8 dB power budget over 40 km fiber transmission using traditional FFE-DFE-Volterra and our simplified FFE-DFE-Volterra, respectively.
Aeropropulsion 1987. Session 2: Aeropropulsion Structures Research
NASA Technical Reports Server (NTRS)
1987-01-01
Aeropropulsion systems present unique problems to the structural engineer. The extremes in operating temperatures, rotational effects, and behaviors of advanced material systems combine into complexities that require advances in many scientific disciplines involved in structural analysis and design procedures. This session provides an overview of the complexities of aeropropulsion structures and the theoretical, computational, and experimental research conducted to achieve the needed advances.
High performance computing in biology: multimillion atom simulations of nanoscale systems
Sanbonmatsu, K. Y.; Tung, C.-S.
2007-01-01
Computational methods have been used in biology for sequence analysis (bioinformatics), all-atom simulation (molecular dynamics and quantum calculations), and more recently for modeling biological networks (systems biology). Of these three techniques, all-atom simulation is currently the most computationally demanding, in terms of compute load, communication speed, and memory load. Breakthroughs in electrostatic force calculation and dynamic load balancing have enabled molecular dynamics simulations of large biomolecular complexes. Here, we report simulation results for the ribosome, using approximately 2.64 million atoms, the largest all-atom biomolecular simulation published to date. Several other nanoscale systems with different numbers of atoms were studied to measure the performance of the NAMD molecular dynamics simulation program on the Los Alamos National Laboratory Q Machine. We demonstrate that multimillion atom systems represent a 'sweet spot' for the NAMD code on large supercomputers. NAMD displays an unprecedented 85% parallel scaling efficiency for the ribosome system on 1024 CPUs. We also review recent targeted molecular dynamics simulations of the ribosome that prove useful for studying conformational changes of this large biomolecular complex in atomic detail. PMID:17187988
NASA Astrophysics Data System (ADS)
Myre, Joseph M.
Heterogeneous computing systems have recently come to the forefront of the High-Performance Computing (HPC) community's interest. HPC computer systems that incorporate special purpose accelerators, such as Graphics Processing Units (GPUs), are said to be heterogeneous. Large scale heterogeneous computing systems have consistently ranked highly on the Top500 list since the beginning of the heterogeneous computing trend. By using heterogeneous computing systems that consist of both general purpose processors and special- purpose accelerators, the speed and problem size of many simulations could be dramatically increased. Ultimately this results in enhanced simulation capabilities that allows, in some cases for the first time, the execution of parameter space and uncertainty analyses, model optimizations, and other inverse modeling techniques that are critical for scientific discovery and engineering analysis. However, simplifying the usage and optimization of codes for heterogeneous computing systems remains a challenge. This is particularly true for scientists and engineers for whom understanding HPC architectures and undertaking performance analysis may not be primary research objectives. To enable scientists and engineers to remain focused on their primary research objectives, a modular environment for geophysical inversion and run-time autotuning on heterogeneous computing systems is presented. This environment is composed of three major components: 1) CUSH---a framework for reducing the complexity of programming heterogeneous computer systems, 2) geophysical inversion routines which can be used to characterize physical systems, and 3) run-time autotuning routines designed to determine configurations of heterogeneous computing systems in an attempt to maximize the performance of scientific and engineering codes. Using three case studies, a lattice-Boltzmann method, a non-negative least squares inversion, and a finite-difference fluid flow method, it is shown that this environment provides scientists and engineers with means to reduce the programmatic complexity of their applications, to perform geophysical inversions for characterizing physical systems, and to determine high-performing run-time configurations of heterogeneous computing systems using a run-time autotuner.
NASA Technical Reports Server (NTRS)
Gordon, S.; Mcbride, B. J.
1976-01-01
A detailed description of the equations and computer program for computations involving chemical equilibria in complex systems is given. A free-energy minimization technique is used. The program permits calculations such as (1) chemical equilibrium for assigned thermodynamic states (T,P), (H,P), (S,P), (T,V), (U,V), or (S,V), (2) theoretical rocket performance for both equilibrium and frozen compositions during expansion, (3) incident and reflected shock properties, and (4) Chapman-Jouguet detonation properties. The program considers condensed species as well as gaseous species.
JPRS Report, Science & Technology, USSR: Computers, Control Systems and Machines
1989-03-14
optimizatsii slozhnykh sistem (Coding Theory and Complex System Optimization ). Alma-Ata, Nauka Press, 1977, pp. 8-16. 11. Author’s certificate number...Interpreter Specifics [0. I. Amvrosova] ............................................. 141 Creation of Modern Computer Systems for Complex Ecological...processor can be designed to decrease degradation upon failure and assure more reliable processor operation, without requiring more complex software or
Data systems and computer science: Software Engineering Program
NASA Technical Reports Server (NTRS)
Zygielbaum, Arthur I.
1991-01-01
An external review of the Integrated Technology Plan for the Civil Space Program is presented. This review is specifically concerned with the Software Engineering Program. The goals of the Software Engineering Program are as follows: (1) improve NASA's ability to manage development, operation, and maintenance of complex software systems; (2) decrease NASA's cost and risk in engineering complex software systems; and (3) provide technology to assure safety and reliability of software in mission critical applications.
SNS programming environment user's guide
NASA Technical Reports Server (NTRS)
Tennille, Geoffrey M.; Howser, Lona M.; Humes, D. Creig; Cronin, Catherine K.; Bowen, John T.; Drozdowski, Joseph M.; Utley, Judith A.; Flynn, Theresa M.; Austin, Brenda A.
1992-01-01
The computing environment is briefly described for the Supercomputing Network Subsystem (SNS) of the Central Scientific Computing Complex of NASA Langley. The major SNS computers are a CRAY-2, a CRAY Y-MP, a CONVEX C-210, and a CONVEX C-220. The software is described that is common to all of these computers, including: the UNIX operating system, computer graphics, networking utilities, mass storage, and mathematical libraries. Also described is file management, validation, SNS configuration, documentation, and customer services.
NASA Astrophysics Data System (ADS)
Engfer, Christian; Pfüller, Enrico; Wiedemann, Manuel; Wolf, Jürgen; Lutz, Thorsten; Krämer, Ewald; Röser, Hans-Peter
2012-09-01
The Stratospheric Observatory for Infrared Astronomy (SOFIA) is a 2.5 m reflecting telescope housed in an open cavity on board of a Boeing 747SP. During observations, the cavity is exposed to transonic flow conditions. The oncoming boundary layer evolves into a free shear layer being responsible for optical aberrations and for aerodynamic and aeroacoustic disturbances within the cavity. While the aero-acoustical excitation of an airborne telescope can be minimized by using passive flow control devices, the aero-optical properties of the flow are difficult to improve. Hence it is important to know how much the image seen through the SOFIA telescope is perturbed by so called seeing effects. Prior to the SOFIA science fights Computational Fluid Dynamics (CFD) simulations using URANS and DES methods were carried out to determine the flow field within and above the cavity and hence in the optical path in order to provide an assessment of the aero-optical properties under baseline conditions. In addition and for validation purposes, out of focus images have been taken during flight with a Super Fast Diagnostic Camera (SFDC). Depending on the binning factor and the sub-array size, the SFDC is able to take and to read out images at very high frame rates. The paper explains the numerical approach based on CFD to evaluate the aero-optical properties of SOFIA. The CFD data is then compared to the high speed images taken by the SFDC during flight.
Plasma-based actuators for turbulent boundary layer control in transonic flow
NASA Astrophysics Data System (ADS)
Budovsky, A. D.; Polivanov, P. A.; Vishnyakov, O. I.; Sidorenko, A. A.
2017-10-01
The study is devoted to development of methods for active control of flow structure typical for the aircraft wings in transonic flow with turbulent boundary layer. The control strategy accepted in the study was based on using of the effects of plasma discharges interaction with miniature geometrical obstacles of various shapes. The conceptions were studied computationally using 3D RANS, URANS approaches. The results of the computations have shown that energy deposition can significantly change the flow pattern over the obstacles increasing their influence on the flow in boundary layer region. Namely, one of the most interesting and promising data were obtained for actuators basing on combination of vertical wedge with asymmetrical plasma discharge. The wedge considered is aligned with the local streamlines and protruding in the flow by 0.4-0.8 of local boundary layer thickness. The actuator produces negligible distortion of the flow at the absence of energy deposition. Energy deposition along the one side of the wedge results in longitudinal vortex formation in the wake of the actuator providing momentum exchange in the boundary layer. The actuator was manufactured and tested in wind tunnel experiments at Mach number 1.5 using the model of flat plate. The experimental data obtained by PIV proved the availability of the actuator.
A Two-Stage Procedure Toward the Efficient Implementation of PANS and Other Hybrid Turbulence Models
NASA Technical Reports Server (NTRS)
Abdol-Hamid, Khaled S.; Girimaji, Sharath S.
2004-01-01
The main objective of this article is to introduce and to show the implementation of a novel two-stage procedure to efficiently estimate the level of scale resolution possible for a given flow on a given grid for Partial Averaged Navier-Stokes (PANS) and other hybrid models. It has been found that the prescribed scale resolution can play a major role in obtaining accurate flow solutions. The first step is to solve the unsteady or steady Reynolds Averaged Navier-Stokes (URANS/RANS) equations. From this preprocessing step, the turbulence length-scale field is obtained. This is then used to compute the characteristic length-scale ratio between the turbulence scale and the grid spacing. Based on this ratio, we can assess the finest scale resolution that a given grid for a given flow can support. Along with other additional criteria, we are able to analytically identify the appropriate hybrid solver resolution for different regions of the flow. This procedure removes the grid dependency issue that affects the results produced by different hybrid procedures in solving unsteady flows. The formulation, implementation methodology, and validation example are presented. We implemented this capability in a production Computational Fluid Dynamics (CFD) code, PAB3D, for the simulation of unsteady flows.
Airfoil optimization for unsteady flows with application to high-lift noise reduction
NASA Astrophysics Data System (ADS)
Rumpfkeil, Markus Peer
The use of steady-state aerodynamic optimization methods in the computational fluid dynamic (CFD) community is fairly well established. In particular, the use of adjoint methods has proven to be very beneficial because their cost is independent of the number of design variables. The application of numerical optimization to airframe-generated noise, however, has not received as much attention, but with the significant quieting of modern engines, airframe noise now competes with engine noise. Optimal control techniques for unsteady flows are needed in order to be able to reduce airframe-generated noise. In this thesis, a general framework is formulated to calculate the gradient of a cost function in a nonlinear unsteady flow environment via the discrete adjoint method. The unsteady optimization algorithm developed in this work utilizes a Newton-Krylov approach since the gradient-based optimizer uses the quasi-Newton method BFGS, Newton's method is applied to the nonlinear flow problem, GMRES is used to solve the resulting linear problem inexactly, and last but not least the linear adjoint problem is solved using Bi-CGSTAB. The flow is governed by the unsteady two-dimensional compressible Navier-Stokes equations in conjunction with a one-equation turbulence model, which are discretized using structured grids and a finite difference approach. The effectiveness of the unsteady optimization algorithm is demonstrated by applying it to several problems of interest including shocktubes, pulses in converging-diverging nozzles, rotating cylinders, transonic buffeting, and an unsteady trailing-edge flow. In order to address radiated far-field noise, an acoustic wave propagation program based on the Ffowcs Williams and Hawkings (FW-H) formulation is implemented and validated. The general framework is then used to derive the adjoint equations for a novel hybrid URANS/FW-H optimization algorithm in order to be able to optimize the shape of airfoils based on their calculated far-field pressure fluctuations. Validation and application results for this novel hybrid URANS/FW-H optimization algorithm show that it is possible to optimize the shape of an airfoil in an unsteady flow environment to minimize its radiated far-field noise while maintaining good aerodynamic performance.
Federal Register 2010, 2011, 2012, 2013, 2014
2012-08-22
... Computer Software and Complex Electronics Used in Safety Systems of Nuclear Power Plants AGENCY: Nuclear...-1209, ``Software Requirement Specifications for Digital Computer Software and Complex Electronics used... Electronics Engineers (ANSI/IEEE) Standard 830-1998, ``IEEE Recommended Practice for Software Requirements...
Complex cellular logic computation using ribocomputing devices.
Green, Alexander A; Kim, Jongmin; Ma, Duo; Silver, Pamela A; Collins, James J; Yin, Peng
2017-08-03
Synthetic biology aims to develop engineering-driven approaches to the programming of cellular functions that could yield transformative technologies. Synthetic gene circuits that combine DNA, protein, and RNA components have demonstrated a range of functions such as bistability, oscillation, feedback, and logic capabilities. However, it remains challenging to scale up these circuits owing to the limited number of designable, orthogonal, high-performance parts, the empirical and often tedious composition rules, and the requirements for substantial resources for encoding and operation. Here, we report a strategy for constructing RNA-only nanodevices to evaluate complex logic in living cells. Our 'ribocomputing' systems are composed of de-novo-designed parts and operate through predictable and designable base-pairing rules, allowing the effective in silico design of computing devices with prescribed configurations and functions in complex cellular environments. These devices operate at the post-transcriptional level and use an extended RNA transcript to co-localize all circuit sensing, computation, signal transduction, and output elements in the same self-assembled molecular complex, which reduces diffusion-mediated signal losses, lowers metabolic cost, and improves circuit reliability. We demonstrate that ribocomputing devices in Escherichia coli can evaluate two-input logic with a dynamic range up to 900-fold and scale them to four-input AND, six-input OR, and a complex 12-input expression (A1 AND A2 AND NOT A1*) OR (B1 AND B2 AND NOT B2*) OR (C1 AND C2) OR (D1 AND D2) OR (E1 AND E2). Successful operation of ribocomputing devices based on programmable RNA interactions suggests that systems employing the same design principles could be implemented in other host organisms or in extracellular settings.
Hybrid and concatenated coding applications.
NASA Technical Reports Server (NTRS)
Hofman, L. B.; Odenwalder, J. P.
1972-01-01
Results of a study to evaluate the performance and implementation complexity of a concatenated and a hybrid coding system for moderate-speed deep-space applications. It is shown that with a total complexity of less than three times that of the basic Viterbi decoder, concatenated coding improves a constraint length 8 rate 1/3 Viterbi decoding system by 1.1 and 2.6 dB at bit error probabilities of 0.0001 and one hundred millionth, respectively. With a somewhat greater total complexity, the hybrid coding system is shown to obtain a 0.9-dB computational performance improvement over the basic rate 1/3 sequential decoding system. Although substantial, these complexities are much less than those required to achieve the same performances with more complex Viterbi or sequential decoder systems.
A resource management architecture based on complex network theory in cloud computing federation
NASA Astrophysics Data System (ADS)
Zhang, Zehua; Zhang, Xuejie
2011-10-01
Cloud Computing Federation is a main trend of Cloud Computing. Resource Management has significant effect on the design, realization, and efficiency of Cloud Computing Federation. Cloud Computing Federation has the typical characteristic of the Complex System, therefore, we propose a resource management architecture based on complex network theory for Cloud Computing Federation (abbreviated as RMABC) in this paper, with the detailed design of the resource discovery and resource announcement mechanisms. Compare with the existing resource management mechanisms in distributed computing systems, a Task Manager in RMABC can use the historical information and current state data get from other Task Managers for the evolution of the complex network which is composed of Task Managers, thus has the advantages in resource discovery speed, fault tolerance and adaptive ability. The result of the model experiment confirmed the advantage of RMABC in resource discovery performance.
NASA Technical Reports Server (NTRS)
Hayden, W. L.; Robinson, L. H.
1972-01-01
Spectral analyses of angle-modulated communication systems is studied by: (1) performing a literature survey of candidate power spectrum computational techniques, determining the computational requirements, and formulating a mathematical model satisfying these requirements; (2) implementing the model on UNIVAC 1230 digital computer as the Spectral Analysis Program (SAP); and (3) developing the hardware specifications for a data acquisition system which will acquire an input modulating signal for SAP. The SAP computational technique uses extended fast Fourier transform and represents a generalized approach for simple and complex modulating signals.
Autonomous control systems: applications to remote sensing and image processing
NASA Astrophysics Data System (ADS)
Jamshidi, Mohammad
2001-11-01
One of the main challenges of any control (or image processing) paradigm is being able to handle complex systems under unforeseen uncertainties. A system may be called complex here if its dimension (order) is too high and its model (if available) is nonlinear, interconnected, and information on the system is uncertain such that classical techniques cannot easily handle the problem. Examples of complex systems are power networks, space robotic colonies, national air traffic control system, and integrated manufacturing plant, the Hubble Telescope, the International Space Station, etc. Soft computing, a consortia of methodologies such as fuzzy logic, neuro-computing, genetic algorithms and genetic programming, has proven to be powerful tools for adding autonomy and semi-autonomy to many complex systems. For such systems the size of soft computing control architecture will be nearly infinite. In this paper new paradigms using soft computing approaches are utilized to design autonomous controllers and image enhancers for a number of application areas. These applications are satellite array formations for synthetic aperture radar interferometry (InSAR) and enhancement of analog and digital images.
Local spatio-temporal analysis in vision systems
NASA Astrophysics Data System (ADS)
Geisler, Wilson S.; Bovik, Alan; Cormack, Lawrence; Ghosh, Joydeep; Gildeen, David
1994-07-01
The aims of this project are the following: (1) develop a physiologically and psychophysically based model of low-level human visual processing (a key component of which are local frequency coding mechanisms); (2) develop image models and image-processing methods based upon local frequency coding; (3) develop algorithms for performing certain complex visual tasks based upon local frequency representations, (4) develop models of human performance in certain complex tasks based upon our understanding of low-level processing; and (5) develop a computational testbed for implementing, evaluating and visualizing the proposed models and algorithms, using a massively parallel computer. Progress has been substantial on all aims. The highlights include the following: (1) completion of a number of psychophysical and physiological experiments revealing new, systematic and exciting properties of the primate (human and monkey) visual system; (2) further development of image models that can accurately represent the local frequency structure in complex images; (3) near completion in the construction of the Texas Active Vision Testbed; (4) development and testing of several new computer vision algorithms dealing with shape-from-texture, shape-from-stereo, and depth-from-focus; (5) implementation and evaluation of several new models of human visual performance; and (6) evaluation, purchase and installation of a MasPar parallel computer.
Simulating complex intracellular processes using object-oriented computational modelling.
Johnson, Colin G; Goldman, Jacki P; Gullick, William J
2004-11-01
The aim of this paper is to give an overview of computer modelling and simulation in cellular biology, in particular as applied to complex biochemical processes within the cell. This is illustrated by the use of the techniques of object-oriented modelling, where the computer is used to construct abstractions of objects in the domain being modelled, and these objects then interact within the computer to simulate the system and allow emergent properties to be observed. The paper also discusses the role of computer simulation in understanding complexity in biological systems, and the kinds of information which can be obtained about biology via simulation.
Al-Sadoon, Mohammed A. G.; Zuid, Abdulkareim; Jones, Stephen M. R.; Noras, James M.
2017-01-01
This paper proposes a new low complexity angle of arrival (AOA) method for signal direction estimation in multi-element smart wireless communication systems. The new method estimates the AOAs of the received signals directly from the received signals with significantly reduced complexity since it does not need to construct the correlation matrix, invert the matrix or apply eigen-decomposition, which are computationally expensive. A mathematical model of the proposed method is illustrated and then verified using extensive computer simulations. Both linear and circular sensors arrays are studied using various numerical examples. The method is systematically compared with other common and recently introduced AOA methods over a wide range of scenarios. The simulated results show that the new method has several advantages in terms of reduced complexity and improved accuracy under the assumptions of correlated signals and limited numbers of snapshots. PMID:29140313
Al-Sadoon, Mohammed A G; Ali, Nazar T; Dama, Yousf; Zuid, Abdulkareim; Jones, Stephen M R; Abd-Alhameed, Raed A; Noras, James M
2017-11-15
This paper proposes a new low complexity angle of arrival (AOA) method for signal direction estimation in multi-element smart wireless communication systems. The new method estimates the AOAs of the received signals directly from the received signals with significantly reduced complexity since it does not need to construct the correlation matrix, invert the matrix or apply eigen-decomposition, which are computationally expensive. A mathematical model of the proposed method is illustrated and then verified using extensive computer simulations. Both linear and circular sensors arrays are studied using various numerical examples. The method is systematically compared with other common and recently introduced AOA methods over a wide range of scenarios. The simulated results show that the new method has several advantages in terms of reduced complexity and improved accuracy under the assumptions of correlated signals and limited numbers of snapshots.
Gao, Jinting; Liu, Yaqing; Lin, Xiaodong; Deng, Jiankang; Yin, Jinjin; Wang, Shuo
2017-10-25
Wiring a series of simple logic gates to process complex data is significantly important and a large challenge for untraditional molecular computing systems. The programmable property of DNA endows its powerful application in molecular computing. In our investigation, it was found that DNA exhibits excellent peroxidase-like activity in a colorimetric system of TMB/H 2 O 2 /Hemin (TMB, 3,3', 5,5'-Tetramethylbenzidine) in the presence of K + and Cu 2+ , which is significantly inhibited by the addition of an antioxidant. According to the modulated catalytic activity of this DNA-based catalyst, three cascade logic gates including AND-OR-INH (INHIBIT), AND-INH and OR-INH were successfully constructed. Interestingly, by only modulating the concentration of Cu 2+ , a majority logic gate with a single-vote veto function was realized following the same threshold value as that of the cascade logic gates. The strategy is quite straightforward and versatile and provides an instructive method for constructing multiple logic gates on a simple platform to implement complex molecular computing.
Kwong, Hoi-Ki; Lo, Po-Kam; Lau, Kai-Chung; Lau, Tai-Chu
2011-04-14
The manganese(V) nitrido complex (PPh(4))(2)[Mn(N)(CN)(4)] is an active catalyst for alkene epoxidation and alcohol oxidation using H(2)O(2) as an oxidant. The catalytic oxidation is greatly enhanced by the addition of just one equivalent of acetic acid. The oxidation of ethene by this system has been studied computationally by the DFT method.
Evaluation of the Cape Cod Advanced Public Transit System : phase 1 and 2
DOT National Transportation Integrated Search
2000-03-01
This case study is one of a series of case studies that examine procurement approaches used to deliver Intelligent Transportation System (ITS) projects. ITS projects are often complex and leverage the latest technology in telecommunications, computer...
Computing with dynamical systems based on insulator-metal-transition oscillators
NASA Astrophysics Data System (ADS)
Parihar, Abhinav; Shukla, Nikhil; Jerry, Matthew; Datta, Suman; Raychowdhury, Arijit
2017-04-01
In this paper, we review recent work on novel computing paradigms using coupled oscillatory dynamical systems. We explore systems of relaxation oscillators based on linear state transitioning devices, which switch between two discrete states with hysteresis. By harnessing the dynamics of complex, connected systems, we embrace the philosophy of "let physics do the computing" and demonstrate how complex phase and frequency dynamics of such systems can be controlled, programmed, and observed to solve computationally hard problems. Although our discussion in this paper is limited to insulator-to-metallic state transition devices, the general philosophy of such computing paradigms can be translated to other mediums including optical systems. We present the necessary mathematical treatments necessary to understand the time evolution of these systems and demonstrate through recent experimental results the potential of such computational primitives.
PROCESS FOR THE RECOVERY OF URANIUM
Morris, G.O.
1955-06-21
This patent relates to a process for the recovery of uranium from impure uranium tetrafluoride. The process consists essentially of the steps of dissolving the impure uranium tetrafluoride in excess dilute sulfuric acid in the presence of excess hydrogen peroxide, precipitating ammonium uranate from the solution so formed by adding an excess of aqueous ammonia, dissolving the precipitate in sulfuric acid and adding hydrogen peroxide to precipitate uranium peroxdde.
Development of Onboard Computer Complex for Russian Segment of ISS
NASA Technical Reports Server (NTRS)
Branets, V.; Brand, G.; Vlasov, R.; Graf, I.; Clubb, J.; Mikrin, E.; Samitov, R.
1998-01-01
Report present a description of the Onboard Computer Complex (CC) that was developed during the period of 1994-1998 for the Russian Segment of ISS. The system was developed in co-operation with NASA and ESA. ESA developed a new computation system under the RSC Energia Technical Assignment, called DMS-R. The CC also includes elements developed by Russian experts and organizations. A general architecture of the computer system and the characteristics of primary elements of this system are described. The system was integrated at RSC Energia with the participation of American and European specialists. The report contains information on software simulators, verification and de-bugging facilities witch were been developed for both stand-alone and integrated tests and verification. This CC serves as the basis for the Russian Segment Onboard Control Complex on ISS.
CCOMP: An efficient algorithm for complex roots computation of determinantal equations
NASA Astrophysics Data System (ADS)
Zouros, Grigorios P.
2018-01-01
In this paper a free Python algorithm, entitled CCOMP (Complex roots COMPutation), is developed for the efficient computation of complex roots of determinantal equations inside a prescribed complex domain. The key to the method presented is the efficient determination of the candidate points inside the domain which, in their close neighborhood, a complex root may lie. Once these points are detected, the algorithm proceeds to a two-dimensional minimization problem with respect to the minimum modulus eigenvalue of the system matrix. In the core of CCOMP exist three sub-algorithms whose tasks are the efficient estimation of the minimum modulus eigenvalues of the system matrix inside the prescribed domain, the efficient computation of candidate points which guarantee the existence of minima, and finally, the computation of minima via bound constrained minimization algorithms. Theoretical results and heuristics support the development and the performance of the algorithm, which is discussed in detail. CCOMP supports general complex matrices, and its efficiency, applicability and validity is demonstrated to a variety of microwave applications.
Synthetic mixed-signal computation in living cells
Rubens, Jacob R.; Selvaggio, Gianluca; Lu, Timothy K.
2016-01-01
Living cells implement complex computations on the continuous environmental signals that they encounter. These computations involve both analogue- and digital-like processing of signals to give rise to complex developmental programs, context-dependent behaviours and homeostatic activities. In contrast to natural biological systems, synthetic biological systems have largely focused on either digital or analogue computation separately. Here we integrate analogue and digital computation to implement complex hybrid synthetic genetic programs in living cells. We present a framework for building comparator gene circuits to digitize analogue inputs based on different thresholds. We then demonstrate that comparators can be predictably composed together to build band-pass filters, ternary logic systems and multi-level analogue-to-digital converters. In addition, we interface these analogue-to-digital circuits with other digital gene circuits to enable concentration-dependent logic. We expect that this hybrid computational paradigm will enable new industrial, diagnostic and therapeutic applications with engineered cells. PMID:27255669
Enabling Co-Design of Multi-Layer Exascale Storage Architectures
DOE Office of Scientific and Technical Information (OSTI.GOV)
Carothers, Christopher
Growing demands for computing power in applications such as energy production, climate analysis, computational chemistry, and bioinformatics have propelled computing systems toward the exascale: systems with 10 18 floating-point operations per second. These systems, to be designed and constructed over the next decade, will create unprecedented challenges in component counts, power consumption, resource limitations, and system complexity. Data storage and access are an increasingly important and complex component in extreme-scale computing systems, and significant design work is needed to develop successful storage hardware and software architectures at exascale. Co-design of these systems will be necessary to find the best possiblemore » design points for exascale systems. The goal of this work has been to enable the exploration and co-design of exascale storage systems by providing a detailed, accurate, and highly parallel simulation of exascale storage and the surrounding environment. Specifically, this simulation has (1) portrayed realistic application checkpointing and analysis workloads, (2) captured the complexity, scale, and multilayer nature of exascale storage hardware and software, and (3) executed in a timeframe that enables “what if'” exploration of design concepts. We developed models of the major hardware and software components in an exascale storage system, as well as the application I/O workloads that drive them. We used our simulation system to investigate critical questions in reliability and concurrency at exascale, helping guide the design of future exascale hardware and software architectures. Additionally, we provided this system to interested vendors and researchers so that others can explore the design space. We validated the capabilities of our simulation environment by configuring the simulation to represent the Argonne Leadership Computing Facility Blue Gene/Q system and comparing simulation results for application I/O patterns to the results of executions of these I/O kernels on the actual system.« less
NASA Astrophysics Data System (ADS)
Sarkar, Shubhra; Ramanathan, N.; Gopi, R.; Sundararajan, K.
2017-12-01
Hydrogen bonded interaction of pyrrole multimer and acetylene-pyrrole complexes were studied in N2 and p-H2 matrixes. DFT computations showed T-shaped geometry for the pyrrole dimer and cyclic complex for the trimer and tetramer were the most stable structures, stabilized by Nsbnd H⋯π interactions. The experimental vibrational wavenumbers observed in N2 and p-H2 matrixes for the pyrrole multimers were correlated with the computed wavenumbers. Computations performed at MP2/aug-cc-pVDZ level of theory showed that C2H2 and C4H5N forms 1:1 hydrogen-bonded complexes stabilized by Csbnd H⋯π interaction (Complex A), Nsbnd H⋯π interaction (Complex B) and π⋯π interaction (Complex C), where the former complex is the global minimum and latter two complexes were the first and second local minima, respectively. Experimentally, 1:1 C2H2sbnd C4H5N complexes A (global minimum) and B (first local minimum) were identified from the shifts in the Nsbnd H stretching, Nsbnd H bending, Csbnd H bending region of pyrrole and Csbnd H asymmetric stretching and bending region of C2H2 in N2 and p-H2 matrixes. Computations were also performed for the higher complexes and found two minima corresponding to the 1:2 C2H2sbnd C4H5N and three minima for the 2:1 C2H2sbnd C4H5N complexes. Experimentally the global minimum 1:2 and 2:1 C2H2sbnd C4H5N complexes were identified in N2 and p-H2 matrixes.
Butchosa, C; Simon, S; Blancafort, L; Voityuk, A
2012-07-12
Because hole transfer from nucleobases to amino acid residues in DNA-protein complexes can prevent oxidative damage of DNA in living cells, computational modeling of the process is of high interest. We performed MS-CASPT2 calculations of several model structures of π-stacked guanine and indole and derived electron-transfer (ET) parameters for these systems using the generalized Mulliken-Hush (GMH) method. We show that the two-state model commonly applied to treat thermal ET between adjacent donor and acceptor is of limited use for the considered systems because of the small gap between the ground and first excited states in the indole radical cation. The ET parameters obtained within the two-state GMH scheme can deviate significantly from the corresponding matrix elements of the two-state effective Hamiltonian based on the GMH treatment of three adiabatic states. The computed values of diabatic energies and electronic couplings provide benchmarks to assess the performance of less sophisticated computational methods.
Schillinger, Dean; McNamara, Danielle; Crossley, Scott; Lyles, Courtney; Moffet, Howard H; Sarkar, Urmimala; Duran, Nicholas; Allen, Jill; Liu, Jennifer; Oryn, Danielle; Ratanawongsa, Neda; Karter, Andrew J
2017-01-01
Health systems are heavily promoting patient portals. However, limited health literacy (HL) can restrict online communication via secure messaging (SM) because patients' literacy skills must be sufficient to convey and comprehend content while clinicians must encourage and elicit communication from patients and match patients' literacy level. This paper describes the Employing Computational Linguistics to Improve Patient-Provider Secure Email (ECLIPPSE) study, an interdisciplinary effort bringing together scientists in communication, computational linguistics, and health services to employ computational linguistic methods to (1) create a novel Linguistic Complexity Profile (LCP) to characterize communications of patients and clinicians and demonstrate its validity and (2) examine whether providers accommodate communication needs of patients with limited HL by tailoring their SM responses. We will study >5 million SMs generated by >150,000 ethnically diverse type 2 diabetes patients and >9000 clinicians from two settings: an integrated delivery system and a public (safety net) system. Finally, we will then create an LCP-based automated aid that delivers real-time feedback to clinicians to reduce the linguistic complexity of their SMs. This research will support health systems' journeys to become health literate healthcare organizations and reduce HL-related disparities in diabetes care.
Increasing complexity with quantum physics.
Anders, Janet; Wiesner, Karoline
2011-09-01
We argue that complex systems science and the rules of quantum physics are intricately related. We discuss a range of quantum phenomena, such as cryptography, computation and quantum phases, and the rules responsible for their complexity. We identify correlations as a central concept connecting quantum information and complex systems science. We present two examples for the power of correlations: using quantum resources to simulate the correlations of a stochastic process and to implement a classically impossible computational task.
NASA Technical Reports Server (NTRS)
Svehla, R. A.; Mcbride, B. J.
1973-01-01
A FORTRAN IV computer program for the calculation of the thermodynamic and transport properties of complex mixtures is described. The program has the capability of performing calculations such as:(1) chemical equilibrium for assigned thermodynamic states, (2) theoretical rocket performance for both equilibrium and frozen compositions during expansion, (3) incident and reflected shock properties, and (4) Chapman-Jouguet detonation properties. Condensed species, as well as gaseous species, are considered in the thermodynamic calculation; but only the gaseous species are considered in the transport calculations.
High Sensitive Scintillation Observations At Very Low Frequencies
NASA Astrophysics Data System (ADS)
Konovalenko, A. A.; Falkovich, I. S.; Kalinichenko, N. N.; Olyak, M. R.; Lecacheux, A.; Rosolen, C.; Bougeret, J.-L.; Rucker, H. O.; Tokarev, Yu.
The observation of interplanetary scintillations of compact radio sources is powerful method of solar wind diagnostics. This method is developed mainly at decimeter- meter wavelengths. New possibilities are opened at extremely low frequencies (decameter waves) especially at large elongations. Now this approach is being actively developed using high effective decameter antennas UTR-2, URAN and Nancay Decameter Array. New class of back-end facility like high dynamic range, high resolution digital spectral processors, as well as dynamic spectra determination ideology give us new opportunities for distinguishing of the ionospheric and interplanetary scintillations and for observations of large number of radio sources, whith different angular sizes and elongations, even for the cases of rather weak objects.
PROCESSES OF RECLAIMING URANIUM FROM SOLUTIONS
Zumwalt, L.R.
1959-02-10
A process is described for reclaiming residual enriched uranium from calutron wash solutions containing Fe, Cr, Cu, Ni, and Mn as impurities. The solution is adjusted to a pH of between 2 and 4 and is contacted with a metallic reducing agent, such as iron or zinc, in order to reduce the copper to metal and thereby remove it from the solution. At the same time the uranium present is reduced to the uranous state The solution is then contacted with a precipitate of zinc hydroxide or barium carbonate in order to precipitate and carry uranium, iron, and chromium away from the nickel and manganese ions in the solution. The uranium is then recovered fronm this precipitate.
Yang, Tzuhsiung; Berry, John F
2018-06-04
The computation of nuclear second derivatives of energy, or the nuclear Hessian, is an essential routine in quantum chemical investigations of ground and transition states, thermodynamic calculations, and molecular vibrations. Analytic nuclear Hessian computations require the resolution of costly coupled-perturbed self-consistent field (CP-SCF) equations, while numerical differentiation of analytic first derivatives has an unfavorable 6 N ( N = number of atoms) prefactor. Herein, we present a new method in which grid computing is used to accelerate and/or enable the evaluation of the nuclear Hessian via numerical differentiation: NUMFREQ@Grid. Nuclear Hessians were successfully evaluated by NUMFREQ@Grid at the DFT level as well as using RIJCOSX-ZORA-MP2 or RIJCOSX-ZORA-B2PLYP for a set of linear polyacenes with systematically increasing size. For the larger members of this group, NUMFREQ@Grid was found to outperform the wall clock time of analytic Hessian evaluation; at the MP2 or B2LYP levels, these Hessians cannot even be evaluated analytically. We also evaluated a 156-atom catalytically relevant open-shell transition metal complex and found that NUMFREQ@Grid is faster (7.7 times shorter wall clock time) and less demanding (4.4 times less memory requirement) than an analytic Hessian. Capitalizing on the capabilities of parallel grid computing, NUMFREQ@Grid can outperform analytic methods in terms of wall time, memory requirements, and treatable system size. The NUMFREQ@Grid method presented herein demonstrates how grid computing can be used to facilitate embarrassingly parallel computational procedures and is a pioneer for future implementations.
Statistical Techniques Complement UML When Developing Domain Models of Complex Dynamical Biosystems.
Williams, Richard A; Timmis, Jon; Qwarnstrom, Eva E
2016-01-01
Computational modelling and simulation is increasingly being used to complement traditional wet-lab techniques when investigating the mechanistic behaviours of complex biological systems. In order to ensure computational models are fit for purpose, it is essential that the abstracted view of biology captured in the computational model, is clearly and unambiguously defined within a conceptual model of the biological domain (a domain model), that acts to accurately represent the biological system and to document the functional requirements for the resultant computational model. We present a domain model of the IL-1 stimulated NF-κB signalling pathway, which unambiguously defines the spatial, temporal and stochastic requirements for our future computational model. Through the development of this model, we observe that, in isolation, UML is not sufficient for the purpose of creating a domain model, and that a number of descriptive and multivariate statistical techniques provide complementary perspectives, in particular when modelling the heterogeneity of dynamics at the single-cell level. We believe this approach of using UML to define the structure and interactions within a complex system, along with statistics to define the stochastic and dynamic nature of complex systems, is crucial for ensuring that conceptual models of complex dynamical biosystems, which are developed using UML, are fit for purpose, and unambiguously define the functional requirements for the resultant computational model.
Statistical Techniques Complement UML When Developing Domain Models of Complex Dynamical Biosystems
Timmis, Jon; Qwarnstrom, Eva E.
2016-01-01
Computational modelling and simulation is increasingly being used to complement traditional wet-lab techniques when investigating the mechanistic behaviours of complex biological systems. In order to ensure computational models are fit for purpose, it is essential that the abstracted view of biology captured in the computational model, is clearly and unambiguously defined within a conceptual model of the biological domain (a domain model), that acts to accurately represent the biological system and to document the functional requirements for the resultant computational model. We present a domain model of the IL-1 stimulated NF-κB signalling pathway, which unambiguously defines the spatial, temporal and stochastic requirements for our future computational model. Through the development of this model, we observe that, in isolation, UML is not sufficient for the purpose of creating a domain model, and that a number of descriptive and multivariate statistical techniques provide complementary perspectives, in particular when modelling the heterogeneity of dynamics at the single-cell level. We believe this approach of using UML to define the structure and interactions within a complex system, along with statistics to define the stochastic and dynamic nature of complex systems, is crucial for ensuring that conceptual models of complex dynamical biosystems, which are developed using UML, are fit for purpose, and unambiguously define the functional requirements for the resultant computational model. PMID:27571414
Mathematical and Computational Modeling in Complex Biological Systems
Li, Wenyang; Zhu, Xiaoliang
2017-01-01
The biological process and molecular functions involved in the cancer progression remain difficult to understand for biologists and clinical doctors. Recent developments in high-throughput technologies urge the systems biology to achieve more precise models for complex diseases. Computational and mathematical models are gradually being used to help us understand the omics data produced by high-throughput experimental techniques. The use of computational models in systems biology allows us to explore the pathogenesis of complex diseases, improve our understanding of the latent molecular mechanisms, and promote treatment strategy optimization and new drug discovery. Currently, it is urgent to bridge the gap between the developments of high-throughput technologies and systemic modeling of the biological process in cancer research. In this review, we firstly studied several typical mathematical modeling approaches of biological systems in different scales and deeply analyzed their characteristics, advantages, applications, and limitations. Next, three potential research directions in systems modeling were summarized. To conclude, this review provides an update of important solutions using computational modeling approaches in systems biology. PMID:28386558
Mathematical and Computational Modeling in Complex Biological Systems.
Ji, Zhiwei; Yan, Ke; Li, Wenyang; Hu, Haigen; Zhu, Xiaoliang
2017-01-01
The biological process and molecular functions involved in the cancer progression remain difficult to understand for biologists and clinical doctors. Recent developments in high-throughput technologies urge the systems biology to achieve more precise models for complex diseases. Computational and mathematical models are gradually being used to help us understand the omics data produced by high-throughput experimental techniques. The use of computational models in systems biology allows us to explore the pathogenesis of complex diseases, improve our understanding of the latent molecular mechanisms, and promote treatment strategy optimization and new drug discovery. Currently, it is urgent to bridge the gap between the developments of high-throughput technologies and systemic modeling of the biological process in cancer research. In this review, we firstly studied several typical mathematical modeling approaches of biological systems in different scales and deeply analyzed their characteristics, advantages, applications, and limitations. Next, three potential research directions in systems modeling were summarized. To conclude, this review provides an update of important solutions using computational modeling approaches in systems biology.
A Computer Story: Complexity from Simplicity
ERIC Educational Resources Information Center
DeLeo, Gary; Weidenhammer, Amanda; Wecht, Kristen
2012-01-01
In this technological age, digital devices are conspicuous examples of extraordinary complexity. When a user clicks on computer icons or presses calculator buttons, these devices channel electricity through a complex system of decision-making circuits. Yet, in spite of this remarkable complexity, the hearts of these devices are components that…
Parallel algorithms for mapping pipelined and parallel computations
NASA Technical Reports Server (NTRS)
Nicol, David M.
1988-01-01
Many computational problems in image processing, signal processing, and scientific computing are naturally structured for either pipelined or parallel computation. When mapping such problems onto a parallel architecture it is often necessary to aggregate an obvious problem decomposition. Even in this context the general mapping problem is known to be computationally intractable, but recent advances have been made in identifying classes of problems and architectures for which optimal solutions can be found in polynomial time. Among these, the mapping of pipelined or parallel computations onto linear array, shared memory, and host-satellite systems figures prominently. This paper extends that work first by showing how to improve existing serial mapping algorithms. These improvements have significantly lower time and space complexities: in one case a published O(nm sup 3) time algorithm for mapping m modules onto n processors is reduced to an O(nm log m) time complexity, and its space requirements reduced from O(nm sup 2) to O(m). Run time complexity is further reduced with parallel mapping algorithms based on these improvements, which run on the architecture for which they create the mappings.
HyperForest: A high performance multi-processor architecture for real-time intelligent systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Garcia, P. Jr.; Rebeil, J.P.; Pollard, H.
1997-04-01
Intelligent Systems are characterized by the intensive use of computer power. The computer revolution of the last few years is what has made possible the development of the first generation of Intelligent Systems. Software for second generation Intelligent Systems will be more complex and will require more powerful computing engines in order to meet real-time constraints imposed by new robots, sensors, and applications. A multiprocessor architecture was developed that merges the advantages of message-passing and shared-memory structures: expendability and real-time compliance. The HyperForest architecture will provide an expandable real-time computing platform for computationally intensive Intelligent Systems and open the doorsmore » for the application of these systems to more complex tasks in environmental restoration and cleanup projects, flexible manufacturing systems, and DOE`s own production and disassembly activities.« less
Investigations of Flow Over a Hemisphere Using Numerical Simulations (Postprint)
2015-06-22
ranging from missile defense, remote sensing , and imaging . An important aspect of these applications is determining the effective beam-on-target...Stokes (URANS), detached eddy simulation (DES), and hybrid RANS/LES. The numerical results were compared with the experiment conducted at Auburn...turret. Using the DES and hybrid RANS/LES turbulence models, Loci-Chem was able to capture the unsteady flow structures, such as the shear layer
Seaborg, G.T.; Orlemann, E.F.; Jensen, L.H.
1958-12-23
A method of obtaining substantially pure uranium from a uranium composition contaminated with light element impurities such as sodium, magnesium, beryllium, and the like is described. An acidic aqueous solution containing tetravalent uranium is treated with a soluble molybdate to form insoluble uranous molybdate which is removed. This material after washing is dissolved in concentrated nitric acid to obtaln a uranyl nitrate solution from which highly purified uranium is obtained by extraction with ether.
NASA Technical Reports Server (NTRS)
Shih, Tsan-Hsing; Liu, Nan-Suey
2012-01-01
This paper presents the numerical simulations of the Jet-A spray reacting flow in a single element lean direct injection (LDI) injector by using the National Combustion Code (NCC) with and without invoking the Eulerian scalar probability density function (PDF) method. The flow field is calculated by using the Reynolds averaged Navier-Stokes equations (RANS and URANS) with nonlinear turbulence models, and when the scalar PDF method is invoked, the energy and compositions or species mass fractions are calculated by solving the equation of an ensemble averaged density-weighted fine-grained probability density function that is referred to here as the averaged probability density function (APDF). A nonlinear model for closing the convection term of the scalar APDF equation is used in the presented simulations and will be briefly described. Detailed comparisons between the results and available experimental data are carried out. Some positive findings of invoking the Eulerian scalar PDF method in both improving the simulation quality and reducing the computing cost are observed.
NASA Technical Reports Server (NTRS)
Bailey, R. T.; Shih, T. I.-P.; Nguyen, H. L.; Roelke, R. J.
1990-01-01
An efficient computer program, called GRID2D/3D, was developed to generate single and composite grid systems within geometrically complex two- and three-dimensional (2- and 3-D) spatial domains that can deform with time. GRID2D/3D generates single grid systems by using algebraic grid generation methods based on transfinite interpolation in which the distribution of grid points within the spatial domain is controlled by stretching functions. All single grid systems generated by GRID2D/3D can have grid lines that are continuous and differentiable everywhere up to the second-order. Also, grid lines can intersect boundaries of the spatial domain orthogonally. GRID2D/3D generates composite grid systems by patching together two or more single grid systems. The patching can be discontinuous or continuous. For continuous composite grid systems, the grid lines are continuous and differentiable everywhere up to the second-order except at interfaces where different single grid systems meet. At interfaces where different single grid systems meet, the grid lines are only differentiable up to the first-order. For 2-D spatial domains, the boundary curves are described by using either cubic or tension spline interpolation. For 3-D spatial domains, the boundary surfaces are described by using either linear Coon's interpolation, bi-hyperbolic spline interpolation, or a new technique referred to as 3-D bi-directional Hermite interpolation. Since grid systems generated by algebraic methods can have grid lines that overlap one another, GRID2D/3D contains a graphics package for evaluating the grid systems generated. With the graphics package, the user can generate grid systems in an interactive manner with the grid generation part of GRID2D/3D. GRID2D/3D is written in FORTRAN 77 and can be run on any IBM PC, XT, or AT compatible computer. In order to use GRID2D/3D on workstations or mainframe computers, some minor modifications must be made in the graphics part of the program; no modifications are needed in the grid generation part of the program. The theory and method used in GRID2D/3D is described.
BeeSim: Leveraging Wearable Computers in Participatory Simulations with Young Children
ERIC Educational Resources Information Center
Peppler, Kylie; Danish, Joshua; Zaitlen, Benjamin; Glosson, Diane; Jacobs, Alexander; Phelps, David
2010-01-01
New technologies have enabled students to become active participants in computational simulations of dynamic and complex systems (called Participatory Simulations), providing a "first-person"perspective on complex systems. However, most existing Participatory Simulations have targeted older children, teens, and adults assuming that such concepts…
Predictor - Predictive Reaction Design via Informatics, Computation and Theories of Reactivity
2017-10-10
into more complex and valuable molecules, but are limited by: 1. The extensive time it takes to design and optimize a synthesis 2. Multi-step...system. As it is fully compatible to the industry standard SQL, designing a server- based system at a later time will be trivial. Producing a JAVA front...Report: PREDICTOR - Predictive REaction Design via Informatics, Computation and Theories of Reactivity The goal of this program was to create a cyber
Planetary Data Workshop, Part 2
NASA Technical Reports Server (NTRS)
1984-01-01
Technical aspects of the Planetary Data System (PDS) are addressed. Methods and tools for maintaining and accessing large, complex sets of data are discussed. The specific software and applications needed for processing imaging and non-imaging science data are reviewed. The need for specific software that provides users with information on the location and geometry of scientific observations is discussed. Computer networks and user interface to the PDS are covered along with Computer hardware available to this data system.
NASA Technical Reports Server (NTRS)
Torres-Pomales, Wilfredo
2014-01-01
A system is safety-critical if its failure can endanger human life or cause significant damage to property or the environment. State-of-the-art computer systems on commercial aircraft are highly complex, software-intensive, functionally integrated, and network-centric systems of systems. Ensuring that such systems are safe and comply with existing safety regulations is costly and time-consuming as the level of rigor in the development process, especially the validation and verification activities, is determined by considerations of system complexity and safety criticality. A significant degree of care and deep insight into the operational principles of these systems is required to ensure adequate coverage of all design implications relevant to system safety. Model-based development methodologies, methods, tools, and techniques facilitate collaboration and enable the use of common design artifacts among groups dealing with different aspects of the development of a system. This paper examines the application of model-based development to complex and safety-critical aircraft computer systems. Benefits and detriments are identified and an overall assessment of the approach is given.
20170312 - Computer Simulation of Developmental ...
Rationale: Recent progress in systems toxicology and synthetic biology have paved the way to new thinking about in vitro/in silico modeling of developmental processes and toxicities, both for embryological and reproductive impacts. Novel in vitro platforms such as 3D organotypic culture models, engineered microscale tissues and complex microphysiological systems (MPS), together with computational models and computer simulation of tissue dynamics, lend themselves to a integrated testing strategies for predictive toxicology. As these emergent methodologies continue to evolve, they must be integrally tied to maternal/fetal physiology and toxicity of the developing individual across early lifestage transitions, from fertilization to birth, through puberty and beyond. Scope: This symposium will focus on how the novel technology platforms can help now and in the future, with in vitro/in silico modeling of complex biological systems for developmental and reproductive toxicity issues, and translating systems models into integrative testing strategies. The symposium is based on three main organizing principles: (1) that novel in vitro platforms with human cells configured in nascent tissue architectures with a native microphysiological environments yield mechanistic understanding of developmental and reproductive impacts of drug/chemical exposures; (2) that novel in silico platforms with high-throughput screening (HTS) data, biologically-inspired computational models of
Computer Simulation of Developmental Processes and ...
Rationale: Recent progress in systems toxicology and synthetic biology have paved the way to new thinking about in vitro/in silico modeling of developmental processes and toxicities, both for embryological and reproductive impacts. Novel in vitro platforms such as 3D organotypic culture models, engineered microscale tissues and complex microphysiological systems (MPS), together with computational models and computer simulation of tissue dynamics, lend themselves to a integrated testing strategies for predictive toxicology. As these emergent methodologies continue to evolve, they must be integrally tied to maternal/fetal physiology and toxicity of the developing individual across early lifestage transitions, from fertilization to birth, through puberty and beyond. Scope: This symposium will focus on how the novel technology platforms can help now and in the future, with in vitro/in silico modeling of complex biological systems for developmental and reproductive toxicity issues, and translating systems models into integrative testing strategies. The symposium is based on three main organizing principles: (1) that novel in vitro platforms with human cells configured in nascent tissue architectures with a native microphysiological environments yield mechanistic understanding of developmental and reproductive impacts of drug/chemical exposures; (2) that novel in silico platforms with high-throughput screening (HTS) data, biologically-inspired computational models of
Elucidating bonding preferences in tetrakis(imido)uranate(VI) dianions
NASA Astrophysics Data System (ADS)
Anderson, Nickolas H.; Xie, Jing; Ray, Debmalya; Zeller, Matthias; Gagliardi, Laura; Bart, Suzanne C.
2017-09-01
Actinyl species, [AnO2]2+, are well-known derivatives of the f-block because of their natural occurrence and essential roles in the nuclear fuel cycle. Along with their nitrogen analogues, [An(NR)2]2+, actinyls are characterized by their two strong trans-An-element multiple bonds, a consequence of the inverse trans influence. We report that these robust bonds can be weakened significantly by increasing the number of multiple bonds to uranium, as demonstrated by a family of uranium(VI) dianions bearing four U-N multiple bonds, [M]2[U(NR)4] (M = Li, Na, K, Rb, Cs). Their geometry is dictated by cation coordination and sterics rather than by electronic factors. Multiple bond weakening by the addition of strong π donors has the potential for applications in the processing of high-valent actinyls, commonly found in environmental pollutants and spent nuclear fuels.
Krylov Subspace Methods for Complex Non-Hermitian Linear Systems. Thesis
NASA Technical Reports Server (NTRS)
Freund, Roland W.
1991-01-01
We consider Krylov subspace methods for the solution of large sparse linear systems Ax = b with complex non-Hermitian coefficient matrices. Such linear systems arise in important applications, such as inverse scattering, numerical solution of time-dependent Schrodinger equations, underwater acoustics, eddy current computations, numerical computations in quantum chromodynamics, and numerical conformal mapping. Typically, the resulting coefficient matrices A exhibit special structures, such as complex symmetry, or they are shifted Hermitian matrices. In this paper, we first describe a Krylov subspace approach with iterates defined by a quasi-minimal residual property, the QMR method, for solving general complex non-Hermitian linear systems. Then, we study special Krylov subspace methods designed for the two families of complex symmetric respectively shifted Hermitian linear systems. We also include some results concerning the obvious approach to general complex linear systems by solving equivalent real linear systems for the real and imaginary parts of x. Finally, numerical experiments for linear systems arising from the complex Helmholtz equation are reported.
CSM Testbed Development and Large-Scale Structural Applications
NASA Technical Reports Server (NTRS)
Knight, Norman F., Jr.; Gillian, R. E.; Mccleary, Susan L.; Lotts, C. G.; Poole, E. L.; Overman, A. L.; Macy, S. C.
1989-01-01
A research activity called Computational Structural Mechanics (CSM) conducted at the NASA Langley Research Center is described. This activity is developing advanced structural analysis and computational methods that exploit high-performance computers. Methods are developed in the framework of the CSM Testbed software system and applied to representative complex structural analysis problems from the aerospace industry. An overview of the CSM Testbed methods development environment is presented and some new numerical methods developed on a CRAY-2 are described. Selected application studies performed on the NAS CRAY-2 are also summarized.
NASA Technical Reports Server (NTRS)
Shih, T. I.-P.; Bailey, R. T.; Nguyen, H. L.; Roelke, R. J.
1990-01-01
An efficient computer program, called GRID2D/3D was developed to generate single and composite grid systems within geometrically complex two- and three-dimensional (2- and 3-D) spatial domains that can deform with time. GRID2D/3D generates single grid systems by using algebraic grid generation methods based on transfinite interpolation in which the distribution of grid points within the spatial domain is controlled by stretching functions. All single grid systems generated by GRID2D/3D can have grid lines that are continuous and differentiable everywhere up to the second-order. Also, grid lines can intersect boundaries of the spatial domain orthogonally. GRID2D/3D generates composite grid systems by patching together two or more single grid systems. The patching can be discontinuous or continuous. For continuous composite grid systems, the grid lines are continuous and differentiable everywhere up to the second-order except at interfaces where different single grid systems meet. At interfaces where different single grid systems meet, the grid lines are only differentiable up to the first-order. For 2-D spatial domains, the boundary curves are described by using either cubic or tension spline interpolation. For 3-D spatial domains, the boundary surfaces are described by using either linear Coon's interpolation, bi-hyperbolic spline interpolation, or a new technique referred to as 3-D bi-directional Hermite interpolation. Since grid systems generated by algebraic methods can have grid lines that overlap one another, GRID2D/3D contains a graphics package for evaluating the grid systems generated. With the graphics package, the user can generate grid systems in an interactive manner with the grid generation part of GRID2D/3D. GRID2D/3D is written in FORTRAN 77 and can be run on any IBM PC, XT, or AT compatible computer. In order to use GRID2D/3D on workstations or mainframe computers, some minor modifications must be made in the graphics part of the program; no modifications are needed in the grid generation part of the program. This technical memorandum describes the theory and method used in GRID2D/3D.
Queueing Network Models for Parallel Processing of Task Systems: an Operational Approach
NASA Technical Reports Server (NTRS)
Mak, Victor W. K.
1986-01-01
Computer performance modeling of possibly complex computations running on highly concurrent systems is considered. Earlier works in this area either dealt with a very simple program structure or resulted in methods with exponential complexity. An efficient procedure is developed to compute the performance measures for series-parallel-reducible task systems using queueing network models. The procedure is based on the concept of hierarchical decomposition and a new operational approach. Numerical results for three test cases are presented and compared to those of simulations.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mishra, P.; Purdue University, West Lafayette, Indiana 47907; Verma, K.
Borazine is isoelectronic with benzene and is popularly referred to as inorganic benzene. The study of non-covalent interactions with borazine and comparison with its organic counterpart promises to show interesting similarities and differences. The motivation of the present study of the borazine-water interaction, for the first time, stems from such interesting possibilities. Hydrogen-bonded complexes of borazine and water were studied using matrix isolation infrared spectroscopy and quantum chemical calculations. Computations were performed at M06-2X and MP2 levels of theory using 6-311++G(d,p) and aug-cc-pVDZ basis sets. At both the levels of theory, the complex involving an N–H⋯O interaction, where the N–Hmore » of borazine serves as the proton donor to the oxygen of water was found to be the global minimum, in contrast to the benzene-water system, which showed an H–π interaction. The experimentally observed infrared spectra of the complexes corroborated well with our computations for the complex corresponding to the global minimum. In addition to the global minimum, our computations also located two local minima on the borazine-water potential energy surface. Of the two local minima, one corresponded to a structure where the water was the proton donor to the nitrogen of borazine, approaching the borazine ring from above the plane of the ring; a structure that resembled the global minimum in the benzene-water H–π complex. The second local minimum corresponded to an interaction of the oxygen of water with the boron of borazine, which can be termed as the boron bond. Clearly the borazine-water system presents a richer landscape than the benzene-water system.« less
A Low Complexity System Based on Multiple Weighted Decision Trees for Indoor Localization
Sánchez-Rodríguez, David; Hernández-Morera, Pablo; Quinteiro, José Ma.; Alonso-González, Itziar
2015-01-01
Indoor position estimation has become an attractive research topic due to growing interest in location-aware services. Nevertheless, satisfying solutions have not been found with the considerations of both accuracy and system complexity. From the perspective of lightweight mobile devices, they are extremely important characteristics, because both the processor power and energy availability are limited. Hence, an indoor localization system with high computational complexity can cause complete battery drain within a few hours. In our research, we use a data mining technique named boosting to develop a localization system based on multiple weighted decision trees to predict the device location, since it has high accuracy and low computational complexity. The localization system is built using a dataset from sensor fusion, which combines the strength of radio signals from different wireless local area network access points and device orientation information from a digital compass built-in mobile device, so that extra sensors are unnecessary. Experimental results indicate that the proposed system leads to substantial improvements on computational complexity over the widely-used traditional fingerprinting methods, and it has a better accuracy than they have. PMID:26110413
Yu, Xinguang; Li, Lianfeng; Wang, Peng; Yin, Yiheng; Bu, Bo; Zhou, Dingbiao
2014-07-01
This study was designed to report our preliminary experience with stabilization procedures for complex craniovertebral junction malformation (CVJM) using intraoperative computed tomography (iCT) with an integrated neuronavigation system (NNS). To evaluate the workflow, feasibility and clinical outcome of stabilization procedures using iCT image-guided navigation for complex CVJM. The stabilization procedures in CVJM are complex because of the area's intricate geometry and bony structures, its critical relationship to neurovascular structures and the intricate biomechanical issues involved. A sliding gantry 40-slice computed tomography scanner was installed in a preexisting operating room. The images were transferred directly from the scanner to the NNS using an automated registration system. On the basis of the analysis of intraoperative computed tomographic images, 23 cases (11 males, 12 females) with complicated CVJM underwent navigated stabilization procedures to allow more control over screw placement. The age of these patients were 19-52 years (mean: 33.5 y). We performed C1-C2 transarticular screw fixation in 6 patients to produce atlantoaxial arthrodesis with better reliability. Because of a high-riding transverse foramen on at least 1 side of the C2 vertebra and an anomalous vertebral artery position, 7 patients underwent C1 lateral mass and C2 pedicle screw fixation. Ten additional patients were treated with individualized occipitocervical fixation surgery from the hypoplasia of C1 or constraints due to C2 bone structure. In total, 108 screws were inserted into 23 patients using navigational assistance. The screws comprised 20 C1 lateral mass screws, 26 C2, 14 C3, or 4 C4 pedicle screws, 32 occipital screws, and 12 C1-C2 transarticular screws. There were no vascular or neural complications except for pedicle perforations that were detected in 2 (1.9%) patients and were corrected intraoperatively without any persistent nerves or vessel damage. The overall accuracy of the image guidance system was 98.1%. The duration of interruption during the surgical process for the iCT was 8±1.5 minutes. All patients were clinically evaluated using Nurick grade criteria and for neurological deficits 3 months after surgery. Twenty-one patients (91.3%) improved by at least 1 Nurick grade, whereas the grade remained unchanged in 2 (8.7%) patients. Craniovertebral stability and solid bone fusion was achieved in all patients. NNS was found to correlate well with the intraoperative findings, and the recalibration was uneventful in all cases and had an accuracy of 1.8 mm (range, 0.6-2.2 mm). iCT scanning with integrated NNS was found to be both feasible and beneficial in the stabilization procedures for complex CVJM. In this unusual patient population, the technique seemed to be of value for negotiating complex anatomy and for achieving more control over screw placement.
Analysis of Software Systems for Specialized Computers,
computer) with given computer hardware and software . The object of study is the software system of a computer, designed for solving a fixed complex of...purpose of the analysis is to find parameters that characterize the system and its elements during operation, i.e., when servicing the given requirement flow. (Author)
The BioIntelligence Framework: a new computational platform for biomedical knowledge computing.
Farley, Toni; Kiefer, Jeff; Lee, Preston; Von Hoff, Daniel; Trent, Jeffrey M; Colbourn, Charles; Mousses, Spyro
2013-01-01
Breakthroughs in molecular profiling technologies are enabling a new data-intensive approach to biomedical research, with the potential to revolutionize how we study, manage, and treat complex diseases. The next great challenge for clinical applications of these innovations will be to create scalable computational solutions for intelligently linking complex biomedical patient data to clinically actionable knowledge. Traditional database management systems (DBMS) are not well suited to representing complex syntactic and semantic relationships in unstructured biomedical information, introducing barriers to realizing such solutions. We propose a scalable computational framework for addressing this need, which leverages a hypergraph-based data model and query language that may be better suited for representing complex multi-lateral, multi-scalar, and multi-dimensional relationships. We also discuss how this framework can be used to create rapid learning knowledge base systems to intelligently capture and relate complex patient data to biomedical knowledge in order to automate the recovery of clinically actionable information.
Shrestha, Kushal; Jakubikova, Elena
2015-08-20
Light-harvesting antennas are protein-pigment complexes that play a crucial role in natural photosynthesis. The antenna complexes absorb light and transfer energy to photosynthetic reaction centers where charge separation occurs. This work focuses on computational studies of the electronic structure of the pigment networks of light-harvesting complex I (LH1), LH1 with the reaction center (RC-LH1), and light-harvesting complex II (LH2) found in purple bacteria. As the pigment networks of LH1, RC-LH1, and LH2 contain thousands of atoms, conventional density functional theory (DFT) and ab initio calculations of these systems are not computationally feasible. Therefore, we utilize DFT in conjunction with the energy-based fragmentation with molecular orbitals method and a semiempirical approach employing the extended Hückel model Hamiltonian to determine the electronic properties of these pigment assemblies. Our calculations provide a deeper understanding of the electronic structure of natural light-harvesting complexes, especially their pigment networks, which could assist in rational design of artificial photosynthetic devices.
Gas-phase nitrosation of ethylene and related events in the C2H4NO+ landscape.
Gerbaux, Pascal; Dechamps, Noemie; Flammang, Robert; Nam, Pham Cam; Nguyen, Minh Tho; Djazi, Fayçal; Berruyer, Florence; Bouchoux, Guy
2008-06-19
The C2H4NO(+) system has been examined by means of quantum chemical calculations using the G2 and G3B3 approaches and tandem mass spectrometry experiments. Theoretical investigation of the C2H4NO(+) potential-energy surface includes 19 stable C2H4NO(+) structures and a large set of their possible interconnections. These computations provide insights for the understanding of the (i) addition of the nitrosonium cation NO(+) to the ethylene molecule, (ii) skeletal rearrangements evidenced in previous experimental studies on comparable systems, and (iii) experimental identification of new C2H4NO(+) structures. It is predicted from computation that gas-phase nitrosation of ethylene may produce C2H4(*)NO(+) adducts, the most stable structure of which is a pi-complex, 1, stabilized by ca. 65 kJ/mol with respect to its separated components. This complex was produced in the gas phase by a transnitrosation process involving as reactant a complex between water and NO(+) (H2O.NO(+)) and the ethylene molecule and fully characterized by collisional experiments. Among the other C 2H 4NO (+) structures predicted by theory to be protected against dissociation or isomerization by significant energy barriers, five were also experimentally identified. These finding include structures CH3CHNO(+) (5), CH 3CNOH (+) ( 8), CH3NHCO(+) (18), CH3NCOH(+) (19), and an ion/neutral complex CH2O...HCNH(+) (12).
Design consideration in constructing high performance embedded Knowledge-Based Systems (KBS)
NASA Technical Reports Server (NTRS)
Dalton, Shelly D.; Daley, Philip C.
1988-01-01
As the hardware trends for artificial intelligence (AI) involve more and more complexity, the process of optimizing the computer system design for a particular problem will also increase in complexity. Space applications of knowledge based systems (KBS) will often require an ability to perform both numerically intensive vector computations and real time symbolic computations. Although parallel machines can theoretically achieve the speeds necessary for most of these problems, if the application itself is not highly parallel, the machine's power cannot be utilized. A scheme is presented which will provide the computer systems engineer with a tool for analyzing machines with various configurations of array, symbolic, scaler, and multiprocessors. High speed networks and interconnections make customized, distributed, intelligent systems feasible for the application of AI in space. The method presented can be used to optimize such AI system configurations and to make comparisons between existing computer systems. It is an open question whether or not, for a given mission requirement, a suitable computer system design can be constructed for any amount of money.
Computational reacting gas dynamics
NASA Technical Reports Server (NTRS)
Lam, S. H.
1993-01-01
In the study of high speed flows at high altitudes, such as that encountered by re-entry spacecrafts, the interaction of chemical reactions and other non-equilibrium processes in the flow field with the gas dynamics is crucial. Generally speaking, problems of this level of complexity must resort to numerical methods for solutions, using sophisticated computational fluid dynamics (CFD) codes. The difficulties introduced by reacting gas dynamics can be classified into three distinct headings: (1) the usually inadequate knowledge of the reaction rate coefficients in the non-equilibrium reaction system; (2) the vastly larger number of unknowns involved in the computation and the expected stiffness of the equations; and (3) the interpretation of the detailed reacting CFD numerical results. The research performed accepts the premise that reacting flows of practical interest in the future will in general be too complex or 'untractable' for traditional analytical developments. The power of modern computers must be exploited. However, instead of focusing solely on the construction of numerical solutions of full-model equations, attention is also directed to the 'derivation' of the simplified model from the given full-model. In other words, the present research aims to utilize computations to do tasks which have traditionally been done by skilled theoreticians: to reduce an originally complex full-model system into an approximate but otherwise equivalent simplified model system. The tacit assumption is that once the appropriate simplified model is derived, the interpretation of the detailed numerical reacting CFD numerical results will become much easier. The approach of the research is called computational singular perturbation (CSP).
Mousa-Pasandi, Mohammad E; Zhuge, Qunbi; Xu, Xian; Osman, Mohamed M; El-Sahn, Ziad A; Chagnon, Mathieu; Plant, David V
2012-07-02
We experimentally investigate the performance of a low-complexity non-iterative phase noise induced inter-carrier interference (ICI) compensation algorithm in reduced-guard-interval dual-polarization coherent-optical orthogonal-frequency-division-multiplexing (RGI-DP-CO-OFDM) transport systems. This interpolation-based ICI compensator estimates the time-domain phase noise samples by a linear interpolation between the CPE estimates of the consecutive OFDM symbols. We experimentally study the performance of this scheme for a 28 Gbaud QPSK RGI-DP-CO-OFDM employing a low cost distributed feedback (DFB) laser. Experimental results using a DFB laser with the linewidth of 2.6 MHz demonstrate 24% and 13% improvement in transmission reach with respect to the conventional equalizer (CE) in presence of weak and strong dispersion-enhanced-phase-noise (DEPN), respectively. A brief analysis of the computational complexity of this scheme in terms of the number of required complex multiplications is provided. This practical approach does not suffer from error propagation while enjoying low computational complexity.
Philip A. Loring; F. Stuart Chapin; S. Craig Gerlach
2008-01-01
Computational thinking (CT) is a way to solve problems and understand complex systems that draws on concepts fundamental to computer science and is well suited to the challenges that face researchers of complex, linked social-ecological systems. This paper explores CT's usefulness to sustainability science through the application of the services-oriented...
Darwin v. 2.0: an interpreted computer language for the biosciences.
Gonnet, G H; Hallett, M T; Korostensky, C; Bernardin, L
2000-02-01
We announce the availability of the second release of Darwin v. 2.0, an interpreted computer language especially tailored to researchers in the biosciences. The system is a general tool applicable to a wide range of problems. This second release improves Darwin version 1.6 in several ways: it now contains (1) a larger set of libraries touching most of the classical problems from computational biology (pairwise alignment, all versus all alignments, tree construction, multiple sequence alignment), (2) an expanded set of general purpose algorithms (search algorithms for discrete problems, matrix decomposition routines, complex/long integer arithmetic operations), (3) an improved language with a cleaner syntax, (4) better on-line help, and (5) a number of fixes to user-reported bugs. Darwin is made available for most operating systems free of char ge from the Computational Biochemistry Research Group (CBRG), reachable at http://chrg.inf.ethz.ch. darwin@inf.ethz.ch
Design of airborne wind turbine and computational fluid dynamics analysis
NASA Astrophysics Data System (ADS)
Anbreen, Faiqa
Wind energy is a promising alternative to the depleting non-renewable sources. The height of the wind turbines becomes a constraint to their efficiency. Airborne wind turbine can reach much higher altitudes and produce higher power due to high wind velocity and energy density. The focus of this thesis is to design a shrouded airborne wind turbine, capable to generate 70 kW to propel a leisure boat with a capacity of 8-10 passengers. The idea of designing an airborne turbine is to take the advantage of higher velocities in the atmosphere. The Solidworks model has been analyzed numerically using Computational Fluid Dynamics (CFD) software StarCCM+. The Unsteady Reynolds Averaged Navier Stokes Simulation (URANS) with K-epsilon turbulence model has been selected, to study the physical properties of the flow, with emphasis on the performance of the turbine and the increase in air velocity at the throat. The analysis has been done using two ambient velocities of 12 m/s and 6 m/s. At 12 m/s inlet velocity, the velocity of air at the turbine has been recorded as 16 m/s. The power generated by the turbine is 61 kW. At inlet velocity of 6 m/s, the velocity of air at turbine increased to 10 m/s. The power generated by turbine is 25 kW.
Improved neutron activation prediction code system development
NASA Technical Reports Server (NTRS)
Saqui, R. M.
1971-01-01
Two integrated neutron activation prediction code systems have been developed by modifying and integrating existing computer programs to perform the necessary computations to determine neutron induced activation gamma ray doses and dose rates in complex geometries. Each of the two systems is comprised of three computational modules. The first program module computes the spatial and energy distribution of the neutron flux from an input source and prepares input data for the second program which performs the reaction rate, decay chain and activation gamma source calculations. A third module then accepts input prepared by the second program to compute the cumulative gamma doses and/or dose rates at specified detector locations in complex, three-dimensional geometries.
Biosphere 2: a prototype project for a permanent and evolving life system for Mars base.
Nelson, M; Allen, J P; Dempster, W F
1992-01-01
As part of the ground-based preparation for creating long-term life systems needed for space habitation and settlement, Space Biospheres Ventures (SBV) is undertaking the Biosphere 2 project near Oracle, Arizona. Biosphere 2, currently under construction, is scheduled to commence its operations in 1991 with a two-year closure period with a crew of eight people. Biosphere 2 is a facility which will be essentialy materially-closed to exchange with the outside environment. It is open to information and energy flow. Biosphere 2 is designed to achieve a complex life-support system by the integration of seven areas or "biomes"--rainforest, savannah, desert, marsh, ocean, intensive agriculture and human habitat. Unique bioregenerative technologies, such as soil bed reactors for air purification, aquatic waste processing systems, real-time analytic systems and complex computer monitoring and control systems are being developed for the Biosphere 2 project. Its operation should afford valuable insight into the functioning of complex life systems necessary for long-term habitation in space. It will serve as an experimental ground-based prototype and testbed for the stable, permanent life systems needed for human exploration of Mars.
Numerical propulsion system simulation
NASA Technical Reports Server (NTRS)
Lytle, John K.; Remaklus, David A.; Nichols, Lester D.
1990-01-01
The cost of implementing new technology in aerospace propulsion systems is becoming prohibitively expensive. One of the major contributors to the high cost is the need to perform many large scale system tests. Extensive testing is used to capture the complex interactions among the multiple disciplines and the multiple components inherent in complex systems. The objective of the Numerical Propulsion System Simulation (NPSS) is to provide insight into these complex interactions through computational simulations. This will allow for comprehensive evaluation of new concepts early in the design phase before a commitment to hardware is made. It will also allow for rapid assessment of field-related problems, particularly in cases where operational problems were encountered during conditions that would be difficult to simulate experimentally. The tremendous progress taking place in computational engineering and the rapid increase in computing power expected through parallel processing make this concept feasible within the near future. However it is critical that the framework for such simulations be put in place now to serve as a focal point for the continued developments in computational engineering and computing hardware and software. The NPSS concept which is described will provide that framework.
NASA Astrophysics Data System (ADS)
Remko, Milan
Ab initio SCF and DFT methods were used to characterize the gas-phase complexes of selected carbonyl and silacarbonyl bases with Li+ , Na+ and Mg2+ . Geometries were optimized at the Hartree-Fock ab initio and Becke 3LYP DFT levels with the 6-31G* basis set. Frequency computations were performed at the RHF/6-31G* level of theory. Interaction energies of the cation-coordinated systems also were determined with the MP2/6-31G* method. The effect of extension of basis set (to the 6-31+ G* basis) on the computed properties of anion-metal cation complexes was investigated. Calculated energies of formation vary as Mg2+ > Li+ > Na+ . The Becke 3LYP DFT binding energies were comparable with those obtained at the correlated MP2 level and are in good agreement with available experimental data.
High fidelity simulation of non-synchronous vibration for aircraft engine fan/compressor
NASA Astrophysics Data System (ADS)
Im, Hong-Sik
The objectives of this research are to develop a high fidelity simulation methodology for turbomachinery aeromechanical problems and to investigate the mechanism of non-synchronous vibration (NSV) of an aircraft engine axial compressor. A fully conservative rotor/stator sliding technique is developed to accurately capture the unsteadiness and interaction between adjacent blade rows. Phase lag boundary conditions (BC) based on the time shift (direct store) method and the Fourier series phase lag BC are implemented to take into account the effect of phase difference for a sector of annulus simulation. To resolve the nonlinear interaction between flow and vibrating blade structure, a fully coupled fluid-structure interaction (FSI) procedure that solves the structural modal equations and time accurate Navier-Stokes equations simultaneously is adopted. An advanced mesh deformation method that generates the blade tip block mesh moving with the blade displacement is developed to ensure the mesh quality. An efficient and low diffusion E-CUSP (LDE) scheme as a Riemann solver designed to minimize numerical dissipation is used with an improved hybrid RANS/LES turbulence strategy, delayed detached eddy simulation (DDES). High order accuracy (3rd and 5th order) weighted essentially non-oscillatory (WENO) schemes for inviscid flux and a conservative 2nd and 4th order viscous flux differencing are employed. Extensive validations are conducted to demonstrate high accuracy and robustness of the high fidelity FSI simulation methodology. The validated cases include: (1) DDES of NACA 0012 airfoil at high angle of attack with massive separation. The DDES accurately predicts the drag whereas the URANS model significantly over predicts the drag. (2) The AGARD Wing 445.6 flutter boundary is accurately predicted including the point at supersonic incoming flow. (3) NASA Rotor 67 validation for steady state speed line and radial profiles at peak efficiency point and near stall point. The calculated results agree excellently with the experiment. (4) NASA Stage 35 speed line and radial profiles to validate the steady state mixing plane BC for multistage computation. Excellent agreement is obtained between the computation and experiment. (5) NASA Rotor 67 full annulus and single passage FSI simulation at near peak condition to validate phase lag BC. The time shifted phase lag BC accurately predicts blade vibration responses that agrees better with the full annulus FSI simulation. The DDES methodology is used to investigate the stall inception of NASA Rotor 67. The stall process begins with spike inception and develops to full stall. The whole process is simulated with full annulus of the rotor. The fully coupled FSI is then used to simulate the stall flutter of NASA Rotor 67. The multistage simulations of a GE aircraft engine high pressure compressor (HPC) reveal for the first time that the travelling tornado vortex formed on the rotor blade tip region is the root cause for the NSV of the compressor. The rotor blades under NSV have large torsional vibration due to the tornado vortex propagation in the opposite to the rotor rotation. The tornado vortex frequency passing the suction surface of each blade in the tip region agrees with the NSV frequency. The predicted NSV frequency based on URANS model with rigid blades agrees very well with the experimental measurement with only 3.3% under-predicted. The NSV prediction using FSI with vibrating blades also obtain the same frequency as the rigid blades. This is because that the NSV is primarily caused by the flow vortex instability and the no resonance occurs. The blade structures respond passively and the small amplitudes of the blade vibration do not have significant effect on the flow. The predicted frequency using DDES with rigid blades is more deviated from the experiment and is 14.7% lower. The reason is that the DDES tends to predict the rotor stall earlier than the URANS and the NSV can be achieved only at higher mass flow rate, which generates a lower frequency. The possible reason for the DDES to predict the rotor stall early may be because DDES is more sensitive to wave reflection and a non-reflective boundary condition may be necessary. Overall, the high fidelity FSI methodology developed in this thesis for aircraft engine fan/compressor aeromechanics simulation is demonstrated to be very successful and has advanced the forefront of the state of the art. Future work to continue to improve the accuracy and efficiency is discussed at the end of the thesis.
NASA Technical Reports Server (NTRS)
Horvitz, Eric; Ruokangas, Corinne; Srinivas, Sampath; Barry, Matthew
1993-01-01
We describe a collaborative research and development effort between the Palo Alto Laboratory of the Rockwell Science Center, Rockwell Space Operations Company, and the Propulsion Systems Section of NASA JSC to design computational tools that can manage the complexity of information displayed to human operators in high-stakes, time-critical decision contexts. We shall review an application from NASA Mission Control and describe how we integrated a probabilistic diagnostic model and a time-dependent utility model, with techniques for managing the complexity of computer displays. Then, we shall describe the behavior of VPROP, a system constructed to demonstrate promising display-management techniques. Finally, we shall describe our current research directions on the Vista 2 follow-on project.
Gray: a ray tracing-based Monte Carlo simulator for PET
NASA Astrophysics Data System (ADS)
Freese, David L.; Olcott, Peter D.; Buss, Samuel R.; Levin, Craig S.
2018-05-01
Monte Carlo simulation software plays a critical role in PET system design. Performing complex, repeated Monte Carlo simulations can be computationally prohibitive, as even a single simulation can require a large amount of time and a computing cluster to complete. Here we introduce Gray, a Monte Carlo simulation software for PET systems. Gray exploits ray tracing methods used in the computer graphics community to greatly accelerate simulations of PET systems with complex geometries. We demonstrate the implementation of models for positron range, annihilation acolinearity, photoelectric absorption, Compton scatter, and Rayleigh scatter. For validation, we simulate the GATE PET benchmark, and compare energy, distribution of hits, coincidences, and run time. We show a speedup using Gray, compared to GATE for the same simulation, while demonstrating nearly identical results. We additionally simulate the Siemens Biograph mCT system with both the NEMA NU-2 scatter phantom and sensitivity phantom. We estimate the total sensitivity within % when accounting for differences in peak NECR. We also estimate the peak NECR to be kcps, or within % of published experimental data. The activity concentration of the peak is also estimated within 1.3%.
Wind Tunnel Interference Effects on Tilt Rotor Testing Using Computational Fluid Dynamics
NASA Technical Reports Server (NTRS)
Koning, Witold J. F.
2016-01-01
Experimental techniques to measure rotorcraft aerodynamic performance are widely used. However, most of them are either unable to capture interference effects from bodies, or require an extremely large computational budget. The objective of the present research is to develop an XV-15 Tiltrotor Research Aircraft rotor model for investigation of wind tunnel wall interference using a novel Computational Fluid Dynamics (CFD) solver for rotorcraft, RotCFD. In RotCFD, a mid-fidelity Unsteady Reynolds Averaged Navier-Stokes (URANS) solver is used with an incompressible flow model and a realizable k-e turbulence model. The rotor is, however, not modeled using a computationally expensive, unsteady viscous body-fitted grid, but is instead modeled using a blade-element model (BEM) with a momentum source approach. Various flight modes of the XV-15 isolated rotor, including hover, tilt, and airplane mode, have been simulated and correlated to existing experimental and theoretical data. The rotor model is subsequently used for wind tunnel wall interference simulations in the National Full-Scale Aerodynamics Complex (NFAC) at Ames Research Center in California. The results from the validation of the isolated rotor performance showed good correlation with experimental and theoretical data. The results were on par with known theoretical analyses. In RotCFD the setup, grid generation, and running of cases is faster than many CFD codes, which makes it a useful engineering tool. Performance predictions need not be as accurate as high-fidelity CFD codes, as long as wall effects can be properly simulated. For both test sections of the NFAC wall, interference was examined by simulating the XV-15 rotor in the test section of the wind tunnel and with an identical grid but extended boundaries in free field. Both cases were also examined with an isolated rotor or with the rotor mounted on the modeled geometry of the Tiltrotor Test Rig (TTR). A "quasi linear trim" was used to trim the thrust for the rotor to compare the power as a unique variable. Power differences between free field and wind tunnel cases were found from -7 to 0 percent in the 80- by 120-Foot Wind Tunnel and -1.6 to 4.8 percent in the 40- by 80-Foot Wind Tunnel, depending on the TTR orientation, tunnel velocity, and blade setting. The TTR will be used in 2016 to test the Bell 609 rotor in a similar fashion to the research in this report.
Wind Tunnel Interference Effects on Tilt Rotor Testing Using Computational Fluid Dynamics
NASA Technical Reports Server (NTRS)
Koning, Witold J. F.
2015-01-01
Experimental techniques to measure rotorcraft aerodynamic performance are widely used. However, most of them are either unable to capture interference effects from bodies, or require an extremely large computational budget. The objective of the present research is to develop an XV-15 Tilt Rotor Research Aircraft rotor model for investigation of wind tunnel wall interference using a novel Computational Fluid Dynamics (CFD) solver for rotorcraft, RotCFD. In RotCFD, a mid-fidelity URANS solver is used with an incompressible flow model and a realizable k-e turbulence model. The rotor is, however, not modeled using a computationally expensive, unsteady viscous body-fitted grid, but is instead modeled using a blade element model with a momentum source approach. Various flight modes of the XV-15 isolated rotor, including hover, tilt and airplane mode, have been simulated and correlated to existing experimental and theoretical data. The rotor model is subsequently used for wind tunnel wall interference simulations in the National Full-Scale Aerodynamics Complex (NFAC) at NASA Ames Research Center in California. The results from the validation of the isolated rotor performance showed good correlation with experimental and theoretical data. The results were on par with known theoretical analyses. In RotCFD the setup, grid generation and running of cases is faster than many CFD codes, which makes it a useful engineering tool. Performance predictions need not be as accurate as high-fidelity CFD codes, as long as wall effects can be properly simulated. For both test sections of the NFAC wall interference was examined by simulating the XV-15 rotor in the test section of the wind tunnel and with an identical grid but extended boundaries in free field. Both cases were also examined with an isolated rotor or with the rotor mounted on the modeled geometry of the Tiltrotor Test Rig (TTR). A 'quasi linear trim' was used to trim the thrust for the rotor to compare the power as a unique variable. Power differences between free field and wind tunnel cases were found from -7 % to 0 % in the 80- by 120-Foot Wind Tunnel test section and -1.6 % to 4.8 % in the 40- by 80-Foot Wind Tunnel, depending on the TTR orientation, tunnel velocity and blade setting. The TTR will be used in 2016 to test the Bell 609 rotor in a similar fashion to the research in this report.
The engineering design integration (EDIN) system. [digital computer program complex
NASA Technical Reports Server (NTRS)
Glatt, C. R.; Hirsch, G. N.; Alford, G. E.; Colquitt, W. N.; Reiners, S. J.
1974-01-01
A digital computer program complex for the evaluation of aerospace vehicle preliminary designs is described. The system consists of a Univac 1100 series computer and peripherals using the Exec 8 operating system, a set of demand access terminals of the alphanumeric and graphics types, and a library of independent computer programs. Modification of the partial run streams, data base maintenance and construction, and control of program sequencing are provided by a data manipulation program called the DLG processor. The executive control of library program execution is performed by the Univac Exec 8 operating system through a user established run stream. A combination of demand and batch operations is employed in the evaluation of preliminary designs. Applications accomplished with the EDIN system are described.
LES Investigation of Wake Development in a Transonic Fan Stage for Aeroacoustic Analysis
NASA Technical Reports Server (NTRS)
Hah, Chunill; Romeo, Michael
2017-01-01
Detailed development of the rotor wake and its interaction with the stator are investigated with a large eddy simulation (LES). Typical steady and unsteady Navier-Stokes approaches (RANS and URANS) do not calculate wake development accurately and do not provide all the necessary information for an aeroacoustic analysis. It is generally believed that higher fidelity analysis tools are required for an aeroacoustic investigation of transonic fan stages.
[Soft- and hardware support for the setup for computer tracking of radiation teletherapy].
Tarutin, I G; Piliavets, V I; Strakh, A G; Minenko, V F; Golubovskiĭ, A I
1983-06-01
A hard and soft ware computer assisted complex has been worked out for gamma-beam therapy. The complex included all radiotherapeutic units, including a Siemens program controlled betatron with an energy of 42 MEV computer ES-1022, a Medigraf system of the processing of graphic information, a Mars-256 system for control over the homogeneity of distribution of dose rate on the field of irradiation and a package of mathematical programs to select a plan of irradiation of various tumor sites. The prospects of the utilization of such complexes in the dosimetric support of radiation therapy are discussed.
ERIC Educational Resources Information Center
Kiraz, George Anton
This book presents a tractable computational model that can cope with complex morphological operations, especially in Semitic languages, and less complex morphological systems present in Western languages. It outlines a new generalized regular rewrite rule system that uses multiple finite-state automata to cater to root-and-pattern morphology,…
ERIC Educational Resources Information Center
Marek, Michael W.; Wu, Wen-Chi Vivian
2014-01-01
This conceptual, interdisciplinary inquiry explores Complex Dynamic Systems as the concept relates to the internal and external environmental factors affecting computer assisted language learning (CALL). Based on the results obtained by de Rosnay ["World Futures: The Journal of General Evolution", 67(4/5), 304-315 (2011)], who observed…
ERIC Educational Resources Information Center
Loke, Swee-Kin; Al-Sallami, Hesham S.; Wright, Daniel F. B.; McDonald, Jenny; Jadhav, Sheetal; Duffull, Stephen B.
2012-01-01
Complex systems are typically difficult for students to understand and computer simulations offer a promising way forward. However, integrating such simulations into conventional classes presents numerous challenges. Framed within an educational design research, we studied the use of an in-house built simulation of the coagulation network in four…
Biocellion: accelerating computer simulation of multicellular biological system models
Kang, Seunghwa; Kahan, Simon; McDermott, Jason; Flann, Nicholas; Shmulevich, Ilya
2014-01-01
Motivation: Biological system behaviors are often the outcome of complex interactions among a large number of cells and their biotic and abiotic environment. Computational biologists attempt to understand, predict and manipulate biological system behavior through mathematical modeling and computer simulation. Discrete agent-based modeling (in combination with high-resolution grids to model the extracellular environment) is a popular approach for building biological system models. However, the computational complexity of this approach forces computational biologists to resort to coarser resolution approaches to simulate large biological systems. High-performance parallel computers have the potential to address the computing challenge, but writing efficient software for parallel computers is difficult and time-consuming. Results: We have developed Biocellion, a high-performance software framework, to solve this computing challenge using parallel computers. To support a wide range of multicellular biological system models, Biocellion asks users to provide their model specifics by filling the function body of pre-defined model routines. Using Biocellion, modelers without parallel computing expertise can efficiently exploit parallel computers with less effort than writing sequential programs from scratch. We simulate cell sorting, microbial patterning and a bacterial system in soil aggregate as case studies. Availability and implementation: Biocellion runs on x86 compatible systems with the 64 bit Linux operating system and is freely available for academic use. Visit http://biocellion.com for additional information. Contact: seunghwa.kang@pnnl.gov PMID:25064572
DOE Office of Scientific and Technical Information (OSTI.GOV)
Alí-Torres, Jorge; Mirats, Andrea; Maréchal, Jean-Didier
Amyloid plaques formation and oxidative stress are two key events in the pathology of the Alzheimer disease (AD), in which metal cations have been shown to play an important role. In particular, the interaction of the redox active Cu{sup 2+} metal cation with Aβ has been found to interfere in amyloid aggregation and to lead to reactive oxygen species (ROS). A detailed knowledge of the electronic and molecular structure of Cu{sup 2+}-Aβ complexes is thus important to get a better understanding of the role of these complexes in the development and progression of the AD disease. The computational treatment ofmore » these systems requires a combination of several available computational methodologies, because two fundamental aspects have to be addressed: the metal coordination sphere and the conformation adopted by the peptide upon copper binding. In this paper we review the main computational strategies used to deal with the Cu{sup 2+}-Aβ coordination and build plausible Cu{sup 2+}-Aβ models that will afterwards allow determining physicochemical properties of interest, such as their redox potential.« less
Complex network problems in physics, computer science and biology
NASA Astrophysics Data System (ADS)
Cojocaru, Radu Ionut
There is a close relation between physics and mathematics and the exchange of ideas between these two sciences are well established. However until few years ago there was no such a close relation between physics and computer science. Even more, only recently biologists started to use methods and tools from statistical physics in order to study the behavior of complex system. In this thesis we concentrate on applying and analyzing several methods borrowed from computer science to biology and also we use methods from statistical physics in solving hard problems from computer science. In recent years physicists have been interested in studying the behavior of complex networks. Physics is an experimental science in which theoretical predictions are compared to experiments. In this definition, the term prediction plays a very important role: although the system is complex, it is still possible to get predictions for its behavior, but these predictions are of a probabilistic nature. Spin glasses, lattice gases or the Potts model are a few examples of complex systems in physics. Spin glasses and many frustrated antiferromagnets map exactly to computer science problems in the NP-hard class defined in Chapter 1. In Chapter 1 we discuss a common result from artificial intelligence (AI) which shows that there are some problems which are NP-complete, with the implication that these problems are difficult to solve. We introduce a few well known hard problems from computer science (Satisfiability, Coloring, Vertex Cover together with Maximum Independent Set and Number Partitioning) and then discuss their mapping to problems from physics. In Chapter 2 we provide a short review of combinatorial optimization algorithms and their applications to ground state problems in disordered systems. We discuss the cavity method initially developed for studying the Sherrington-Kirkpatrick model of spin glasses. We extend this model to the study of a specific case of spin glass on the Bethe lattice at zero temperature and then we apply this formalism to the K-SAT problem defined in Chapter 1. The phase transition which physicists study often corresponds to a change in the computational complexity of the corresponding computer science problem. Chapter 3 presents phase transitions which are specific to the problems discussed in Chapter 1 and also known results for the K-SAT problem. We discuss the replica method and experimental evidences of replica symmetry breaking. The physics approach to hard problems is based on replica methods which are difficult to understand. In Chapter 4 we develop novel methods for studying hard problems using methods similar to the message passing techniques that were discussed in Chapter 2. Although we concentrated on the symmetric case, cavity methods show promise for generalizing our methods to the un-symmetric case. As has been highlighted by John Hopfield, several key features of biological systems are not shared by physical systems. Although living entities follow the laws of physics and chemistry, the fact that organisms adapt and reproduce introduces an essential ingredient that is missing in the physical sciences. In order to extract information from networks many algorithm have been developed. In Chapter 5 we apply polynomial algorithms like minimum spanning tree in order to study and construct gene regulatory networks from experimental data. As future work we propose the use of algorithms like min-cut/max-flow and Dijkstra for understanding key properties of these networks.
Efficient two-dimensional compressive sensing in MIMO radar
NASA Astrophysics Data System (ADS)
Shahbazi, Nafiseh; Abbasfar, Aliazam; Jabbarian-Jahromi, Mohammad
2017-12-01
Compressive sensing (CS) has been a way to lower sampling rate leading to data reduction for processing in multiple-input multiple-output (MIMO) radar systems. In this paper, we further reduce the computational complexity of a pulse-Doppler collocated MIMO radar by introducing a two-dimensional (2D) compressive sensing. To do so, we first introduce a new 2D formulation for the compressed received signals and then we propose a new measurement matrix design for our 2D compressive sensing model that is based on minimizing the coherence of sensing matrix using gradient descent algorithm. The simulation results show that our proposed 2D measurement matrix design using gradient decent algorithm (2D-MMDGD) has much lower computational complexity compared to one-dimensional (1D) methods while having better performance in comparison with conventional methods such as Gaussian random measurement matrix.
Wu, Naiqi; Zhou, MengChu
2005-12-01
An automated manufacturing system (AMS) contains a number of versatile machines (or workstations), buffers, an automated material handling system (MHS), and is computer-controlled. An effective and flexible alternative for implementing MHS is to use automated guided vehicle (AGV) system. The deadlock issue in AMS is very important in its operation and has extensively been studied. The deadlock problems were separately treated for parts in production and transportation and many techniques were developed for each problem. However, such treatment does not take the advantage of the flexibility offered by multiple AGVs. In general, it is intractable to obtain maximally permissive control policy for either problem. Instead, this paper investigates these two problems in an integrated way. First we model an AGV system and part processing processes by resource-oriented Petri nets, respectively. Then the two models are integrated by using macro transitions. Based on the combined model, a novel control policy for deadlock avoidance is proposed. It is shown to be maximally permissive with computational complexity of O (n2) where n is the number of machines in AMS if the complexity for controlling the part transportation by AGVs is not considered. Thus, the complexity of deadlock avoidance for the whole system is bounded by the complexity in controlling the AGV system. An illustrative example shows its application and power.
NASA Technical Reports Server (NTRS)
Wolpert, David H.; Koga, Dennis (Technical Monitor)
2000-01-01
In this first of two papers, strong limits on the accuracy of physical computation are established. First it is proven that there cannot be a physical computer C to which one can pose any and all computational tasks concerning the physical universe. Next it is proven that no physical computer C can correctly carry out any computational task in the subset of such tasks that can be posed to C. This result holds whether the computational tasks concern a system that is physically isolated from C, or instead concern a system that is coupled to C. As a particular example, this result means that there cannot be a physical computer that can, for any physical system external to that computer, take the specification of that external system's state as input and then correctly predict its future state before that future state actually occurs; one cannot build a physical computer that can be assured of correctly 'processing information faster than the universe does'. The results also mean that there cannot exist an infallible, general-purpose observation apparatus, and that there cannot be an infallible, general-purpose control apparatus. These results do not rely on systems that are infinite, and/or non-classical, and/or obey chaotic dynamics. They also hold even if one uses an infinitely fast, infinitely dense computer, with computational powers greater than that of a Turing Machine. This generality is a direct consequence of the fact that a novel definition of computation - a definition of 'physical computation' - is needed to address the issues considered in these papers. While this definition does not fit into the traditional Chomsky hierarchy, the mathematical structure and impossibility results associated with it have parallels in the mathematics of the Chomsky hierarchy. The second in this pair of papers presents a preliminary exploration of some of this mathematical structure, including in particular that of prediction complexity, which is a 'physical computation analogue' of algorithmic information complexity. It is proven in that second paper that either the Hamiltonian of our universe proscribes a certain type of computation, or prediction complexity is unique (unlike algorithmic information complexity), in that there is one and only version of it that can be applicable throughout our universe.
NASA Technical Reports Server (NTRS)
Malin, Jane T.; Schreckenghost, Debra L.; Woods, David D.; Potter, Scott S.; Johannesen, Leila; Holloway, Matthew; Forbus, Kenneth D.
1991-01-01
Initial results are reported from a multi-year, interdisciplinary effort to provide guidance and assistance for designers of intelligent systems and their user interfaces. The objective is to achieve more effective human-computer interaction (HCI) for systems with real time fault management capabilities. Intelligent fault management systems within the NASA were evaluated for insight into the design of systems with complex HCI. Preliminary results include: (1) a description of real time fault management in aerospace domains; (2) recommendations and examples for improving intelligent systems design and user interface design; (3) identification of issues requiring further research; and (4) recommendations for a development methodology integrating HCI design into intelligent system design.
Space station Simulation Computer System (SCS) study for NASA/MSFC. Volume 2: Concept document
NASA Technical Reports Server (NTRS)
1989-01-01
The Simulation Computer System (SCS) concept document describes and establishes requirements for the functional performance of the SCS system, including interface, logistic, and qualification requirements. The SCS is the computational communications and display segment of the Marshall Space Flight Center (MSFC) Payload Training Complex (PTC). The PTC is the MSFC facility that will train onboard and ground operations personnel to operate the payloads and experiments on board the international Space Station Freedom. The requirements to be satisfied by the system implementation are identified here. The SCS concept document defines the requirements to be satisfied through the implementation of the system capability. The information provides the operational basis for defining the requirements to be allocated to the system components and enables the system organization to assess whether or not the completed system complies with the requirements of the system.
Biologically inspired collision avoidance system for unmanned vehicles
NASA Astrophysics Data System (ADS)
Ortiz, Fernando E.; Graham, Brett; Spagnoli, Kyle; Kelmelis, Eric J.
2009-05-01
In this project, we collaborate with researchers in the neuroscience department at the University of Delaware to develop an Field Programmable Gate Array (FPGA)-based embedded computer, inspired by the brains of small vertebrates (fish). The mechanisms of object detection and avoidance in fish have been extensively studied by our Delaware collaborators. The midbrain optic tectum is a biological multimodal navigation controller capable of processing input from all senses that convey spatial information, including vision, audition, touch, and lateral-line (water current sensing in fish). Unfortunately, computational complexity makes these models too slow for use in real-time applications. These simulations are run offline on state-of-the-art desktop computers, presenting a gap between the application and the target platform: a low-power embedded device. EM Photonics has expertise in developing of high-performance computers based on commodity platforms such as graphic cards (GPUs) and FPGAs. FPGAs offer (1) high computational power, low power consumption and small footprint (in line with typical autonomous vehicle constraints), and (2) the ability to implement massively-parallel computational architectures, which can be leveraged to closely emulate biological systems. Combining UD's brain modeling algorithms and the power of FPGAs, this computer enables autonomous navigation in complex environments, and further types of onboard neural processing in future applications.
Integrating GIS and ABM to Explore Spatiotemporal Dynamics
NASA Astrophysics Data System (ADS)
Sun, M.; Jiang, Y.; Yang, C.
2013-12-01
Agent-based modeling as a methodology for the bottom-up exploration with the account of adaptive behavior and heterogeneity of system components can help discover the development and pattern of the complex social and environmental system. However, ABM is a computationally intensive process especially when the number of system components becomes large and the agent-agent/agent-environmental interaction is modeled very complex. Most of traditional ABM frameworks developed based on CPU do not have a satisfying computing capacity. To address the problem and as the emergence of advanced techniques, GPU computing with CUDA can provide powerful parallel structure to enable the complex simulation of spatiotemporal dynamics. In this study, we first develop a GPU-based ABM system. Secondly, in order to visualize the dynamics generated from the movement of agent and the change of agent/environmental attributes during the simulation, we integrate GIS into the ABM system. Advanced geovisualization technologies can be utilized for representing the spatiotemporal change events, such as proper 2D/3D maps with state-of-the-art symbols, space-time cube and multiple layers each of which presents pattern in one time-stamp, etc. Thirdly, visual analytics which include interactive tools (e.g. grouping, filtering, linking, etc.) is included in our ABM-GIS system to help users conduct real-time data exploration during the progress of simulation. Analysis like flow analysis and spatial cluster analysis can be integrated according to the geographical problem we want to explore.
Predicting equilibrium uranium isotope fractionation in crystals and solution
NASA Astrophysics Data System (ADS)
Schauble, E. A.
2015-12-01
Despite the rapidly growing interest in using 238U/235U measurements as a proxy for changes in oxygen abundance in surface and near-surface environments, the present theoretical understanding of uranium isotope fractionation is limited to a few simple gas-phase molecules and analogues of dissolved species (e.g., 1,2,3). Understanding uranium isotope fractionation behavior in more complicated species, such as crystals and adsorption complexes, will help in the design and interpretation of experiments and field studies, and may suggest other uses for 38U/235U measurements. In this study, a recently developed first-principles method for estimating the nuclear volume component of field shift fractionation in crystals and complex molecular species (4) is combined with mass-dependent fractionation theory to predict equilibrium 38U/235U fractionations in aqueous and crystalline uranium compounds, including uraninite (UO2). The nuclear field shift effect, caused by the interaction of electrons with the finite volume of the positive charge distribution in uranium nuclei, is estimated using Density Functional Theory and the Projector Augmented Wave method (DFT-PAW). Tests against relativistic electronic structure calculations and Mössbauer isomer shift data indicate that the DFT-PAW method is reasonably accurate, while being much better suited to models of complex and crystalline species. Initial results confirm previous predictions that the nuclear volume effect overwhelms mass depdendent fractionation in U(VI)-U(IV) exchange reactions, leading to higher 238U/235U in U(IV) species (i.e., for UO2 xtal vs. UO22+aq, ln αNV ≈ +1.8‰ , ln αMD ≈ -0.8‰, ln αTotal ≈ +1.0‰ at 25ºC). UO2 and U(H2O)94+, are within ~0.4‰ of each other, while U(VI) species appear to be more variable. This suggests that speciation is likely to significantly affect natural uranium isotope fractionations, in addition to oxidation state. Tentatively, it appears that uranyl-type (UO22+-bearing) structures will tend to have higher 238U/235U than uranate-type structures that lack strong U=O bonds. References: 1. Bigeleisen (1996) JACS 118:3676; 2. Schauble (2006) Eos 87:V21B-0570; 3. Abe et al. (2008) J Chem Phys 128:144309, 129:164309, & Abe et al. (2010) J Chem Phys 133:044309; 4. Schauble (2013) PNAS 110:17714.
Experimental control of a fluidic pinball using genetic programming
NASA Astrophysics Data System (ADS)
Raibaudo, Cedric; Zhong, Peng; Noack, Bernd R.; Martinuzzi, Robert J.
2017-11-01
The wake stabilization of a triangular cluster of three rotating cylinders was investigated in the present study. Experiments were performed at Reynolds number Re 6000, and compared with URANS-2D simulations at same flow conditions. 2D2C PIV measurements and constant temperature anemometry were used to characterize the flow without and with actuation. Open-loop actuation was first considered for the identification of particular control strategies. Machine learning control was also implemented for the experimental study. Linear genetic programming has been used for the optimization of open-loop parameters and closed-loop controllers. Considering a cost function J based on the fluctuations of the velocity measured by the hot-wire sensor, significant performances were achieved using the machine learning approach. The present work is supported by the senior author's (R. J. Martinuzzi) NSERC discovery Grant. C. Raibaudo acknowledges the financial support of the University of Calgary Eyes-High PDF program.
Remote control system for high-perfomance computer simulation of crystal growth by the PFC method
NASA Astrophysics Data System (ADS)
Pavlyuk, Evgeny; Starodumov, Ilya; Osipov, Sergei
2017-04-01
Modeling of crystallization process by the phase field crystal method (PFC) - one of the important directions of modern computational materials science. In this paper, the practical side of the computer simulation of the crystallization process by the PFC method is investigated. To solve problems using this method, it is necessary to use high-performance computing clusters, data storage systems and other often expensive complex computer systems. Access to such resources is often limited, unstable and accompanied by various administrative problems. In addition, the variety of software and settings of different computing clusters sometimes does not allow researchers to use unified program code. There is a need to adapt the program code for each configuration of the computer complex. The practical experience of the authors has shown that the creation of a special control system for computing with the possibility of remote use can greatly simplify the implementation of simulations and increase the performance of scientific research. In current paper we show the principal idea of such a system and justify its efficiency.
Huang, Wei; Ravikumar, Krishnakumar M; Parisien, Marc; Yang, Sichun
2016-12-01
Structural determination of protein-protein complexes such as multidomain nuclear receptors has been challenging for high-resolution structural techniques. Here, we present a combined use of multiple biophysical methods, termed iSPOT, an integration of shape information from small-angle X-ray scattering (SAXS), protection factors probed by hydroxyl radical footprinting, and a large series of computationally docked conformations from rigid-body or molecular dynamics (MD) simulations. Specifically tested on two model systems, the power of iSPOT is demonstrated to accurately predict the structures of a large protein-protein complex (TGFβ-FKBP12) and a multidomain nuclear receptor homodimer (HNF-4α), based on the structures of individual components of the complexes. Although neither SAXS nor footprinting alone can yield an unambiguous picture for each complex, the combination of both, seamlessly integrated in iSPOT, narrows down the best-fit structures that are about 3.2Å and 4.2Å in RMSD from their corresponding crystal structures, respectively. Furthermore, this proof-of-principle study based on the data synthetically derived from available crystal structures shows that the iSPOT-using either rigid-body or MD-based flexible docking-is capable of overcoming the shortcomings of standalone computational methods, especially for HNF-4α. By taking advantage of the integration of SAXS-based shape information and footprinting-based protection/accessibility as well as computational docking, this iSPOT platform is set to be a powerful approach towards accurate integrated modeling of many challenging multiprotein complexes. Copyright © 2016 Elsevier Inc. All rights reserved.
Computer simulation of functioning of elements of security systems
NASA Astrophysics Data System (ADS)
Godovykh, A. V.; Stepanov, B. P.; Sheveleva, A. A.
2017-01-01
The article is devoted to issues of development of the informational complex for simulation of functioning of the security system elements. The complex is described from the point of view of main objectives, a design concept and an interrelation of main elements. The proposed conception of the computer simulation provides an opportunity to simulate processes of security system work for training security staff during normal and emergency operation.
Synthetic Analog and Digital Circuits for Cellular Computation and Memory
Purcell, Oliver; Lu, Timothy K.
2014-01-01
Biological computation is a major area of focus in synthetic biology because it has the potential to enable a wide range of applications. Synthetic biologists have applied engineering concepts to biological systems in order to construct progressively more complex gene circuits capable of processing information in living cells. Here, we review the current state of computational genetic circuits and describe artificial gene circuits that perform digital and analog computation. We then discuss recent progress in designing gene circuits that exhibit memory, and how memory and computation have been integrated to yield more complex systems that can both process and record information. Finally, we suggest new directions for engineering biological circuits capable of computation. PMID:24794536
The BioIntelligence Framework: a new computational platform for biomedical knowledge computing
Farley, Toni; Kiefer, Jeff; Lee, Preston; Von Hoff, Daniel; Trent, Jeffrey M; Colbourn, Charles
2013-01-01
Breakthroughs in molecular profiling technologies are enabling a new data-intensive approach to biomedical research, with the potential to revolutionize how we study, manage, and treat complex diseases. The next great challenge for clinical applications of these innovations will be to create scalable computational solutions for intelligently linking complex biomedical patient data to clinically actionable knowledge. Traditional database management systems (DBMS) are not well suited to representing complex syntactic and semantic relationships in unstructured biomedical information, introducing barriers to realizing such solutions. We propose a scalable computational framework for addressing this need, which leverages a hypergraph-based data model and query language that may be better suited for representing complex multi-lateral, multi-scalar, and multi-dimensional relationships. We also discuss how this framework can be used to create rapid learning knowledge base systems to intelligently capture and relate complex patient data to biomedical knowledge in order to automate the recovery of clinically actionable information. PMID:22859646
Marr's levels and the minimalist program.
Johnson, Mark
2017-02-01
A simple change to a cognitive system at Marr's computational level may entail complex changes at the other levels of description of the system. The implementational level complexity of a change, rather than its computational level complexity, may be more closely related to the plausibility of a discrete evolutionary event causing that change. Thus the formal complexity of a change at the computational level may not be a good guide to the plausibility of an evolutionary event introducing that change. For example, while the Minimalist Program's Merge is a simple formal operation (Berwick & Chomsky, 2016), the computational mechanisms required to implement the language it generates (e.g., to parse the language) may be considerably more complex. This has implications for the theory of grammar: theories of grammar which involve several kinds of syntactic operations may be no less evolutionarily plausible than a theory of grammar that involves only one. A deeper understanding of human language at the algorithmic and implementational levels could strengthen Minimalist Program's account of the evolution of language.
Geometry of discrete quantum computing
NASA Astrophysics Data System (ADS)
Hanson, Andrew J.; Ortiz, Gerardo; Sabry, Amr; Tai, Yu-Tsung
2013-05-01
Conventional quantum computing entails a geometry based on the description of an n-qubit state using 2n infinite precision complex numbers denoting a vector in a Hilbert space. Such numbers are in general uncomputable using any real-world resources, and, if we have the idea of physical law as some kind of computational algorithm of the universe, we would be compelled to alter our descriptions of physics to be consistent with computable numbers. Our purpose here is to examine the geometric implications of using finite fields Fp and finite complexified fields \\mathbf {F}_{p^2} (based on primes p congruent to 3 (mod4)) as the basis for computations in a theory of discrete quantum computing, which would therefore become a computable theory. Because the states of a discrete n-qubit system are in principle enumerable, we are able to determine the proportions of entangled and unentangled states. In particular, we extend the Hopf fibration that defines the irreducible state space of conventional continuous n-qubit theories (which is the complex projective space \\mathbf {CP}^{2^{n}-1}) to an analogous discrete geometry in which the Hopf circle for any n is found to be a discrete set of p + 1 points. The tally of unit-length n-qubit states is given, and reduced via the generalized Hopf fibration to \\mathbf {DCP}^{2^{n}-1}, the discrete analogue of the complex projective space, which has p^{2^{n}-1} (p-1)\\,\\prod _{k=1}^{n-1} ( p^{2^{k}}+1) irreducible states. Using a measure of entanglement, the purity, we explore the entanglement features of discrete quantum states and find that the n-qubit states based on the complexified field \\mathbf {F}_{p^2} have pn(p - 1)n unentangled states (the product of the tally for a single qubit) with purity 1, and they have pn + 1(p - 1)(p + 1)n - 1 maximally entangled states with purity zero.
Kramer, Tobias; Noack, Matthias; Reinefeld, Alexander; Rodríguez, Mirta; Zelinskyy, Yaroslav
2018-06-11
Time- and frequency-resolved optical signals provide insights into the properties of light-harvesting molecular complexes, including excitation energies, dipole strengths and orientations, as well as in the exciton energy flow through the complex. The hierarchical equations of motion (HEOM) provide a unifying theory, which allows one to study the combined effects of system-environment dissipation and non-Markovian memory without making restrictive assumptions about weak or strong couplings or separability of vibrational and electronic degrees of freedom. With increasing system size the exact solution of the open quantum system dynamics requires memory and compute resources beyond a single compute node. To overcome this barrier, we developed a scalable variant of HEOM. Our distributed memory HEOM, DM-HEOM, is a universal tool for open quantum system dynamics. It is used to accurately compute all experimentally accessible time- and frequency-resolved processes in light-harvesting molecular complexes with arbitrary system-environment couplings for a wide range of temperatures and complex sizes. © 2018 Wiley Periodicals, Inc. © 2018 Wiley Periodicals, Inc.
ERIC Educational Resources Information Center
Yoon, Susan A.; Anderson, Emma; Koehler-Yom, Jessica; Evans, Chad; Park, Miyoung; Sheldon, Josh; Schoenfeld, Ilana; Wendel, Daniel; Scheintaub, Hal; Klopfer, Eric
2017-01-01
The recent next generation science standards in the United States have emphasized learning about complex systems as a core feature of science learning. Over the past 15 years, a number of educational tools and theories have been investigated to help students learn about complex systems; but surprisingly, little research has been devoted to…
Iterative Minimum Variance Beamformer with Low Complexity for Medical Ultrasound Imaging.
Deylami, Ali Mohades; Asl, Babak Mohammadzadeh
2018-06-04
Minimum variance beamformer (MVB) improves the resolution and contrast of medical ultrasound images compared with delay and sum (DAS) beamformer. The weight vector of this beamformer should be calculated for each imaging point independently, with a cost of increasing computational complexity. The large number of necessary calculations limits this beamformer to application in real-time systems. A beamformer is proposed based on the MVB with lower computational complexity while preserving its advantages. This beamformer avoids matrix inversion, which is the most complex part of the MVB, by solving the optimization problem iteratively. The received signals from two imaging points close together do not vary much in medical ultrasound imaging. Therefore, using the previously optimized weight vector for one point as initial weight vector for the new neighboring point can improve the convergence speed and decrease the computational complexity. The proposed method was applied on several data sets, and it has been shown that the method can regenerate the results obtained by the MVB while the order of complexity is decreased from O(L 3 ) to O(L 2 ). Copyright © 2018 World Federation for Ultrasound in Medicine and Biology. Published by Elsevier Inc. All rights reserved.
NASA Technical Reports Server (NTRS)
Liang, Shoudan
2000-01-01
Our research effort has produced nine publications in peer-reviewed journals listed at the end of this report. The work reported here are in the following areas: (1) genetic network modeling; (2) autocatalytic model of pre-biotic evolution; (3) theoretical and computational studies of strongly correlated electron systems; (4) reducing thermal oscillations in atomic force microscope; (5) transcription termination mechanism in prokaryotic cells; and (6) the low glutamine usage in thennophiles obtained by studying completely sequenced genomes. We discuss the main accomplishments of these publications.
2015-09-30
Clark (2014), "Using High Performance Computing to Explore Large Complex Bioacoustic Soundscapes : Case Study for Right Whale Acoustics," Procedia...34Using High Performance Computing to Explore Large Complex Bioacoustic Soundscapes : Case Study for Right Whale Acoustics," Procedia Computer Science 20
Study of Liquid Breakup Process in Solid Rocket Motors
2014-01-01
waves. The breakup level increases with the surrounding gas velocity; more liquid breakup in the nozzle throat reduces the liquid alumina droplet size...process of a liquid film that flows along the wall of a straight channel while a high-speed gas moves over it. We have used an unsteady-flow Reynolds...Averaged Navier-Stokes code (URANS) to investigate the interaction of the liquid film flow with the gas flow, and analyzed the breakup process for
Exponential rise of dynamical complexity in quantum computing through projections.
Burgarth, Daniel Klaus; Facchi, Paolo; Giovannetti, Vittorio; Nakazato, Hiromichi; Pascazio, Saverio; Yuasa, Kazuya
2014-10-10
The ability of quantum systems to host exponentially complex dynamics has the potential to revolutionize science and technology. Therefore, much effort has been devoted to developing of protocols for computation, communication and metrology, which exploit this scaling, despite formidable technical difficulties. Here we show that the mere frequent observation of a small part of a quantum system can turn its dynamics from a very simple one into an exponentially complex one, capable of universal quantum computation. After discussing examples, we go on to show that this effect is generally to be expected: almost any quantum dynamics becomes universal once 'observed' as outlined above. Conversely, we show that any complex quantum dynamics can be 'purified' into a simpler one in larger dimensions. We conclude by demonstrating that even local noise can lead to an exponentially complex dynamics.
Biocellion: accelerating computer simulation of multicellular biological system models.
Kang, Seunghwa; Kahan, Simon; McDermott, Jason; Flann, Nicholas; Shmulevich, Ilya
2014-11-01
Biological system behaviors are often the outcome of complex interactions among a large number of cells and their biotic and abiotic environment. Computational biologists attempt to understand, predict and manipulate biological system behavior through mathematical modeling and computer simulation. Discrete agent-based modeling (in combination with high-resolution grids to model the extracellular environment) is a popular approach for building biological system models. However, the computational complexity of this approach forces computational biologists to resort to coarser resolution approaches to simulate large biological systems. High-performance parallel computers have the potential to address the computing challenge, but writing efficient software for parallel computers is difficult and time-consuming. We have developed Biocellion, a high-performance software framework, to solve this computing challenge using parallel computers. To support a wide range of multicellular biological system models, Biocellion asks users to provide their model specifics by filling the function body of pre-defined model routines. Using Biocellion, modelers without parallel computing expertise can efficiently exploit parallel computers with less effort than writing sequential programs from scratch. We simulate cell sorting, microbial patterning and a bacterial system in soil aggregate as case studies. Biocellion runs on x86 compatible systems with the 64 bit Linux operating system and is freely available for academic use. Visit http://biocellion.com for additional information. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
Digitized adiabatic quantum computing with a superconducting circuit.
Barends, R; Shabani, A; Lamata, L; Kelly, J; Mezzacapo, A; Las Heras, U; Babbush, R; Fowler, A G; Campbell, B; Chen, Yu; Chen, Z; Chiaro, B; Dunsworth, A; Jeffrey, E; Lucero, E; Megrant, A; Mutus, J Y; Neeley, M; Neill, C; O'Malley, P J J; Quintana, C; Roushan, P; Sank, D; Vainsencher, A; Wenner, J; White, T C; Solano, E; Neven, H; Martinis, John M
2016-06-09
Quantum mechanics can help to solve complex problems in physics and chemistry, provided they can be programmed in a physical device. In adiabatic quantum computing, a system is slowly evolved from the ground state of a simple initial Hamiltonian to a final Hamiltonian that encodes a computational problem. The appeal of this approach lies in the combination of simplicity and generality; in principle, any problem can be encoded. In practice, applications are restricted by limited connectivity, available interactions and noise. A complementary approach is digital quantum computing, which enables the construction of arbitrary interactions and is compatible with error correction, but uses quantum circuit algorithms that are problem-specific. Here we combine the advantages of both approaches by implementing digitized adiabatic quantum computing in a superconducting system. We tomographically probe the system during the digitized evolution and explore the scaling of errors with system size. We then let the full system find the solution to random instances of the one-dimensional Ising problem as well as problem Hamiltonians that involve more complex interactions. This digital quantum simulation of the adiabatic algorithm consists of up to nine qubits and up to 1,000 quantum logic gates. The demonstration of digitized adiabatic quantum computing in the solid state opens a path to synthesizing long-range correlations and solving complex computational problems. When combined with fault-tolerance, our approach becomes a general-purpose algorithm that is scalable.
A new decision sciences for complex systems.
Lempert, Robert J
2002-05-14
Models of complex systems can capture much useful information but can be difficult to apply to real-world decision-making because the type of information they contain is often inconsistent with that required for traditional decision analysis. New approaches, which use inductive reasoning over large ensembles of computational experiments, now make possible systematic comparison of alternative policy options using models of complex systems. This article describes Computer-Assisted Reasoning, an approach to decision-making under conditions of deep uncertainty that is ideally suited to applying complex systems to policy analysis. The article demonstrates the approach on the policy problem of global climate change, with a particular focus on the role of technology policies in a robust, adaptive strategy for greenhouse gas abatement.
PANORAMA: An approach to performance modeling and diagnosis of extreme-scale workflows
DOE Office of Scientific and Technical Information (OSTI.GOV)
Deelman, Ewa; Carothers, Christopher; Mandal, Anirban
Here we report that computational science is well established as the third pillar of scientific discovery and is on par with experimentation and theory. However, as we move closer toward the ability to execute exascale calculations and process the ensuing extreme-scale amounts of data produced by both experiments and computations alike, the complexity of managing the compute and data analysis tasks has grown beyond the capabilities of domain scientists. Therefore, workflow management systems are absolutely necessary to ensure current and future scientific discoveries. A key research question for these workflow management systems concerns the performance optimization of complex calculation andmore » data analysis tasks. The central contribution of this article is a description of the PANORAMA approach for modeling and diagnosing the run-time performance of complex scientific workflows. This approach integrates extreme-scale systems testbed experimentation, structured analytical modeling, and parallel systems simulation into a comprehensive workflow framework called Pegasus for understanding and improving the overall performance of complex scientific workflows.« less
PANORAMA: An approach to performance modeling and diagnosis of extreme-scale workflows
Deelman, Ewa; Carothers, Christopher; Mandal, Anirban; ...
2015-07-14
Here we report that computational science is well established as the third pillar of scientific discovery and is on par with experimentation and theory. However, as we move closer toward the ability to execute exascale calculations and process the ensuing extreme-scale amounts of data produced by both experiments and computations alike, the complexity of managing the compute and data analysis tasks has grown beyond the capabilities of domain scientists. Therefore, workflow management systems are absolutely necessary to ensure current and future scientific discoveries. A key research question for these workflow management systems concerns the performance optimization of complex calculation andmore » data analysis tasks. The central contribution of this article is a description of the PANORAMA approach for modeling and diagnosing the run-time performance of complex scientific workflows. This approach integrates extreme-scale systems testbed experimentation, structured analytical modeling, and parallel systems simulation into a comprehensive workflow framework called Pegasus for understanding and improving the overall performance of complex scientific workflows.« less
A full computation-relevant topological dynamics classification of elementary cellular automata.
Schüle, Martin; Stoop, Ruedi
2012-12-01
Cellular automata are both computational and dynamical systems. We give a complete classification of the dynamic behaviour of elementary cellular automata (ECA) in terms of fundamental dynamic system notions such as sensitivity and chaoticity. The "complex" ECA emerge to be sensitive, but not chaotic and not eventually weakly periodic. Based on this classification, we conjecture that elementary cellular automata capable of carrying out complex computations, such as needed for Turing-universality, are at the "edge of chaos."
The computational challenges of Earth-system science.
O'Neill, Alan; Steenman-Clark, Lois
2002-06-15
The Earth system--comprising atmosphere, ocean, land, cryosphere and biosphere--is an immensely complex system, involving processes and interactions on a wide range of space- and time-scales. To understand and predict the evolution of the Earth system is one of the greatest challenges of modern science, with success likely to bring enormous societal benefits. High-performance computing, along with the wealth of new observational data, is revolutionizing our ability to simulate the Earth system with computer models that link the different components of the system together. There are, however, considerable scientific and technical challenges to be overcome. This paper will consider four of them: complexity, spatial resolution, inherent uncertainty and time-scales. Meeting these challenges requires a significant increase in the power of high-performance computers. The benefits of being able to make reliable predictions about the evolution of the Earth system should, on their own, amply repay this investment.
NASA Astrophysics Data System (ADS)
Aono, Masashi; Gunji, Yukio-Pegio
2004-08-01
How can non-algorithmic/non-deterministic computational syntax be computed? "The hyperincursive system" introduced by Dubois is an anticipatory system embracing the contradiction/uncertainty. Although it may provide a novel viewpoint for the understanding of complex systems, conventional digital computers cannot run faithfully as the hyperincursive computational syntax specifies, in a strict sense. Then is it an imaginary story? In this paper we try to argue that it is not. We show that a model of complex systems "Elementary Conflictable Cellular Automata (ECCA)" proposed by Aono and Gunji is embracing the hyperincursivity and the nonlocality. ECCA is based on locality-only type settings basically as well as other CA models, and/but at the same time, each cell is required to refer to globality-dominant regularity. Due to this contradictory locality-globality loop, the time evolution equation specifies that the system reaches the deadlock/infinite-loop. However, we show that there is a possibility of the resolution of these problems if the computing system has parallel and/but non-distributed property like an amoeboid organism. This paper is an introduction to "the slime mold computing" that is an attempt to cultivate an unconventional notion of computation.
NASA Astrophysics Data System (ADS)
Nguyen, L.; Chee, T.; Minnis, P.; Spangenberg, D.; Ayers, J. K.; Palikonda, R.; Vakhnin, A.; Dubois, R.; Murphy, P. R.
2014-12-01
The processing, storage and dissemination of satellite cloud and radiation products produced at NASA Langley Research Center are key activities for the Climate Science Branch. A constellation of systems operates in sync to accomplish these goals. Because of the complexity involved with operating such intricate systems, there are both high failure rates and high costs for hardware and system maintenance. Cloud computing has the potential to ameliorate cost and complexity issues. Over time, the cloud computing model has evolved and hybrid systems comprising off-site as well as on-site resources are now common. Towards our mission of providing the highest quality research products to the widest audience, we have explored the use of the Amazon Web Services (AWS) Cloud and Storage and present a case study of our results and efforts. This project builds upon NASA Langley Cloud and Radiation Group's experience with operating large and complex computing infrastructures in a reliable and cost effective manner to explore novel ways to leverage cloud computing resources in the atmospheric science environment. Our case study presents the project requirements and then examines the fit of AWS with the LaRC computing model. We also discuss the evaluation metrics, feasibility, and outcomes and close the case study with the lessons we learned that would apply to others interested in exploring the implementation of the AWS system in their own atmospheric science computing environments.
2010-01-01
Background The robust storage, updating and utilization of information are necessary for the maintenance and perpetuation of dynamic systems. These systems can exist as constructs of metal-oxide semiconductors and silicon, as in a digital computer, or in the "wetware" of organic compounds, proteins and nucleic acids that make up biological organisms. We propose that there are essential functional properties of centralized information-processing systems; for digital computers these properties reside in the computer's hard drive, and for eukaryotic cells they are manifest in the DNA and associated structures. Methods Presented herein is a descriptive framework that compares DNA and its associated proteins and sub-nuclear structure with the structure and function of the computer hard drive. We identify four essential properties of information for a centralized storage and processing system: (1) orthogonal uniqueness, (2) low level formatting, (3) high level formatting and (4) translation of stored to usable form. The corresponding aspects of the DNA complex and a computer hard drive are categorized using this classification. This is intended to demonstrate a functional equivalence between the components of the two systems, and thus the systems themselves. Results Both the DNA complex and the computer hard drive contain components that fulfill the essential properties of a centralized information storage and processing system. The functional equivalence of these components provides insight into both the design process of engineered systems and the evolved solutions addressing similar system requirements. However, there are points where the comparison breaks down, particularly when there are externally imposed information-organizing structures on the computer hard drive. A specific example of this is the imposition of the File Allocation Table (FAT) during high level formatting of the computer hard drive and the subsequent loading of an operating system (OS). Biological systems do not have an external source for a map of their stored information or for an operational instruction set; rather, they must contain an organizational template conserved within their intra-nuclear architecture that "manipulates" the laws of chemistry and physics into a highly robust instruction set. We propose that the epigenetic structure of the intra-nuclear environment and the non-coding RNA may play the roles of a Biological File Allocation Table (BFAT) and biological operating system (Bio-OS) in eukaryotic cells. Conclusions The comparison of functional and structural characteristics of the DNA complex and the computer hard drive leads to a new descriptive paradigm that identifies the DNA as a dynamic storage system of biological information. This system is embodied in an autonomous operating system that inductively follows organizational structures, data hierarchy and executable operations that are well understood in the computer science industry. Characterizing the "DNA hard drive" in this fashion can lead to insights arising from discrepancies in the descriptive framework, particularly with respect to positing the role of epigenetic processes in an information-processing context. Further expansions arising from this comparison include the view of cells as parallel computing machines and a new approach towards characterizing cellular control systems. PMID:20092652
D'Onofrio, David J; An, Gary
2010-01-21
The robust storage, updating and utilization of information are necessary for the maintenance and perpetuation of dynamic systems. These systems can exist as constructs of metal-oxide semiconductors and silicon, as in a digital computer, or in the "wetware" of organic compounds, proteins and nucleic acids that make up biological organisms. We propose that there are essential functional properties of centralized information-processing systems; for digital computers these properties reside in the computer's hard drive, and for eukaryotic cells they are manifest in the DNA and associated structures. Presented herein is a descriptive framework that compares DNA and its associated proteins and sub-nuclear structure with the structure and function of the computer hard drive. We identify four essential properties of information for a centralized storage and processing system: (1) orthogonal uniqueness, (2) low level formatting, (3) high level formatting and (4) translation of stored to usable form. The corresponding aspects of the DNA complex and a computer hard drive are categorized using this classification. This is intended to demonstrate a functional equivalence between the components of the two systems, and thus the systems themselves. Both the DNA complex and the computer hard drive contain components that fulfill the essential properties of a centralized information storage and processing system. The functional equivalence of these components provides insight into both the design process of engineered systems and the evolved solutions addressing similar system requirements. However, there are points where the comparison breaks down, particularly when there are externally imposed information-organizing structures on the computer hard drive. A specific example of this is the imposition of the File Allocation Table (FAT) during high level formatting of the computer hard drive and the subsequent loading of an operating system (OS). Biological systems do not have an external source for a map of their stored information or for an operational instruction set; rather, they must contain an organizational template conserved within their intra-nuclear architecture that "manipulates" the laws of chemistry and physics into a highly robust instruction set. We propose that the epigenetic structure of the intra-nuclear environment and the non-coding RNA may play the roles of a Biological File Allocation Table (BFAT) and biological operating system (Bio-OS) in eukaryotic cells. The comparison of functional and structural characteristics of the DNA complex and the computer hard drive leads to a new descriptive paradigm that identifies the DNA as a dynamic storage system of biological information. This system is embodied in an autonomous operating system that inductively follows organizational structures, data hierarchy and executable operations that are well understood in the computer science industry. Characterizing the "DNA hard drive" in this fashion can lead to insights arising from discrepancies in the descriptive framework, particularly with respect to positing the role of epigenetic processes in an information-processing context. Further expansions arising from this comparison include the view of cells as parallel computing machines and a new approach towards characterizing cellular control systems.
DISCRETE EVENT SIMULATION OF OPTICAL SWITCH MATRIX PERFORMANCE IN COMPUTER NETWORKS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Imam, Neena; Poole, Stephen W
2013-01-01
In this paper, we present application of a Discrete Event Simulator (DES) for performance modeling of optical switching devices in computer networks. Network simulators are valuable tools in situations where one cannot investigate the system directly. This situation may arise if the system under study does not exist yet or the cost of studying the system directly is prohibitive. Most available network simulators are based on the paradigm of discrete-event-based simulation. As computer networks become increasingly larger and more complex, sophisticated DES tool chains have become available for both commercial and academic research. Some well-known simulators are NS2, NS3, OPNET,more » and OMNEST. For this research, we have applied OMNEST for the purpose of simulating multi-wavelength performance of optical switch matrices in computer interconnection networks. Our results suggest that the application of DES to computer interconnection networks provides valuable insight in device performance and aids in topology and system optimization.« less
Geometry of Quantum Computation with Qudits
Luo, Ming-Xing; Chen, Xiu-Bo; Yang, Yi-Xian; Wang, Xiaojun
2014-01-01
The circuit complexity of quantum qubit system evolution as a primitive problem in quantum computation has been discussed widely. We investigate this problem in terms of qudit system. Using the Riemannian geometry the optimal quantum circuits are equivalent to the geodetic evolutions in specially curved parametrization of SU(dn). And the quantum circuit complexity is explicitly dependent of controllable approximation error bound. PMID:24509710
Decreasing the temporal complexity for nonlinear, implicit reduced-order models by forecasting
Carlberg, Kevin; Ray, Jaideep; van Bloemen Waanders, Bart
2015-02-14
Implicit numerical integration of nonlinear ODEs requires solving a system of nonlinear algebraic equations at each time step. Each of these systems is often solved by a Newton-like method, which incurs a sequence of linear-system solves. Most model-reduction techniques for nonlinear ODEs exploit knowledge of system's spatial behavior to reduce the computational complexity of each linear-system solve. However, the number of linear-system solves for the reduced-order simulation often remains roughly the same as that for the full-order simulation. We propose exploiting knowledge of the model's temporal behavior to (1) forecast the unknown variable of the reduced-order system of nonlinear equationsmore » at future time steps, and (2) use this forecast as an initial guess for the Newton-like solver during the reduced-order-model simulation. To compute the forecast, we propose using the Gappy POD technique. As a result, the goal is to generate an accurate initial guess so that the Newton solver requires many fewer iterations to converge, thereby decreasing the number of linear-system solves in the reduced-order-model simulation.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wolverton, Christopher; Ozolins, Vidvuds; Kung, Harold H.
The objective of the proposed program is to discover novel mixed hydrides for hydrogen storage, which enable the DOE 2010 system-level goals. Our goal is to find a material that desorbs 8.5 wt.% H 2 or more at temperatures below 85°C. The research program will combine first-principles calculations of reaction thermodynamics and kinetics with material and catalyst synthesis, testing, and characterization. We will combine materials from distinct categories (e.g., chemical and complex hydrides) to form novel multicomponent reactions. Systems to be studied include mixtures of complex hydrides and chemical hydrides [e.g. LiNH 2+NH 3BH 3] and nitrogen-hydrogen based borohydrides [e.g.more » Al(BH 4) 3(NH 3) 3]. The 2010 and 2015 FreedomCAR/DOE targets for hydrogen storage systems are very challenging, and cannot be met with existing materials. The vast majority of the work to date has delineated materials into various classes, e.g., complex and metal hydrides, chemical hydrides, and sorbents. However, very recent studies indicate that mixtures of storage materials, particularly mixtures between various classes, hold promise to achieve technological attributes that materials within an individual class cannot reach. Our project involves a systematic, rational approach to designing novel multicomponent mixtures of materials with fast hydrogenation/dehydrogenation kinetics and favorable thermodynamics using a combination of state-of-the-art scientific computing and experimentation. We will use the accurate predictive power of first-principles modeling to understand the thermodynamic and microscopic kinetic processes involved in hydrogen release and uptake and to design new material/catalyst systems with improved properties. Detailed characterization and atomic-scale catalysis experiments will elucidate the effect of dopants and nanoscale catalysts in achieving fast kinetics and reversibility. And, state-of-the-art storage experiments will give key storage attributes of the investigated reactions, validate computational predictions, and help guide and improve computational methods. In sum, our approach involves a powerful blend of: 1) H2 Storage measurements and characterization, 2) State-of-the-art computational modeling, 3) Detailed catalysis experiments, 4) In-depth automotive perspective.« less
A Computational Workflow for the Automated Generation of Models of Genetic Designs.
Misirli, Göksel; Nguyen, Tramy; McLaughlin, James Alastair; Vaidyanathan, Prashant; Jones, Timothy S; Densmore, Douglas; Myers, Chris; Wipat, Anil
2018-06-05
Computational models are essential to engineer predictable biological systems and to scale up this process for complex systems. Computational modeling often requires expert knowledge and data to build models. Clearly, manual creation of models is not scalable for large designs. Despite several automated model construction approaches, computational methodologies to bridge knowledge in design repositories and the process of creating computational models have still not been established. This paper describes a workflow for automatic generation of computational models of genetic circuits from data stored in design repositories using existing standards. This workflow leverages the software tool SBOLDesigner to build structural models that are then enriched by the Virtual Parts Repository API using Systems Biology Open Language (SBOL) data fetched from the SynBioHub design repository. The iBioSim software tool is then utilized to convert this SBOL description into a computational model encoded using the Systems Biology Markup Language (SBML). Finally, this SBML model can be simulated using a variety of methods. This workflow provides synthetic biologists with easy to use tools to create predictable biological systems, hiding away the complexity of building computational models. This approach can further be incorporated into other computational workflows for design automation.
Moving alcohol prevention research forward-Part I: introducing a complex systems paradigm.
Apostolopoulos, Yorghos; Lemke, Michael K; Barry, Adam E; Lich, Kristen Hassmiller
2018-02-01
The drinking environment is a complex system consisting of a number of heterogeneous, evolving and interacting components, which exhibit circular causality and emergent properties. These characteristics reduce the efficacy of commonly used research approaches, which typically do not account for the underlying dynamic complexity of alcohol consumption and the interdependent nature of diverse factors influencing misuse over time. We use alcohol misuse among college students in the United States as an example for framing our argument for a complex systems paradigm. A complex systems paradigm, grounded in socio-ecological and complex systems theories and computational modeling and simulation, is introduced. Theoretical, conceptual, methodological and analytical underpinnings of this paradigm are described in the context of college drinking prevention research. The proposed complex systems paradigm can transcend limitations of traditional approaches, thereby fostering new directions in alcohol prevention research. By conceptualizing student alcohol misuse as a complex adaptive system, computational modeling and simulation methodologies and analytical techniques can be used. Moreover, use of participatory model-building approaches to generate simulation models can further increase stakeholder buy-in, understanding and policymaking. A complex systems paradigm for research into alcohol misuse can provide a holistic understanding of the underlying drinking environment and its long-term trajectory, which can elucidate high-leverage preventive interventions. © 2017 Society for the Study of Addiction.
Computed Tomography Inspection and Analysis for Additive Manufacturing Components
NASA Technical Reports Server (NTRS)
Beshears, Ronald D.
2017-01-01
Computed tomography (CT) inspection was performed on test articles additively manufactured from metallic materials. Metallic AM and machined wrought alloy test articles with programmed flaws and geometric features were inspected using a 2-megavolt linear accelerator based CT system. Performance of CT inspection on identically configured wrought and AM components and programmed flaws was assessed to determine the impact of additive manufacturing on inspectability of objects with complex geometries.
Prognostics Methodology for Complex Systems
NASA Technical Reports Server (NTRS)
Gulati, Sandeep; Mackey, Ryan
2003-01-01
An automatic method to schedule maintenance and repair of complex systems is produced based on a computational structure called the Informed Maintenance Grid (IMG). This method provides solutions to the two fundamental problems in autonomic logistics: (1) unambiguous detection of deterioration or impending loss of function and (2) determination of the time remaining to perform maintenance or other corrective action based upon information from the system. The IMG provides a health determination over the medium-to-longterm operation of the system, from one or more days to years of study. The IMG is especially applicable to spacecraft and both piloted and autonomous aircraft, or industrial control processes.
Understanding System of Systems Development Using an Agent-Based Wave Model
2012-01-01
Procedia Computer Science Procedia Computer Science 00 (2012) 000–000 www.elsevier.com/locate/ procedia Complex Adaptive Systems...integration of technical systems as well as cognitive and social processes, which alter system behavior [6]. As mentioned before * Corresponding...Prescribed by ANSI Std Z39-18 Acheson/ Procedia Computer Science 00 (2012) 000–000 most system architects assume that SoS participants exhibit
Simulated parallel annealing within a neighborhood for optimization of biomechanical systems.
Higginson, J S; Neptune, R R; Anderson, F C
2005-09-01
Optimization problems for biomechanical systems have become extremely complex. Simulated annealing (SA) algorithms have performed well in a variety of test problems and biomechanical applications; however, despite advances in computer speed, convergence to optimal solutions for systems of even moderate complexity has remained prohibitive. The objective of this study was to develop a portable parallel version of a SA algorithm for solving optimization problems in biomechanics. The algorithm for simulated parallel annealing within a neighborhood (SPAN) was designed to minimize interprocessor communication time and closely retain the heuristics of the serial SA algorithm. The computational speed of the SPAN algorithm scaled linearly with the number of processors on different computer platforms for a simple quadratic test problem and for a more complex forward dynamic simulation of human pedaling.
The discrete hungry Lotka Volterra system and a new algorithm for computing matrix eigenvalues
NASA Astrophysics Data System (ADS)
Fukuda, Akiko; Ishiwata, Emiko; Iwasaki, Masashi; Nakamura, Yoshimasa
2009-01-01
The discrete hungry Lotka-Volterra (dhLV) system is a generalization of the discrete Lotka-Volterra (dLV) system which stands for a prey-predator model in mathematical biology. In this paper, we show that (1) some invariants exist which are expressed by dhLV variables and are independent from the discrete time and (2) a dhLV variable converges to some positive constant or zero as the discrete time becomes sufficiently large. Some characteristic polynomial is then factorized with the help of the dhLV system. The asymptotic behaviour of the dhLV system enables us to design an algorithm for computing complex eigenvalues of a certain band matrix.
Molecular Dynamics Simulations of Protein-Ligand Complexes in Near Physiological Conditions
NASA Astrophysics Data System (ADS)
Wambo, Thierry Oscar
Proteins are important molecules for their key functions. However, under certain circumstances, the function of these proteins needs to be regulated to keep us healthy. Ligands are small molecules often used to modulate the function of proteins. The binding affinity is a quantitative measure of how strong the ligand will modulate the function of the protein: a strong binding affinity will highly impact the performance of the protein. It becomes clear that it is critical to have appropriate techniques to accurately compute the binding affinity. The most difficult task in computer simulations is how to efficiently sample the space spanned by the ligand during the binding process. In this work, we have developed some schemes to compute the binding affinity of a ligand to a protein, and of a metal ion to a protein. Application of these techniques to some complexes yield results in agreement with experimental values. These methods are a brute force approach and make no assumption other than that the complexes are governed by the force field used. Specifically, we computed the free energy of binding between (1) human carbonic anhydrase II and the drug acetazolamide (hcaII-AZM), (2) human carbonic anhydrase II and the zinc ion (hcaII-Zinc), and (3) beta-lactoglobulin and five fatty acids complexes (BLG-FAs). We found the following free energies of binding in unit of kcal/mol: -12.96 +/-2.44 (-15.74) for hcaII-Zinc complex, -5.76+/-0.76 (-5.57) for BLG-OCA , -4.44+/-1.08 (-5.22) for BLG-DKA,-6.89+/-1.25 (-7.24) for BLG-DAO, -8.57+/-0.82 (-8.14) for BLG-MYR, -8.99+/-0.87 (-8.72) for BLG-PLM, and -11.87+/-1.8 (-10.8) for hcaII-AZM. The values inside the parentheses are experimental results. The simulations and quantitative analysis of each system provide interesting insights into the interactions between each entity and helps us to better understand the dynamics of these systems.
The role of the host in a cooperating mainframe and workstation environment, volumes 1 and 2
NASA Technical Reports Server (NTRS)
Kusmanoff, Antone; Martin, Nancy L.
1989-01-01
In recent years, advancements made in computer systems have prompted a move from centralized computing based on timesharing a large mainframe computer to distributed computing based on a connected set of engineering workstations. A major factor in this advancement is the increased performance and lower cost of engineering workstations. The shift to distributed computing from centralized computing has led to challenges associated with the residency of application programs within the system. In a combined system of multiple engineering workstations attached to a mainframe host, the question arises as to how does a system designer assign applications between the larger mainframe host and the smaller, yet powerful, workstation. The concepts related to real time data processing are analyzed and systems are displayed which use a host mainframe and a number of engineering workstations interconnected by a local area network. In most cases, distributed systems can be classified as having a single function or multiple functions and as executing programs in real time or nonreal time. In a system of multiple computers, the degree of autonomy of the computers is important; a system with one master control computer generally differs in reliability, performance, and complexity from a system in which all computers share the control. This research is concerned with generating general criteria principles for software residency decisions (host or workstation) for a diverse yet coupled group of users (the clustered workstations) which may need the use of a shared resource (the mainframe) to perform their functions.
Egri-Nagy, Attila; Nehaniv, Chrystopher L
2008-01-01
Beyond complexity measures, sometimes it is worthwhile in addition to investigate how complexity changes structurally, especially in artificial systems where we have complete knowledge about the evolutionary process. Hierarchical decomposition is a useful way of assessing structural complexity changes of organisms modeled as automata, and we show how recently developed computational tools can be used for this purpose, by computing holonomy decompositions and holonomy complexity. To gain insight into the evolution of complexity, we investigate the smoothness of the landscape structure of complexity under minimal transitions. As a proof of concept, we illustrate how the hierarchical complexity analysis reveals symmetries and irreversible structure in biological networks by applying the methods to the lac operon mechanism in the genetic regulatory network of Escherichia coli.
Synthetic analog and digital circuits for cellular computation and memory.
Purcell, Oliver; Lu, Timothy K
2014-10-01
Biological computation is a major area of focus in synthetic biology because it has the potential to enable a wide range of applications. Synthetic biologists have applied engineering concepts to biological systems in order to construct progressively more complex gene circuits capable of processing information in living cells. Here, we review the current state of computational genetic circuits and describe artificial gene circuits that perform digital and analog computation. We then discuss recent progress in designing gene networks that exhibit memory, and how memory and computation have been integrated to yield more complex systems that can both process and record information. Finally, we suggest new directions for engineering biological circuits capable of computation. Copyright © 2014 The Authors. Published by Elsevier Ltd.. All rights reserved.
Urban Typologies: Towards an ORNL Urban Information System (UrbIS)
NASA Astrophysics Data System (ADS)
KC, B.; King, A. W.; Sorokine, A.; Crow, M. C.; Devarakonda, R.; Hilbert, N. L.; Karthik, R.; Patlolla, D.; Surendran Nair, S.
2016-12-01
Urban environments differ in a large number of key attributes; these include infrastructure, morphology, demography, and economic and social variables, among others. These attributes determine many urban properties such as energy and water consumption, greenhouse gas emissions, air quality, public health, sustainability, and vulnerability and resilience to climate change. Characterization of urban environments by a single property such as population size does not sufficiently capture this complexity. In addressing this multivariate complexity one typically faces such problems as disparate and scattered data, challenges of big data management, spatial searching, insufficient computational capacity for data-driven analysis and modelling, and the lack of tools to quickly visualize the data and compare the analytical results across different cities and regions. We have begun the development of an Urban Information System (UrbIS) to address these issues, one that embraces the multivariate "big data" of urban areas and their environments across the United States utilizing the Big Data as a Service (BDaaS) concept. With technological roots in High-performance Computing (HPC), BDaaS is based on the idea of outsourcing computations to different computing paradigms, scalable to super-computers. UrbIS aims to incorporate federated metadata search, integrated modeling and analysis, and geovisualization into a single seamless workflow. The system includes web-based 2D/3D visualization with an iGlobe interface, fast cloud-based and server-side data processing and analysis, and a metadata search engine based on the Mercury data search system developed at Oak Ridge National Laboratory (ORNL). Results of analyses will be made available through web services. We are implementing UrbIS in ORNL's Compute and Data Environment for Science (CADES) and are leveraging ORNL experience in complex data and geospatial projects. The development of UrbIS is being guided by an investigation of urban heat islands (UHI) using high-dimensional clustering and statistics to define urban typologies (types of cities) in an investigation of how UHI vary with urban type across the United States.
Visual Complexity in Orthographic Learning: Modeling Learning across Writing System Variations
ERIC Educational Resources Information Center
Chang, Li-Yun; Plaut, David C.; Perfetti, Charles A.
2016-01-01
The visual complexity of orthographies varies across writing systems. Prior research has shown that complexity strongly influences the initial stage of reading development: the perceptual learning of grapheme forms. This study presents a computational simulation that examines the degree to which visual complexity leads to grapheme learning…
Computer-Aided Reliability Estimation
NASA Technical Reports Server (NTRS)
Bavuso, S. J.; Stiffler, J. J.; Bryant, L. A.; Petersen, P. L.
1986-01-01
CARE III (Computer-Aided Reliability Estimation, Third Generation) helps estimate reliability of complex, redundant, fault-tolerant systems. Program specifically designed for evaluation of fault-tolerant avionics systems. However, CARE III general enough for use in evaluation of other systems as well.
Channel Model Optimization with Reflection Residual Component for Indoor MIMO-VLC System
NASA Astrophysics Data System (ADS)
Chen, Yong; Li, Tengfei; Liu, Huanlin; Li, Yichao
2017-12-01
A fast channel modeling method is studied to solve the problem of reflection channel gain for multiple input multiple output-visible light communications (MIMO-VLC) in the paper. For reducing the computational complexity when associating with the reflection times, no more than 3 reflections are taken into consideration in VLC. We think that higher order reflection link consists of corresponding many times line of sight link and firstly present reflection residual component to characterize higher reflection (more than 2 reflections). We perform computer simulation results for point-to-point channel impulse response, receiving optical power and receiving signal to noise ratio. Based on theoretical analysis and simulation results, the proposed method can effectively reduce the computational complexity of higher order reflection in channel modeling.
Gray: a ray tracing-based Monte Carlo simulator for PET.
Freese, David L; Olcott, Peter D; Buss, Samuel R; Levin, Craig S
2018-05-21
Monte Carlo simulation software plays a critical role in PET system design. Performing complex, repeated Monte Carlo simulations can be computationally prohibitive, as even a single simulation can require a large amount of time and a computing cluster to complete. Here we introduce Gray, a Monte Carlo simulation software for PET systems. Gray exploits ray tracing methods used in the computer graphics community to greatly accelerate simulations of PET systems with complex geometries. We demonstrate the implementation of models for positron range, annihilation acolinearity, photoelectric absorption, Compton scatter, and Rayleigh scatter. For validation, we simulate the GATE PET benchmark, and compare energy, distribution of hits, coincidences, and run time. We show a [Formula: see text] speedup using Gray, compared to GATE for the same simulation, while demonstrating nearly identical results. We additionally simulate the Siemens Biograph mCT system with both the NEMA NU-2 scatter phantom and sensitivity phantom. We estimate the total sensitivity within [Formula: see text]% when accounting for differences in peak NECR. We also estimate the peak NECR to be [Formula: see text] kcps, or within [Formula: see text]% of published experimental data. The activity concentration of the peak is also estimated within 1.3%.
Building an adiabatic quantum computer simulation in the classroom
NASA Astrophysics Data System (ADS)
Rodríguez-Laguna, Javier; Santalla, Silvia N.
2018-05-01
We present a didactic introduction to adiabatic quantum computation (AQC) via the explicit construction of a classical simulator of quantum computers. This constitutes a suitable route to introduce several important concepts for advanced undergraduates in physics: quantum many-body systems, quantum phase transitions, disordered systems, spin-glasses, and computational complexity theory.
Shen, Weifeng; Jiang, Libing; Zhang, Mao; Ma, Yuefeng; Jiang, Guanyu; He, Xiaojun
2014-01-01
To review the research methods of mass casualty incident (MCI) systematically and introduce the concept and characteristics of complexity science and artificial system, computational experiments and parallel execution (ACP) method. We searched PubMed, Web of Knowledge, China Wanfang and China Biology Medicine (CBM) databases for relevant studies. Searches were performed without year or language restrictions and used the combinations of the following key words: "mass casualty incident", "MCI", "research method", "complexity science", "ACP", "approach", "science", "model", "system" and "response". Articles were searched using the above keywords and only those involving the research methods of mass casualty incident (MCI) were enrolled. Research methods of MCI have increased markedly over the past few decades. For now, dominating research methods of MCI are theory-based approach, empirical approach, evidence-based science, mathematical modeling and computer simulation, simulation experiment, experimental methods, scenario approach and complexity science. This article provides an overview of the development of research methodology for MCI. The progresses of routine research approaches and complexity science are briefly presented in this paper. Furthermore, the authors conclude that the reductionism underlying the exact science is not suitable for MCI complex systems. And the only feasible alternative is complexity science. Finally, this summary is followed by a review that ACP method combining artificial systems, computational experiments and parallel execution provides a new idea to address researches for complex MCI.
Jungle Computing: Distributed Supercomputing Beyond Clusters, Grids, and Clouds
NASA Astrophysics Data System (ADS)
Seinstra, Frank J.; Maassen, Jason; van Nieuwpoort, Rob V.; Drost, Niels; van Kessel, Timo; van Werkhoven, Ben; Urbani, Jacopo; Jacobs, Ceriel; Kielmann, Thilo; Bal, Henri E.
In recent years, the application of high-performance and distributed computing in scientific practice has become increasingly wide spread. Among the most widely available platforms to scientists are clusters, grids, and cloud systems. Such infrastructures currently are undergoing revolutionary change due to the integration of many-core technologies, providing orders-of-magnitude speed improvements for selected compute kernels. With high-performance and distributed computing systems thus becoming more heterogeneous and hierarchical, programming complexity is vastly increased. Further complexities arise because urgent desire for scalability and issues including data distribution, software heterogeneity, and ad hoc hardware availability commonly force scientists into simultaneous use of multiple platforms (e.g., clusters, grids, and clouds used concurrently). A true computing jungle.
NASA Technical Reports Server (NTRS)
Johnson, S. C.
1982-01-01
An interface system for passing data between a relational information management (RIM) data base complex and engineering analysis language (EAL), a finite element structural analysis program is documented. The interface system, implemented on a CDC Cyber computer, is composed of two FORTRAN programs called RIM2EAL and EAL2RIM. The RIM2EAL reads model definition data from RIM and creates a file of EAL commands to define the model. The EAL2RIM reads model definition and EAL generated analysis data from EAL's data library and stores these data dirctly in a RIM data base. These two interface programs and the format for the RIM data complex are described.
Casey, M
1996-08-15
Recurrent neural networks (RNNs) can learn to perform finite state computations. It is shown that an RNN performing a finite state computation must organize its state space to mimic the states in the minimal deterministic finite state machine that can perform that computation, and a precise description of the attractor structure of such systems is given. This knowledge effectively predicts activation space dynamics, which allows one to understand RNN computation dynamics in spite of complexity in activation dynamics. This theory provides a theoretical framework for understanding finite state machine (FSM) extraction techniques and can be used to improve training methods for RNNs performing FSM computations. This provides an example of a successful approach to understanding a general class of complex systems that has not been explicitly designed, e.g., systems that have evolved or learned their internal structure.
Complex Systems Simulation and Optimization | Computational Science | NREL
account. Stochastic Optimization and Control: Formulation and implementation of advanced optimization and account uncertainty. Contact Wesley Jones Group Manager, Complex Systems Simulation and Optimiziation
Computational complexities and storage requirements of some Riccati equation solvers
NASA Technical Reports Server (NTRS)
Utku, Senol; Garba, John A.; Ramesh, A. V.
1989-01-01
The linear optimal control problem of an nth-order time-invariant dynamic system with a quadratic performance functional is usually solved by the Hamilton-Jacobi approach. This leads to the solution of the differential matrix Riccati equation with a terminal condition. The bulk of the computation for the optimal control problem is related to the solution of this equation. There are various algorithms in the literature for solving the matrix Riccati equation. However, computational complexities and storage requirements as a function of numbers of state variables, control variables, and sensors are not available for all these algorithms. In this work, the computational complexities and storage requirements for some of these algorithms are given. These expressions show the immensity of the computational requirements of the algorithms in solving the Riccati equation for large-order systems such as the control of highly flexible space structures. The expressions are also needed to compute the speedup and efficiency of any implementation of these algorithms on concurrent machines.
Applications of complex systems theory in nursing education, research, and practice.
Clancy, Thomas R; Effken, Judith A; Pesut, Daniel
2008-01-01
The clinical and administrative processes in today's healthcare environment are becoming increasingly complex. Multiple providers, new technology, competition, and the growing ubiquity of information all contribute to the notion of health care as a complex system. A complex system (CS) is characterized by a highly connected network of entities (e.g., physical objects, people or groups of people) from which higher order behavior emerges. Research in the transdisciplinary field of CS has focused on the use of computational modeling and simulation as a methodology for analyzing CS behavior. The creation of virtual worlds through computer simulation allows researchers to analyze multiple variables simultaneously and begin to understand behaviors that are common regardless of the discipline. The application of CS principles, mediated through computer simulation, informs nursing practice of the benefits and drawbacks of new procedures, protocols and practices before having to actually implement them. The inclusion of new computational tools and their applications in nursing education is also gaining attention. For example, education in CSs and applied computational applications has been endorsed by The Institute of Medicine, the American Organization of Nurse Executives and the American Association of Colleges of Nursing as essential training of nurse leaders. The purpose of this article is to review current research literature regarding CS science within the context of expert practice and implications for the education of nurse leadership roles. The article focuses on 3 broad areas: CS defined, literature review and exemplars from CS research and applications of CS theory in nursing leadership education. The article also highlights the key role nursing informaticists play in integrating emerging computational tools in the analysis of complex nursing systems.
GMXPBSA 2.1: A GROMACS tool to perform MM/PBSA and computational alanine scanning
NASA Astrophysics Data System (ADS)
Paissoni, C.; Spiliotopoulos, D.; Musco, G.; Spitaleri, A.
2015-01-01
GMXPBSA 2.1 is a user-friendly suite of Bash/Perl scripts for streamlining MM/PBSA calculations on structural ensembles derived from GROMACS trajectories, to automatically calculate binding free energies for protein-protein or ligand-protein complexes [R.T. Bradshaw et al., Protein Eng. Des. Sel. 24 (2011) 197-207]. GMXPBSA 2.1 is flexible and can easily be customized to specific needs and it is an improvement of the previous GMXPBSA 2.0 [C. Paissoni et al., Comput. Phys. Commun. (2014), 185, 2920-2929]. Additionally, it performs computational alanine scanning (CAS) to study the effects of ligand and/or receptor alanine mutations on the free energy of binding. Calculations require only for protein-protein or protein-ligand MD simulations. GMXPBSA 2.1 performs different comparative analyses, including a posteriori generation of alanine mutants of the wild-type complex, calculation of the binding free energy values of the mutant complexes and comparison of the results with the wild-type system. Moreover, it compares the binding free energy of different complex trajectories, allowing the study of the effects of non-alanine mutations, post-translational modifications or unnatural amino acids on the binding free energy of the system under investigation. Finally, it can calculate and rank relative affinity to the same receptor utilizing MD simulations of proteins in complex with different ligands. In order to dissect the different MM/PBSA energy contributions, including molecular mechanic (MM), electrostatic contribution to solvation (PB) and nonpolar contribution to solvation (SA), the tool combines two freely available programs: the MD simulations software GROMACS [S. Pronk et al., Bioinformatics 29 (2013) 845-854] and the Poisson-Boltzmann equation solver APBS [N.A. Baker et al., Proc. Natl. Acad. Sci. U.S.A 98 (2001) 10037-10041]. All the calculations can be performed in single or distributed automatic fashion on a cluster facility in order to increase the calculation by dividing frames across the available processors. This new version with respect to our previously published GMXPBSA 2.0 fixes some problem and allows additional kind of calculations, such as CAS on single protein in order to individuate the hot-spots, more custom options to perform APBS calculations, improvements of speed calculation of APBS (precF set to 0), possibility to work with multichain systems (see Summary of revisions for more details). The program is freely available under the GPL license.
Simulations of Turbulent Momentum and Scalar Transport in Confined Swirling Coaxial Jets
NASA Technical Reports Server (NTRS)
Shih, Tsan-Hsing; Liu, Nan-Suey; Moder, Jeffrey P.
2015-01-01
This paper presents the numerical simulations of confined three-dimensional coaxial water jets. The objectives are to validate the newly proposed nonlinear turbulence models of momentum and scalar transport, and to evaluate the newly introduced scalar APDF and DWFDF equation along with its Eulerian implementation in the National Combustion Code(NCC). Simulations conducted include the steady RANS, the unsteady RANS (URANS), and the time-filtered Navier-Stokes (TFNS); both without and with invoking the APDF or DWFDF equation.
ELECTROLYTIC PRODUCTION OF URANIUM TETRAFLUORIDE
Lofthouse, E.
1954-08-31
This patent relates to electrolytic methods for the production of uranium tetrafluoride. According to the present invention a process for the production of uranium tetrafluoride comprises submitting to electrolysis an aqueous solution of uranyl fluoride containing free hydrofluoric acid. Advantageously the aqueous solution of uranyl fluoride is obtained by dissolving uranium hexafluoride in water. On electrolysis, the uranyl ions are reduced to uranous tons at the cathode and immediately combine with the fluoride ions in solution to form the insoluble uranium tetrafluoride which is precipitated.
Study of Liquid Breakup Process in Solid Rocket Motor Nozzle
2016-02-16
liquid film flow with the gas flow. The rate of the wave breakup was characterized by introducing Breakup-length, Ohnesorge Number (Oh) and Weber Number... liquid film that flows along the wall of a strraight test channel while a relatively higher-speed gas moves over it. We have used an unsteady-flow...Reynolds- Averaged Navier-Stokes code (URANS) to investigate the interaction of the liquid film flow with the gas flow. The rate of the wave breakup was
NASA Technical Reports Server (NTRS)
Yakimovsky, Y.
1974-01-01
An approach to simultaneous interpretation of objects in complex structures so as to maximize a combined utility function is presented. Results of the application of a computer software system to assign meaning to regions in a segmented image based on the principles described in this paper and on a special interactive sequential classification learning system, which is referenced, are demonstrated.
Latent Computational Complexity of Symmetry-Protected Topological Order with Fractional Symmetry.
Miller, Jacob; Miyake, Akimasa
2018-04-27
An emerging insight is that ground states of symmetry-protected topological orders (SPTOs) possess latent computational complexity in terms of their many-body entanglement. By introducing a fractional symmetry of SPTO, which requires the invariance under 3-colorable symmetries of a lattice, we prove that every renormalization fixed-point state of 2D (Z_{2})^{m} SPTO with fractional symmetry can be utilized for universal quantum computation using only Pauli measurements, as long as it belongs to a nontrivial 2D SPTO phase. Our infinite family of fixed-point states may serve as a base model to demonstrate the idea of a "quantum computational phase" of matter, whose states share universal computational complexity ubiquitously.
Methodology for extracting local constants from petroleum cracking flows
Chang, Shen-Lin; Lottes, Steven A.; Zhou, Chenn Q.
2000-01-01
A methodology provides for the extraction of local chemical kinetic model constants for use in a reacting flow computational fluid dynamics (CFD) computer code with chemical kinetic computations to optimize the operating conditions or design of the system, including retrofit design improvements to existing systems. The coupled CFD and kinetic computer code are used in combination with data obtained from a matrix of experimental tests to extract the kinetic constants. Local fluid dynamic effects are implicitly included in the extracted local kinetic constants for each particular application system to which the methodology is applied. The extracted local kinetic model constants work well over a fairly broad range of operating conditions for specific and complex reaction sets in specific and complex reactor systems. While disclosed in terms of use in a Fluid Catalytic Cracking (FCC) riser, the inventive methodology has application in virtually any reaction set to extract constants for any particular application and reaction set formulation. The methodology includes the step of: (1) selecting the test data sets for various conditions; (2) establishing the general trend of the parametric effect on the measured product yields; (3) calculating product yields for the selected test conditions using coupled computational fluid dynamics and chemical kinetics; (4) adjusting the local kinetic constants to match calculated product yields with experimental data; and (5) validating the determined set of local kinetic constants by comparing the calculated results with experimental data from additional test runs at different operating conditions.
Numerical estimation of cavitation intensity
NASA Astrophysics Data System (ADS)
Krumenacker, L.; Fortes-Patella, R.; Archer, A.
2014-03-01
Cavitation may appear in turbomachinery and in hydraulic orifices, venturis or valves, leading to performance losses, vibrations and material erosion. This study propose a new method to predict the cavitation intensity of the flow, based on a post-processing of unsteady CFD calculations. The paper presents the analyses of cavitating structures' evolution at two different scales: • A macroscopic one in which the growth of cavitating structures is calculated using an URANS software based on a homogeneous model. Simulations of cavitating flows are computed using a barotropic law considering presence of air and interfacial tension, and Reboud's correction on the turbulence model. • Then a small one where a Rayleigh-Plesset software calculates the acoustic energy generated by the implosion of the vapor/gas bubbles with input parameters from macroscopic scale. The volume damage rate of the material during incubation time is supposed to be a part of the cumulated acoustic energy received by the solid wall. The proposed analysis method is applied to calculations on hydrofoil and orifice geometries. Comparisons between model results and experimental works concerning flow characteristic (size of cavity, pressure,velocity) as well as pitting (erosion area, relative cavitation intensity) are presented.
NASA Astrophysics Data System (ADS)
Pond, Ian; Edabi, Alireza; Dubief, Yves; White, Christopher
2015-11-01
Reynolds Average Navier Stokes (RANS) modeling has established itself as a critical design tool in many engineering applications, thanks to its superior computational efficiency. The drawbacks of RANS models are well known, but not necessarily well understood: poor prediction of transition, non equilibrium flows, mixing and heat transfer, to name the ones relevant to our study. In the present study, we use a DNS of a reciprocating channel flow driven by an oscillating pressure gradient to test several low- and high-Reynolds RANS models. Temperature is introduced as a passive scalar to study heat transfer modeling. Low-Reynolds models manage to capture the overall physics of wall shear and heat flux well, yet with some phase discrepancies, whereas high Reynolds models fail. Under the microscope of the integral method for wall shear and wall heat flux, the qualitative agreement appears more serendipitous than driven by the ability of the models to capture the correct physics. The integral method is shown to be more insightful in the benchmarking of RANS models than the typical comparisons of statistical quantities. The authors acknowledges the support of NSF and DOE under grant NSF/DOE 1258697 (VT) and 1258702 (NH).
Method and apparatus for transfer function simulator for testing complex systems
NASA Technical Reports Server (NTRS)
Kavaya, M. J. (Inventor)
1985-01-01
A method and apparatus for testing the operation of a complex stabilization circuit in a closed loop system is presented. The method is comprised of a programmed analog or digital computing system for implementing the transfer function of a load thereby providing a predictable load. The digital computing system employs a table stored in a microprocessor in which precomputed values of the load transfer function are stored for values of input signal from the stabilization circuit over the range of interest. This technique may be used not only for isolating faults in the stabilization circuit, but also for analyzing a fault in a faulty load by so varying parameters of the computing system as to simulate operation of the actual load with the fault.
Computer program for determining mass properties of a rigid structure
NASA Technical Reports Server (NTRS)
Hull, R. A.; Gilbert, J. L.; Klich, P. J.
1978-01-01
A computer program was developed for the rapid computation of the mass properties of complex structural systems. The program uses rigid body analyses and permits differences in structural material throughout the total system. It is based on the premise that complex systems can be adequately described by a combination of basic elemental shapes. Simple geometric data describing size and location of each element and the respective material density or weight of each element were the only required input data. From this minimum input, the program yields system weight, center of gravity, moments of inertia and products of inertia with respect to mutually perpendicular axes through the system center of gravity. The program also yields mass properties of the individual shapes relative to component axes.
Sensitivity analysis of dynamic biological systems with time-delays.
Wu, Wu Hsiung; Wang, Feng Sheng; Chang, Maw Shang
2010-10-15
Mathematical modeling has been applied to the study and analysis of complex biological systems for a long time. Some processes in biological systems, such as the gene expression and feedback control in signal transduction networks, involve a time delay. These systems are represented as delay differential equation (DDE) models. Numerical sensitivity analysis of a DDE model by the direct method requires the solutions of model and sensitivity equations with time-delays. The major effort is the computation of Jacobian matrix when computing the solution of sensitivity equations. The computation of partial derivatives of complex equations either by the analytic method or by symbolic manipulation is time consuming, inconvenient, and prone to introduce human errors. To address this problem, an automatic approach to obtain the derivatives of complex functions efficiently and accurately is necessary. We have proposed an efficient algorithm with an adaptive step size control to compute the solution and dynamic sensitivities of biological systems described by ordinal differential equations (ODEs). The adaptive direct-decoupled algorithm is extended to solve the solution and dynamic sensitivities of time-delay systems describing by DDEs. To save the human effort and avoid the human errors in the computation of partial derivatives, an automatic differentiation technique is embedded in the extended algorithm to evaluate the Jacobian matrix. The extended algorithm is implemented and applied to two realistic models with time-delays: the cardiovascular control system and the TNF-α signal transduction network. The results show that the extended algorithm is a good tool for dynamic sensitivity analysis on DDE models with less user intervention. By comparing with direct-coupled methods in theory, the extended algorithm is efficient, accurate, and easy to use for end users without programming background to do dynamic sensitivity analysis on complex biological systems with time-delays.
Azzopardi, George; Petkov, Nicolai
2014-01-01
The remarkable abilities of the primate visual system have inspired the construction of computational models of some visual neurons. We propose a trainable hierarchical object recognition model, which we call S-COSFIRE (S stands for Shape and COSFIRE stands for Combination Of Shifted FIlter REsponses) and use it to localize and recognize objects of interests embedded in complex scenes. It is inspired by the visual processing in the ventral stream (V1/V2 → V4 → TEO). Recognition and localization of objects embedded in complex scenes is important for many computer vision applications. Most existing methods require prior segmentation of the objects from the background which on its turn requires recognition. An S-COSFIRE filter is automatically configured to be selective for an arrangement of contour-based features that belong to a prototype shape specified by an example. The configuration comprises selecting relevant vertex detectors and determining certain blur and shift parameters. The response is computed as the weighted geometric mean of the blurred and shifted responses of the selected vertex detectors. S-COSFIRE filters share similar properties with some neurons in inferotemporal cortex, which provided inspiration for this work. We demonstrate the effectiveness of S-COSFIRE filters in two applications: letter and keyword spotting in handwritten manuscripts and object spotting in complex scenes for the computer vision system of a domestic robot. S-COSFIRE filters are effective to recognize and localize (deformable) objects in images of complex scenes without requiring prior segmentation. They are versatile trainable shape detectors, conceptually simple and easy to implement. The presented hierarchical shape representation contributes to a better understanding of the brain and to more robust computer vision algorithms. PMID:25126068
Iterative Demodulation and Decoding of Non-Square QAM
NASA Technical Reports Server (NTRS)
Li, Lifang; Divsalar, Dariush; Dolinar, Samuel
2004-01-01
It has been shown that a non-square (NS) 2(sup 2n+1)-ary (where n is a positive integer) quadrature amplitude modulation [(NS)2(sup 2n+1)-QAM] has inherent memory that can be exploited to obtain coding gains. Moreover, it should not be necessary to build new hardware to realize these gains. The present scheme is a product of theoretical calculations directed toward reducing the computational complexity of decoding coded 2(sup 2n+1)-QAM. In the general case of 2(sup 2n+1)-QAM, the signal constellation is not square and it is impossible to have independent in-phase (I) and quadrature-phase (Q) mapping and demapping. However, independent I and Q mapping and demapping are desirable for reducing the complexity of computing the log likelihood ratio (LLR) between a bit and a received symbol (such computations are essential operations in iterative decoding). This is because in modulation schemes that include independent I and Q mapping and demapping, each bit of a signal point is involved in only one-dimensional mapping and demapping. As a result, the computation of the LLR is equivalent to that of a one-dimensional pulse amplitude modulation (PAM) system. Therefore, it is desirable to find a signal constellation that enables independent I and Q mapping and demapping for 2(sup 2n+1)-QAM.
Condor-COPASI: high-throughput computing for biochemical networks
2012-01-01
Background Mathematical modelling has become a standard technique to improve our understanding of complex biological systems. As models become larger and more complex, simulations and analyses require increasing amounts of computational power. Clusters of computers in a high-throughput computing environment can help to provide the resources required for computationally expensive model analysis. However, exploiting such a system can be difficult for users without the necessary expertise. Results We present Condor-COPASI, a server-based software tool that integrates COPASI, a biological pathway simulation tool, with Condor, a high-throughput computing environment. Condor-COPASI provides a web-based interface, which makes it extremely easy for a user to run a number of model simulation and analysis tasks in parallel. Tasks are transparently split into smaller parts, and submitted for execution on a Condor pool. Result output is presented to the user in a number of formats, including tables and interactive graphical displays. Conclusions Condor-COPASI can effectively use a Condor high-throughput computing environment to provide significant gains in performance for a number of model simulation and analysis tasks. Condor-COPASI is free, open source software, released under the Artistic License 2.0, and is suitable for use by any institution with access to a Condor pool. Source code is freely available for download at http://code.google.com/p/condor-copasi/, along with full instructions on deployment and usage. PMID:22834945
Practical Measurement of Complexity In Dynamic Systems
2012-01-01
policies that produce highly complex behaviors , yet yield no benefit. 21Jason B. Clark and David R. Jacques / Procedia Computer Science 8 (2012) 14... Procedia Computer Science 8 (2012) 14 – 21 1877-0509 © 2012 Published by Elsevier B.V. doi:10.1016/j.procs.2012.01.008 Available online at...www.sciencedirect.com Procedia Computer Science Procedia Computer Science 00 (2012) 000–000 www.elsevier.com/locate/ procedia Available online at
Computer-aided programming for message-passing system; Problems and a solution
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wu, M.Y.; Gajski, D.D.
1989-12-01
As the number of processors and the complexity of problems to be solved increase, programming multiprocessing systems becomes more difficult and error-prone. Program development tools are necessary since programmers are not able to develop complex parallel programs efficiently. Parallel models of computation, parallelization problems, and tools for computer-aided programming (CAP) are discussed. As an example, a CAP tool that performs scheduling and inserts communication primitives automatically is described. It also generates the performance estimates and other program quality measures to help programmers in improving their algorithms and programs.
Radio Jove: Citizen Science for Jupiter Radio Astronomy
NASA Astrophysics Data System (ADS)
Higgins, C. A.; Thieman, J.; Reyes, F. J.; Typinski, D.; Flagg, R. F.; Greenman, W.; Brown, J.; Ashcraft, T.; Sky, J.; Cecconi, B.; Garcia, L. N.
2016-12-01
The Radio Jove Project (http://radiojove.gsfc.nasa.gov) has been operating as an educational activity for 18 years to introduce radio astronomy activities to students, teachers, and the general public. Participants may build a simple radio telescope kit, make scientific observations, and interact with radio observatories in real-time over the Internet. Recently some of our dedicated citizen science observers have upgraded their systems to better study radio emission from Jupiter and the Sun by adding dual-polarization spectrographs and wide-band antennas in the frequency range of 15-30 MHz. Some of these observations are being used in conjunction with professional telescopes such as the Long Wavelength Array (LWA), the Nancay Decametric Array, and the Ukrainian URAN2 Radio Telescope. In particular, there is an effort to support the Juno Mission radio waves instrument at Jupiter by using citizen science ground-based data for comparison and polarization verification. These data will be archived through a Virtual European Solar and Planetary Access (VESPA) archive (https://voparis-radiojove.obspm.fr/radiojove/welcome) for use by the amateur and professional radio science community. We overview the program and display recent observations that will be of interest to the science community.
NASA Astrophysics Data System (ADS)
Rundle, J.; Rundle, P.; Donnellan, A.; Li, P.
2003-12-01
We consider the problem of the complex dynamics of earthquake fault systems, and whether numerical simulations can be used to define an ensemble forecasting technology similar to that used in weather and climate research. To effectively carry out such a program, we need 1) a topological realistic model to simulate the fault system; 2) data sets to constrain the model parameters through a systematic program of data assimilation; 3) a computational technology making use of modern paradigms of high performance and parallel computing systems; and 4) software to visualize and analyze the results. In particular, we focus attention of a new version of our code Virtual California (version 2001) in which we model all of the major strike slip faults extending throughout California, from the Mexico-California border to the Mendocino Triple Junction. We use the historic data set of earthquakes larger than magnitude M > 6 to define the frictional properties of all 654 fault segments (degrees of freedom) in the model. Previous versions of Virtual California had used only 215 fault segments to model the strike slip faults in southern California. To compute the dynamics and the associated surface deformation, we use message passing as implemented in the MPICH standard distribution on a small Beowulf cluster consisting of 10 cpus. We are also planning to run the code on significantly larger machines so that we can begin to examine much finer spatial scales of resolution, and to assess scaling properties of the code. We present results of simulations both as static images and as mpeg movies, so that the dynamical aspects of the computation can be assessed by the viewer. We also compute a variety of statistics from the simulations, including magnitude-frequency relations, and compare these with data from real fault systems.
NASA Astrophysics Data System (ADS)
Myint, L. M. M.; Warisarn, C.
2017-05-01
Two-dimensional (2-D) interference is one of the prominent challenges in ultra-high density recording system such as bit patterned media recording (BPMR). The multi-track joint 2-D detection technique with the help of the array-head reading can tackle this problem effectively by jointly processing the multiple readback signals from the adjacent tracks. Moreover, it can robustly alleviate the impairments due to track mis-registration (TMR) and media noise. However, the computational complexity of such detectors is normally too high and hard to implement in a reality, even for a few multiple tracks. Therefore, in this paper, we mainly focus on reducing the complexity of multi-track joint 2-D Viterbi detector without paying a large penalty in terms of the performance. We propose a simplified multi-track joint 2-D Viterbi detector with a manageable complexity level for the BPMR's multi-track multi-head (MTMH) system. In the proposed method, the complexity of detector's trellis is reduced with the help of the joint-track equalization method which employs 1-D equalizers and 2-D generalized partial response (GPR) target. Moreover, we also examine the performance of a full-fledged multi-track joint 2-D detector and the conventional 2-D detection. The results show that the simplified detector can perform close to the full-fledge detector, especially when the system faces high media noise, with the significant low complexity.
NASA Astrophysics Data System (ADS)
Mhashilkar, Parag; Tiradani, Anthony; Holzman, Burt; Larson, Krista; Sfiligoi, Igor; Rynge, Mats
2014-06-01
Scientific communities have been in the forefront of adopting new technologies and methodologies in the computing. Scientific computing has influenced how science is done today, achieving breakthroughs that were impossible to achieve several decades ago. For the past decade several such communities in the Open Science Grid (OSG) and the European Grid Infrastructure (EGI) have been using GlideinWMS to run complex application workflows to effectively share computational resources over the grid. GlideinWMS is a pilot-based workload management system (WMS) that creates on demand, a dynamically sized overlay HTCondor batch system on grid resources. At present, the computational resources shared over the grid are just adequate to sustain the computing needs. We envision that the complexity of the science driven by "Big Data" will further push the need for computational resources. To fulfill their increasing demands and/or to run specialized workflows, some of the big communities like CMS are investigating the use of cloud computing as Infrastructure-As-A-Service (IAAS) with GlideinWMS as a potential alternative to fill the void. Similarly, communities with no previous access to computing resources can use GlideinWMS to setup up a batch system on the cloud infrastructure. To enable this, the architecture of GlideinWMS has been extended to enable support for interfacing GlideinWMS with different Scientific and commercial cloud providers like HLT, FutureGrid, FermiCloud and Amazon EC2. In this paper, we describe a solution for cloud bursting with GlideinWMS. The paper describes the approach, architectural changes and lessons learned while enabling support for cloud infrastructures in GlideinWMS.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mhashilkar, Parag; Tiradani, Anthony; Holzman, Burt
Scientific communities have been in the forefront of adopting new technologies and methodologies in the computing. Scientific computing has influenced how science is done today, achieving breakthroughs that were impossible to achieve several decades ago. For the past decade several such communities in the Open Science Grid (OSG) and the European Grid Infrastructure (EGI) have been using GlideinWMS to run complex application workflows to effectively share computational resources over the grid. GlideinWMS is a pilot-based workload management system (WMS) that creates on demand, a dynamically sized overlay HTCondor batch system on grid resources. At present, the computational resources shared overmore » the grid are just adequate to sustain the computing needs. We envision that the complexity of the science driven by 'Big Data' will further push the need for computational resources. To fulfill their increasing demands and/or to run specialized workflows, some of the big communities like CMS are investigating the use of cloud computing as Infrastructure-As-A-Service (IAAS) with GlideinWMS as a potential alternative to fill the void. Similarly, communities with no previous access to computing resources can use GlideinWMS to setup up a batch system on the cloud infrastructure. To enable this, the architecture of GlideinWMS has been extended to enable support for interfacing GlideinWMS with different Scientific and commercial cloud providers like HLT, FutureGrid, FermiCloud and Amazon EC2. In this paper, we describe a solution for cloud bursting with GlideinWMS. The paper describes the approach, architectural changes and lessons learned while enabling support for cloud infrastructures in GlideinWMS.« less
A technique for computation of noise temperature due to a beam waveguide shroud
NASA Technical Reports Server (NTRS)
Veruttipong, W.; Franco, M. M.
1993-01-01
Direct analytical computation of the noise temperature of real beam waveguide (BWG) systems, including all mirrors and the surrounding shroud, is an extremely complex problem and virtually impossible to achieve. Yet the DSN antennas are required to be ultra low-noise in order to be effective, and a reasonably accurate prediction is essential. This article presents a relatively simple technique to compute a real BWG system noise temperature by combining analytical techniques with data from experimental tests. Specific expressions and parameters for X-band (8.45-GHz) BWG noise computation are obtained for DSS 13 and DSS 24, now under construction. These expressions are also valid for various conditions of the BWG feed systems, including horn sizes and positions, and mirror sizes, curvatures, and positions. Parameters for S- and Ka-bands (2.3 and 32.0 GHz) have not been determined; however, those can be obtained following the same procedure as for X-band.
The "Biologically-Inspired Computing" Column
NASA Technical Reports Server (NTRS)
Hinchey, Mike
2007-01-01
Self-managing systems, whether viewed from the perspective of Autonomic Computing, or from that of another initiative, offers a holistic vision for the development and evolution of biologically-inspired computer-based systems. It aims to bring new levels of automation and dependability to systems, while simultaneously hiding their complexity and reducing costs. A case can certainly be made that all computer-based systems should exhibit autonomic properties [6], and we envisage greater interest in, and uptake of, autonomic principles in future system development.
Discovering Tradeoffs, Vulnerabilities, and Dependencies within Water Resources Systems
NASA Astrophysics Data System (ADS)
Reed, P. M.
2015-12-01
There is a growing recognition and interest in using emerging computational tools for discovering the tradeoffs that emerge across complex combinations infrastructure options, adaptive operations, and sign posts. As a field concerned with "deep uncertainties", it is logically consistent to include a more direct acknowledgement that our choices for dealing with computationally demanding simulations, advanced search algorithms, and sensitivity analysis tools are themselves subject to failures that could adversely bias our understanding of how systems' vulnerabilities change with proposed actions. Balancing simplicity versus complexity in our computational frameworks is nontrivial given that we are often exploring high impact irreversible decisions. It is not always clear that accepted models even encompass important failure modes. Moreover as they become more complex and computationally demanding the benefits and consequences of simplifications are often untested. This presentation discusses our efforts to address these challenges through our "many-objective robust decision making" (MORDM) framework for the design and management water resources systems. The MORDM framework has four core components: (1) elicited problem conception and formulation, (2) parallel many-objective search, (3) interactive visual analytics, and (4) negotiated selection of robust alternatives. Problem conception and formulation is the process of abstracting a practical design problem into a mathematical representation. We build on the emerging work in visual analytics to exploit interactive visualization of both the design space and the objective space in multiple heterogeneous linked views that permit exploration and discovery. Many-objective search produces tradeoff solutions from potentially competing problem formulations that can each consider up to ten conflicting objectives based on current computational search capabilities. Negotiated design selection uses interactive visualization, reformulation, and optimization to discover desirable designs for implementation. Multi-city urban water supply portfolio planning will be used to illustrate the MORDM framework.
Integrated modeling tool for performance engineering of complex computer systems
NASA Technical Reports Server (NTRS)
Wright, Gary; Ball, Duane; Hoyt, Susan; Steele, Oscar
1989-01-01
This report summarizes Advanced System Technologies' accomplishments on the Phase 2 SBIR contract NAS7-995. The technical objectives of the report are: (1) to develop an evaluation version of a graphical, integrated modeling language according to the specification resulting from the Phase 2 research; and (2) to determine the degree to which the language meets its objectives by evaluating ease of use, utility of two sets of performance predictions, and the power of the language constructs. The technical approach followed to meet these objectives was to design, develop, and test an evaluation prototype of a graphical, performance prediction tool. The utility of the prototype was then evaluated by applying it to a variety of test cases found in the literature and in AST case histories. Numerous models were constructed and successfully tested. The major conclusion of this Phase 2 SBIR research and development effort is that complex, real-time computer systems can be specified in a non-procedural manner using combinations of icons, windows, menus, and dialogs. Such a specification technique provides an interface that system designers and architects find natural and easy to use. In addition, PEDESTAL's multiview approach provides system engineers with the capability to perform the trade-offs necessary to produce a design that meets timing performance requirements. Sample system designs analyzed during the development effort showed that models could be constructed in a fraction of the time required by non-visual system design capture tools.
Computed Tomography Inspection and Analysis for Additive Manufacturing Components
NASA Technical Reports Server (NTRS)
Beshears, Ronald D.
2016-01-01
Computed tomography (CT) inspection was performed on test articles additively manufactured from metallic materials. Metallic AM and machined wrought alloy test articles with programmed flaws were inspected using a 2MeV linear accelerator based CT system. Performance of CT inspection on identically configured wrought and AM components and programmed flaws was assessed using standard image analysis techniques to determine the impact of additive manufacturing on inspectability of objects with complex geometries.
Modeling And Simulation Of Bar Code Scanners Using Computer Aided Design Software
NASA Astrophysics Data System (ADS)
Hellekson, Ron; Campbell, Scott
1988-06-01
Many optical systems have demanding requirements to package the system in a small 3 dimensional space. The use of computer graphic tools can be a tremendous aid to the designer in analyzing the optical problems created by smaller and less costly systems. The Spectra Physics grocery store bar code scanner employs an especially complex 3 dimensional scan pattern to read bar code labels. By using a specially written program which interfaces with a computer aided design system, we have simulated many of the functions of this complex optical system. In this paper we will illustrate how a recent version of the scanner has been designed. We will discuss the use of computer graphics in the design process including interactive tweaking of the scan pattern, analysis of collected light, analysis of the scan pattern density, and analysis of the manufacturing tolerances used to build the scanner.
NASA Astrophysics Data System (ADS)
Mondal, Apurba Sau; Jana, Mahendra Sekhar; Manna, Chandan Kumar; Naskar, Rahul; Mondal, Tapan Kumar
2018-07-01
A new zinc(II) complex, [Zn(L)](ClO4) with hexadentate N4S2 donor azo-thioether ligand (HL) was synthesized and characterized by several spectroscopic techniques. The structure was confirmed by single crystal X-ray analysis. The interaction of the complex with CT DNA was investigated by UV-vis method and binding constant is found to be 6.6 × 104 M-1. Competitive binding titration with ethidium bromide (EB) by fluorescence titration method reveals that the complex efficiently displaces EB from EB-DNA system and the Stern-Volmer dynamic quenching constant, Ksv is found to be 2.6 × 104 M-1. DFT and TDDFT calculations were carried out to interpret the electronic structure and electronic spectra of the complex.
Lazareva, Svetlana; Ismagilov, Zinfer; Kuznetsov, Vadim; Shikina, Nadezhda; Kerzhentsev, Mikhail
2018-02-05
Huge amounts of nuclear waste, including depleted uranium, significantly contribute to the adverse environmental situation throughout the world. An approach to the effective use of uranium oxides in catalysts for the deep oxidation of chlorine-containing hydrocarbons is suggested. Investigation of the catalytic activity of the synthesized supported uranium oxide catalysts doped with Cr, Mn and Co transition metals in the chlorobenzene oxidation showed that these catalysts are comparable with conventional commercial ones. Physicochemical properties of the catalysts were studied by X-ray diffraction, temperature-programmed reduction with hydrogen (H 2 -TPR), and Fourier transform infrared spectroscopy. The higher activity of Mn- and Co-containing uranium oxide catalysts in the H 2 -TPR and oxidation of chlorobenzene in comparison with non-uranium catalysts may be related to the formation of a new disperse phase represented by uranates. The study of chlorobenzene adsorption revealed that the surface oxygen is involved in the catalytic process.
Unsteady flow sensing and optimal sensor placement using machine learning
NASA Astrophysics Data System (ADS)
Semaan, Richard
2016-11-01
Machine learning is used to estimate the flow state and to determine the optimal sensor placement over a two-dimensional (2D) airfoil equipped with a Coanda actuator. The analysis is based on flow field data obtained from 2D unsteady Reynolds averaged Navier-Stokes (uRANS) simulations with different jet blowing intensities and actuation frequencies, characterizing different flow separation states. This study shows how the "random forests" algorithm is utilized beyond its typical usage in fluid mechanics estimating the flow state to determine the optimal sensor placement. The results are compared against the current de-facto standard of maximum modal amplitude location and against a brute force approach that scans all possible sensor combinations. The results show that it is possible to simultaneously infer the state of flow and to determine the optimal sensor location without the need to perform proper orthogonal decomposition. Collaborative Research Center (CRC) 880, DFG.
Impurity characterization of magnesium diuranate using simultaneous TG-DTA-FTIR measurements
NASA Astrophysics Data System (ADS)
Raje, Naina; Ghonge, Darshana K.; Hemantha Rao, G. V. S.; Reddy, A. V. R.
2013-05-01
Current studies describe the application of simultaneous thermogravimetry-differential thermal analysis - evolved gas analysis techniques for the compositional characterization of magnesium diuranate (MDU) with respect to the impurities present in the matrix. The stoichiometric composition of MDU was identified as MgU2O7ṡ3H2O. Presence of carbonate and sulphate as impurities in the matrix was confirmed through the evolved gas analysis using Fourier Transformation Infrared Spectrometry detection. Carbon and magnesium hydroxide content present as impurities in magnesium diuranate have been determined quantitatively using TG and FTIR techniques and the results are in good agreement. Powder X-ray diffraction analysis of magnesium diuranate suggests the presence of magnesium hydroxide as impurity in the matrix. Also these studies confirm the formation of magnesium uranate, uranium sesquioxide and uranium dioxide above 1000 °C, due to the decomposition of magnesium diuranate.
An introduction to the spectrum, symmetries, and dynamics of spin-1/2 Heisenberg chains
NASA Astrophysics Data System (ADS)
Joel, Kira; Kollmar, Davida; Santos, Lea F.
2013-06-01
Quantum spin chains are prototype quantum many-body systems that are employed in the description of various complex physical phenomena. We provide an introduction to this subject by focusing on the time evolution of a Heisenberg spin-1/2 chain and interpreting the results based on the analysis of the eigenvalues, eigenstates, and symmetries of the system. We make available online all computer codes used to obtain our data.
Chakraborty, Saumen; Reed, Julian; Ross, Matthew; Nilges, Mark J; Petrik, Igor D; Ghosh, Soumya; Hammes-Schiffer, Sharon; Sage, J Timothy; Zhang, Yong; Schulz, Charles E; Lu, Yi
2014-02-24
A major barrier to understanding the mechanism of nitric oxide reductases (NORs) is the lack of a selective probe of NO binding to the nonheme FeB center. By replacing the heme in a biosynthetic model of NORs, which structurally and functionally mimics NORs, with isostructural ZnPP, the electronic structure and functional properties of the FeB nitrosyl complex was probed. This approach allowed observation of the first S=3/2 nonheme {FeNO}(7) complex in a protein-based model system of NOR. Detailed spectroscopic and computational studies show that the electronic state of the {FeNO}(7) complex is best described as a high spin ferrous iron (S=2) antiferromagnetically coupled to an NO radical (S=1/2) [Fe(2+)-NO(.)]. The radical nature of the FeB -bound NO would facilitate N-N bond formation by radical coupling with the heme-bound NO. This finding, therefore, supports the proposed trans mechanism of NO reduction by NORs. Copyright © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Complexation and phase evolution at dimethylformamide-Ag(111) interfaces
DOE Office of Scientific and Technical Information (OSTI.GOV)
Song, Wentao; Leung, Kevin; Shao, Qian
The interaction of solvent molecules with metallic surfaces impacts many interfacial chemical processes. We investigate the chemical and structure evolution that follows adsorption of the polar solvent dimethylformamide (DMF) on Ag(111). An Ag(DMF) 2 coordination complex forms spontaneously by DMF etching of Ag(111), yielding mixed films of the complexes and DMF. Utilizing ultrahigh vacuum scanning tunneling microscopy (UHV-STM), in combination with X-ray photoelectron spectroscopy (XPS) and density functional theory (DFT) computations, we map monolayer phases from the 2-D gas regime, consisting of a binary mixture of DMF and Ag(DMF) 2, through the saturation monolayer limit, in which these two chemicalmore » species phase separate into ordered islands. Structural models for the near-square DMF phase and the chain-like Ag(DMF) 2 phase are presented and supported by DFT computation. Interface evolution is summarized in a surface pressure-composition phase diagram, which allows structure prediction over arbitrary experimental conditions. In conclusion, this work reveals new surface coordination chemistry for an important electrolyte-electrode system, and illustrates how surface pressure can be used to tune monolayer phases.« less
Complexation and phase evolution at dimethylformamide-Ag(111) interfaces
Song, Wentao; Leung, Kevin; Shao, Qian; ...
2016-09-15
The interaction of solvent molecules with metallic surfaces impacts many interfacial chemical processes. We investigate the chemical and structure evolution that follows adsorption of the polar solvent dimethylformamide (DMF) on Ag(111). An Ag(DMF) 2 coordination complex forms spontaneously by DMF etching of Ag(111), yielding mixed films of the complexes and DMF. Utilizing ultrahigh vacuum scanning tunneling microscopy (UHV-STM), in combination with X-ray photoelectron spectroscopy (XPS) and density functional theory (DFT) computations, we map monolayer phases from the 2-D gas regime, consisting of a binary mixture of DMF and Ag(DMF) 2, through the saturation monolayer limit, in which these two chemicalmore » species phase separate into ordered islands. Structural models for the near-square DMF phase and the chain-like Ag(DMF) 2 phase are presented and supported by DFT computation. Interface evolution is summarized in a surface pressure-composition phase diagram, which allows structure prediction over arbitrary experimental conditions. In conclusion, this work reveals new surface coordination chemistry for an important electrolyte-electrode system, and illustrates how surface pressure can be used to tune monolayer phases.« less
An Approach to Experimental Design for the Computer Analysis of Complex Phenomenon
NASA Technical Reports Server (NTRS)
Rutherford, Brian
2000-01-01
The ability to make credible system assessments, predictions and design decisions related to engineered systems and other complex phenomenon is key to a successful program for many large-scale investigations in government and industry. Recently, many of these large-scale analyses have turned to computational simulation to provide much of the required information. Addressing specific goals in the computer analysis of these complex phenomenon is often accomplished through the use of performance measures that are based on system response models. The response models are constructed using computer-generated responses together with physical test results where possible. They are often based on probabilistically defined inputs and generally require estimation of a set of response modeling parameters. As a consequence, the performance measures are themselves distributed quantities reflecting these variabilities and uncertainties. Uncertainty in the values of the performance measures leads to uncertainties in predicted performance and can cloud the decisions required of the analysis. A specific goal of this research has been to develop methodology that will reduce this uncertainty in an analysis environment where limited resources and system complexity together restrict the number of simulations that can be performed. An approach has been developed that is based on evaluation of the potential information provided for each "intelligently selected" candidate set of computer runs. Each candidate is evaluated by partitioning the performance measure uncertainty into two components - one component that could be explained through the additional computational simulation runs and a second that would remain uncertain. The portion explained is estimated using a probabilistic evaluation of likely results for the additional computational analyses based on what is currently known about the system. The set of runs indicating the largest potential reduction in uncertainty is then selected and the computational simulations are performed. Examples are provided to demonstrate this approach on small scale problems. These examples give encouraging results. Directions for further research are indicated.
2015-01-01
Economies are instances of complex socio-technical systems that are shaped by the interactions of large numbers of individuals. The individual behavior and decision-making of consumer agents is determined by complex psychological dynamics that include their own assessment of present and future economic conditions as well as those of others, potentially leading to feedback loops that affect the macroscopic state of the economic system. We propose that the large-scale interactions of a nation's citizens with its online resources can reveal the complex dynamics of their collective psychology, including their assessment of future system states. Here we introduce a behavioral index of Chinese Consumer Confidence (C3I) that computationally relates large-scale online search behavior recorded by Google Trends data to the macroscopic variable of consumer confidence. Our results indicate that such computational indices may reveal the components and complex dynamics of consumer psychology as a collective socio-economic phenomenon, potentially leading to improved and more refined economic forecasting. PMID:25826692
Dong, Xianlei; Bollen, Johan
2015-01-01
Economies are instances of complex socio-technical systems that are shaped by the interactions of large numbers of individuals. The individual behavior and decision-making of consumer agents is determined by complex psychological dynamics that include their own assessment of present and future economic conditions as well as those of others, potentially leading to feedback loops that affect the macroscopic state of the economic system. We propose that the large-scale interactions of a nation's citizens with its online resources can reveal the complex dynamics of their collective psychology, including their assessment of future system states. Here we introduce a behavioral index of Chinese Consumer Confidence (C3I) that computationally relates large-scale online search behavior recorded by Google Trends data to the macroscopic variable of consumer confidence. Our results indicate that such computational indices may reveal the components and complex dynamics of consumer psychology as a collective socio-economic phenomenon, potentially leading to improved and more refined economic forecasting.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dress, W.B.
Rosen's modeling relation is embedded in Popper's three worlds to provide an heuristic tool for model building and a guide for thinking about complex systems. The utility of this construct is demonstrated by suggesting a solution to the problem of pseudo science and a resolution of the famous Bohr-Einstein debates. A theory of bizarre systems is presented by an analogy with entangled particles of quantum mechanics. This theory underscores the poverty of present-day computational systems (e.g., computers) for creating complex and bizarre entities by distinguishing between mechanism and organism.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Marzouk, Youssef
Predictive simulation of complex physical systems increasingly rests on the interplay of experimental observations with computational models. Key inputs, parameters, or structural aspects of models may be incomplete or unknown, and must be developed from indirect and limited observations. At the same time, quantified uncertainties are needed to qualify computational predictions in the support of design and decision-making. In this context, Bayesian statistics provides a foundation for inference from noisy and limited data, but at prohibitive computional expense. This project intends to make rigorous predictive modeling *feasible* in complex physical systems, via accelerated and scalable tools for uncertainty quantification, Bayesianmore » inference, and experimental design. Specific objectives are as follows: 1. Develop adaptive posterior approximations and dimensionality reduction approaches for Bayesian inference in high-dimensional nonlinear systems. 2. Extend accelerated Bayesian methodologies to large-scale {\\em sequential} data assimilation, fully treating nonlinear models and non-Gaussian state and parameter distributions. 3. Devise efficient surrogate-based methods for Bayesian model selection and the learning of model structure. 4. Develop scalable simulation/optimization approaches to nonlinear Bayesian experimental design, for both parameter inference and model selection. 5. Demonstrate these inferential tools on chemical kinetic models in reacting flow, constructing and refining thermochemical and electrochemical models from limited data. Demonstrate Bayesian filtering on canonical stochastic PDEs and in the dynamic estimation of inhomogeneous subsurface properties and flow fields.« less
Technology and Transformation in Academic Libraries.
ERIC Educational Resources Information Center
Shaw, Ward
Academic library computing systems, which are among the most complex found in academic environments, now include external systems, such as online commercial search services and nationwide networks, and local systems that control and support internal operations. As librarians have realized the benefit of using computer systems to perform…
Drawert, Brian; Trogdon, Michael; Toor, Salman; Petzold, Linda; Hellander, Andreas
2016-01-01
Computational experiments using spatial stochastic simulations have led to important new biological insights, but they require specialized tools and a complex software stack, as well as large and scalable compute and data analysis resources due to the large computational cost associated with Monte Carlo computational workflows. The complexity of setting up and managing a large-scale distributed computation environment to support productive and reproducible modeling can be prohibitive for practitioners in systems biology. This results in a barrier to the adoption of spatial stochastic simulation tools, effectively limiting the type of biological questions addressed by quantitative modeling. In this paper, we present PyURDME, a new, user-friendly spatial modeling and simulation package, and MOLNs, a cloud computing appliance for distributed simulation of stochastic reaction-diffusion models. MOLNs is based on IPython and provides an interactive programming platform for development of sharable and reproducible distributed parallel computational experiments.
1991 Annual report on scientific programs: A broad research program on the sciences of complexity
DOE Office of Scientific and Technical Information (OSTI.GOV)
Not Available
1991-01-01
1991 was continued rapid growth for the Santa Fe Institute (SFI) as it broadened its interdisciplinary research into the organization, evolution and operation of complex systems and sought deeply the principles underlying their dynamic behavior. Research on complex systems--the focus of work at SFI--involves an extraordinary range of topics normally studied in seemingly disparate fields. Natural systems displaying complex behavior range upwards from proteins and DNA through cells and evolutionary systems to human societies. Research models exhibiting complexity include nonlinear equations, spin glasses, cellular automata, genetic algorithms, classifier systems, and an array of other computational models. Some of the majormore » questions facing complex systems researchers are: (1) explaining how complexity arises from the nonlinear interaction of simples components, (2) describing the mechanisms underlying high-level aggregate behavior of complex systems (such as the overt behavior of an organism, the flow of energy in an ecology, the GNP of an economy), and (3) creating a theoretical framework to enable predictions about the likely behavior of such systems in various conditions. The importance of understanding such systems in enormous: many of the most serious challenges facing humanity--e.g., environmental sustainability, economic stability, the control of disease--as well as many of the hardest scientific questions--e.g., protein folding, the distinction between self and non-self in the immune system, the nature of intelligence, the origin of life--require deep understanding of complex systems.« less
1991 Annual report on scientific programs: A broad research program on the sciences of complexity
DOE Office of Scientific and Technical Information (OSTI.GOV)
Not Available
1991-12-31
1991 was continued rapid growth for the Santa Fe Institute (SFI) as it broadened its interdisciplinary research into the organization, evolution and operation of complex systems and sought deeply the principles underlying their dynamic behavior. Research on complex systems--the focus of work at SFI--involves an extraordinary range of topics normally studied in seemingly disparate fields. Natural systems displaying complex behavior range upwards from proteins and DNA through cells and evolutionary systems to human societies. Research models exhibiting complexity include nonlinear equations, spin glasses, cellular automata, genetic algorithms, classifier systems, and an array of other computational models. Some of the majormore » questions facing complex systems researchers are: (1) explaining how complexity arises from the nonlinear interaction of simples components, (2) describing the mechanisms underlying high-level aggregate behavior of complex systems (such as the overt behavior of an organism, the flow of energy in an ecology, the GNP of an economy), and (3) creating a theoretical framework to enable predictions about the likely behavior of such systems in various conditions. The importance of understanding such systems in enormous: many of the most serious challenges facing humanity--e.g., environmental sustainability, economic stability, the control of disease--as well as many of the hardest scientific questions--e.g., protein folding, the distinction between self and non-self in the immune system, the nature of intelligence, the origin of life--require deep understanding of complex systems.« less
ADAM: analysis of discrete models of biological systems using computer algebra.
Hinkelmann, Franziska; Brandon, Madison; Guang, Bonny; McNeill, Rustin; Blekherman, Grigoriy; Veliz-Cuba, Alan; Laubenbacher, Reinhard
2011-07-20
Many biological systems are modeled qualitatively with discrete models, such as probabilistic Boolean networks, logical models, Petri nets, and agent-based models, to gain a better understanding of them. The computational complexity to analyze the complete dynamics of these models grows exponentially in the number of variables, which impedes working with complex models. There exist software tools to analyze discrete models, but they either lack the algorithmic functionality to analyze complex models deterministically or they are inaccessible to many users as they require understanding the underlying algorithm and implementation, do not have a graphical user interface, or are hard to install. Efficient analysis methods that are accessible to modelers and easy to use are needed. We propose a method for efficiently identifying attractors and introduce the web-based tool Analysis of Dynamic Algebraic Models (ADAM), which provides this and other analysis methods for discrete models. ADAM converts several discrete model types automatically into polynomial dynamical systems and analyzes their dynamics using tools from computer algebra. Specifically, we propose a method to identify attractors of a discrete model that is equivalent to solving a system of polynomial equations, a long-studied problem in computer algebra. Based on extensive experimentation with both discrete models arising in systems biology and randomly generated networks, we found that the algebraic algorithms presented in this manuscript are fast for systems with the structure maintained by most biological systems, namely sparseness and robustness. For a large set of published complex discrete models, ADAM identified the attractors in less than one second. Discrete modeling techniques are a useful tool for analyzing complex biological systems and there is a need in the biological community for accessible efficient analysis tools. ADAM provides analysis methods based on mathematical algorithms as a web-based tool for several different input formats, and it makes analysis of complex models accessible to a larger community, as it is platform independent as a web-service and does not require understanding of the underlying mathematics.
Dynamic Identification for Control of Large Space Structures
NASA Technical Reports Server (NTRS)
Ibrahim, S. R.
1985-01-01
This is a compilation of reports by the one author on one subject. It consists of the following five journal articles: (1) A Parametric Study of the Ibrahim Time Domain Modal Identification Algorithm; (2) Large Modal Survey Testing Using the Ibrahim Time Domain Identification Technique; (3) Computation of Normal Modes from Identified Complex Modes; (4) Dynamic Modeling of Structural from Measured Complex Modes; and (5) Time Domain Quasi-Linear Identification of Nonlinear Dynamic Systems.
Hernández-Valdés, Daniel; Rodríguez-Riera, Zalua; Díaz-García, Alicia; Benoist, Eric; Jáuregui-Haza, Ulises
2016-08-01
The development of novel radiopharmaceuticals for nuclear medicine based on M(CO)3 (M = Tc, Re) complexes has attracted great attention. The versatility of this core and the easy production of the fac-[M(CO)3(H2O)3](+) precursor could explain this interest. The main characteristics of these tricarbonyl complexes are the high substitution stability of the three CO ligands and the corresponding lability of the coordinated water molecules, yielding, via easy exchange of a variety of bi- and tridentate ligands, complexes xof very high kinetic stability. Here, a computational study of different tricarbonyl complexes of Re(I) and Tc(I) was performed using density functional theory. The solvent effect was simulated using the polarizable continuum model. These structures were used as a starting point to investigate the relative stabilities of tricarbonyl complexes with various tridentate ligands. These complexes included an iminodiacetic acid unit for tridentate coordination to the fac-[M(CO)3](+) moiety (M = Re, Tc), an aromatic ring system bearing a functional group (-NO2, -NH2, and -Cl) as a linking site model, and a tethering moiety (a methylene, ethylene, propylene butylene, or pentylene bridge) between the linking and coordinating sites. The optimized complexes showed geometries comparable to those inferred from X-ray data. In general, the Re complexes were more stable than the corresponding Tc complexes. Furthermore, using NH2 as the functional group, a medium length carbon chain, and ortho substitution increased complex stability. All of the bonds involving the metal center presented a closed shell interaction with dative or covalent character, and the strength of these bonds decreased in the sequence Tc-CO > Tc-O > Tc-N.
NASA Astrophysics Data System (ADS)
Sherwin, Jason
At the start of the 21st century, the topic of complexity remains a formidable challenge in engineering, science and other aspects of our world. It seems that when disaster strikes it is because some complex and unforeseen interaction causes the unfortunate outcome. Why did the financial system of the world meltdown in 2008--2009? Why are global temperatures on the rise? These questions and other ones like them are difficult to answer because they pertain to contexts that require lengthy descriptions. In other words, these contexts are complex. But we as human beings are able to observe and recognize this thing we call 'complexity'. Furthermore, we recognize that there are certain elements of a context that form a system of complex interactions---i.e., a complex system. Many researchers have even noted similarities between seemingly disparate complex systems. Do sub-atomic systems bear resemblance to weather patterns? Or do human-based economic systems bear resemblance to macroscopic flows? Where do we draw the line in their resemblance? These are the kinds of questions that are asked in complex systems research. And the ability to recognize complexity is not only limited to analytic research. Rather, there are many known examples of humans who, not only observe and recognize but also, operate complex systems. How do they do it? Is there something superhuman about these people or is there something common to human anatomy that makes it possible to fly a plane? Or to drive a bus? Or to operate a nuclear power plant? Or to play Chopin's etudes on the piano? In each of these examples, a human being operates a complex system of machinery, whether it is a plane, a bus, a nuclear power plant or a piano. What is the common thread running through these abilities? The study of situational awareness (SA) examines how people do these types of remarkable feats. It is not a bottom-up science though because it relies on finding general principles running through a host of varied human activities. Nevertheless, since it is not constrained by computational details, the study of situational awareness provides a unique opportunity to approach complex tasks of operation from an analytical perspective. In other words, with SA, we get to see how humans observe, recognize and react to complex systems on which they exert some control. Reconciling this perspective on complexity with complex systems research, it might be possible to further our understanding of complex phenomena if we can probe the anatomical mechanisms by which we, as humans, do it naturally. At this unique intersection of two disciplines, a hybrid approach is needed. So in this work, we propose just such an approach. In particular, this research proposes a computational approach to the situational awareness (SA) of complex systems. Here we propose to implement certain aspects of situational awareness via a biologically-inspired machine-learning technique called Hierarchical Temporal Memory (HTM). In doing so, we will use either simulated or actual data to create and to test computational implementations of situational awareness. This will be tested in two example contexts, one being more complex than the other. The ultimate goal of this research is to demonstrate a possible approach to analyzing and understanding complex systems. By using HTM and carefully developing techniques to analyze the SA formed from data, it is believed that this goal can be obtained.
Computer program determines gas flow rates in piping systems
NASA Technical Reports Server (NTRS)
Franke, R.
1966-01-01
Computer program calculates the steady state flow characteristics of an ideal compressible gas in a complex piping system. The program calculates the stagnation and total temperature, static and total pressure, loss factor, and forces on each element in the piping system.
Computing Systems | High-Performance Computing | NREL
investigate, build, and test models of complex phenomena or entire integrated systems-that cannot be directly observed or manipulated in the lab, or would be too expensive or time consuming. Models and visualizations
Computational Investigations on the Aerodynamics of a Generic Car Model in Proximity to a Side Wall
NASA Astrophysics Data System (ADS)
Mallapragada, Srivatsa
A moving road vehicle is subjected to many fluid interferences caused by a number of external agents apart from the vehicle itself. Vehicles moving in proximity to a side wall is an interesting aspect that has been little investigated in the literature. This is of great interest in motorsports, more specifically in NASCAR racing. The aim of this thesis is to develop a Computational Fluid Dynamics (CFD) model that can simulate the motion of a race car moving close to a side wall with an objective of understanding the influence of this side barrier on the overall aerodynamic characteristics of the vehicle, like the force and moment coefficients. Additionally, flow visualization tools are used to gain insights into the flow field and to explain the causes of the observed aerodynamic characteristics of the vehicle. This is accomplished by using a generic car model, a 25-degree slant angle Ahmed Body, in proximity to a side wall in a virtual wind tunnel where the vehicle body is allowed to move at constant velocity. This methodology is different from the traditional CFD approach where the air is blown over a stationary vehicle. The simulation process used in this thesis requires the use of a recently developed meshing methodology called the Overset mesh. All simulations were run using a commercial finite volume CFD code called StarCCM+ where the Unsteady Reynolds Averaged Navier-Stokes URANS fluid flow solver was used to model turbulence. However, the existing literature suggests that no URANS model can correctly predict the flow field around a 25-degree slant Ahmed body model; all models under-predict turbulence in the initial separated shear layer and over-predict the separation region. Subsequently, the first phase of this thesis involved the determination of a modeling methodology that can accurately predict the flow-field over a 25-degree Ahmed body. Two two-equation eddy-viscosity turbulence models, the AKN and SST preferred by many researchers for CFD simulations of massively separated flows, were tested. It turned out that only the latter with modified model coefficients was capable of reproducing the experimental results with a reasonable accuracy. Compared to the eddy viscosity CFD simulations of an isolated 25-degree slant angle Ahmed body seen in existing literature, the results presented in this thesis show significantly better correlations with experiments. The wall proximity studies show a strong influence of the presence of the wall on the overall aerodynamic characteristics of the vehicle body. When compared with the experimental studies, although both show similar trends, however, there exists a significant difference between the experimental and CFD predicted results which tend to worsen as one approaches the wall. These differences can be attributed to fact that the CFD emulation of the flow around the side-wall is more realistic compared to the experimental implementation.
Close to real life. [solving for transonic flow about lifting airfoils using supercomputers
NASA Technical Reports Server (NTRS)
Peterson, Victor L.; Bailey, F. Ron
1988-01-01
NASA's Numerical Aerodynamic Simulation (NAS) facility for CFD modeling of highly complex aerodynamic flows employs as its basic hardware two Cray-2s, an ETA-10 Model Q, an Amdahl 5880 mainframe computer that furnishes both support processing and access to 300 Gbytes of disk storage, several minicomputers and superminicomputers, and a Thinking Machines 16,000-device 'connection machine' processor. NAS, which was the first supercomputer facility to standardize operating-system and communication software on all processors, has done important Space Shuttle aerodynamics simulations and will be critical to the configurational refinement of the National Aerospace Plane and its intergrated powerplant, which will involve complex, high temperature reactive gasdynamic computations.
Verification of Space Station Secondary Power System Stability Using Design of Experiment
NASA Technical Reports Server (NTRS)
Karimi, Kamiar J.; Booker, Andrew J.; Mong, Alvin C.; Manners, Bruce
1998-01-01
This paper describes analytical methods used in verification of large DC power systems with applications to the International Space Station (ISS). Large DC power systems contain many switching power converters with negative resistor characteristics. The ISS power system presents numerous challenges with respect to system stability such as complex sources and undefined loads. The Space Station program has developed impedance specifications for sources and loads. The overall approach to system stability consists of specific hardware requirements coupled with extensive system analysis and testing. Testing of large complex distributed power systems is not practical due to size and complexity of the system. Computer modeling has been extensively used to develop hardware specifications as well as to identify system configurations for lab testing. The statistical method of Design of Experiments (DoE) is used as an analysis tool for verification of these large systems. DOE reduces the number of computer runs which are necessary to analyze the performance of a complex power system consisting of hundreds of DC/DC converters. DoE also provides valuable information about the effect of changes in system parameters on the performance of the system. DoE provides information about various operating scenarios and identification of the ones with potential for instability. In this paper we will describe how we have used computer modeling to analyze a large DC power system. A brief description of DoE is given. Examples using applications of DoE to analysis and verification of the ISS power system are provided.
A unifying framework for rigid multibody dynamics and serial and parallel computational issues
NASA Technical Reports Server (NTRS)
Fijany, Amir; Jain, Abhinandan
1989-01-01
A unifying framework for various formulations of the dynamics of open-chain rigid multibody systems is discussed. Their suitability for serial and parallel processing is assessed. The framework is based on the derivation of intrinsic, i.e., coordinate-free, equations of the algorithms which provides a suitable abstraction and permits a distinction to be made between the computational redundancy in the intrinsic and extrinsic equations. A set of spatial notation is used which allows the derivation of the various algorithms in a common setting and thus clarifies the relationships among them. The three classes of algorithms viz., O(n), O(n exp 2) and O(n exp 3) or the solution of the dynamics problem are investigated. Researchers begin with the derivation of O(n exp 3) algorithms based on the explicit computation of the mass matrix and it provides insight into the underlying basis of the O(n) algorithms. From a computational perspective, the optimal choice of a coordinate frame for the projection of the intrinsic equations is discussed and the serial computational complexity of the different algorithms is evaluated. The three classes of algorithms are also analyzed for suitability for parallel processing. It is shown that the problem belongs to the class of N C and the time and processor bounds are of O(log2/2(n)) and O(n exp 4), respectively. However, the algorithm that achieves the above bounds is not stable. Researchers show that the fastest stable parallel algorithm achieves a computational complexity of O(n) with O(n exp 4), respectively. However, the algorithm that achieves the above bounds is not stable. Researchers show that the fastest stable parallel algorithm achieves a computational complexity of O(n) with O(n exp 2) processors, and results from the parallelization of the O(n exp 3) serial algorithm.
NASA Technical Reports Server (NTRS)
Shooman, Martin L.
1991-01-01
Many of the most challenging reliability problems of our present decade involve complex distributed systems such as interconnected telephone switching computers, air traffic control centers, aircraft and space vehicles, and local area and wide area computer networks. In addition to the challenge of complexity, modern fault-tolerant computer systems require very high levels of reliability, e.g., avionic computers with MTTF goals of one billion hours. Most analysts find that it is too difficult to model such complex systems without computer aided design programs. In response to this need, NASA has developed a suite of computer aided reliability modeling programs beginning with CARE 3 and including a group of new programs such as: HARP, HARP-PC, Reliability Analysts Workbench (Combination of model solvers SURE, STEM, PAWS, and common front-end model ASSIST), and the Fault Tree Compiler. The HARP program is studied and how well the user can model systems using this program is investigated. One of the important objectives will be to study how user friendly this program is, e.g., how easy it is to model the system, provide the input information, and interpret the results. The experiences of the author and his graduate students who used HARP in two graduate courses are described. Some brief comparisons were made with the ARIES program which the students also used. Theoretical studies of the modeling techniques used in HARP are also included. Of course no answer can be any more accurate than the fidelity of the model, thus an Appendix is included which discusses modeling accuracy. A broad viewpoint is taken and all problems which occurred in the use of HARP are discussed. Such problems include: computer system problems, installation manual problems, user manual problems, program inconsistencies, program limitations, confusing notation, long run times, accuracy problems, etc.
Distributed intrusion detection system based on grid security model
NASA Astrophysics Data System (ADS)
Su, Jie; Liu, Yahui
2008-03-01
Grid computing has developed rapidly with the development of network technology and it can solve the problem of large-scale complex computing by sharing large-scale computing resource. In grid environment, we can realize a distributed and load balance intrusion detection system. This paper first discusses the security mechanism in grid computing and the function of PKI/CA in the grid security system, then gives the application of grid computing character in the distributed intrusion detection system (IDS) based on Artificial Immune System. Finally, it gives a distributed intrusion detection system based on grid security system that can reduce the processing delay and assure the detection rates.
NASA Astrophysics Data System (ADS)
Samant, Hrishikesh; Pundalik, Ashwin; D'souza, Joseph; Sheth, Hetu; Lobo, Keegan Carmo; D'souza, Kyle; Patel, Vanit
2017-02-01
The Panvel flexure is a 150-km long tectonic structure, comprising prominently seaward-dipping Deccan flood basalts, on the western Indian rifted margin. Given the active tectonic faulting beneath the Panvel flexure zone inferred from microseismicity, better structural understanding of the region is needed. The geology of Elephanta Island in the Mumbai harbour, famous for the ca. mid-6th century A.D. Hindu rock-cut caves in Deccan basalt (a UNESCO World Heritage site) is poorly known. We describe a previously unreported but well-exposed fault zone on Elephanta Island, consisting of two large faults dipping steeply east-southeast and producing easterly downthrows. Well-developed slickensides and structural measurements indicate oblique slip on both faults. The Elephanta Island fault zone may be the northern extension of the Alibag-Uran fault zone previously described. This and two other known regional faults (Nhava-Sheva and Belpada faults) indicate a progressively eastward step-faulted structure of the Panvel flexure, with the important result that the individual movements were not simply downdip but also oblique-slip and locally even rotational (as at Uran). An interesting problem is the normal faulting, block tectonics and rifting of this region of the crust for which seismological data indicate a normal thickness (up to 41.3 km). A model of asymmetric rifting by simple shear may explain this observation and the consistently landward dips of the rifted margin faults.
Bots, Pieter; Morris, Katherine; Hibberd, Rosemary; Law, Gareth T W; Mosselmans, J Frederick W; Brown, Andy P; Doutch, James; Smith, Andrew J; Shaw, Samuel
2014-12-09
The favored pathway for disposal of higher activity radioactive wastes is via deep geological disposal. Many geological disposal facility designs include cement in their engineering design. Over the long term, interaction of groundwater with the cement and waste will form a plume of a hyperalkaline leachate (pH 10-13), and the behavior of radionuclides needs to be constrained under these extreme conditions to minimize the environmental hazard from the wastes. For uranium, a key component of many radioactive wastes, thermodynamic modeling predicts that, at high pH, U(VI) solubility will be very low (nM or lower) and controlled by equilibrium with solid phase alkali and alkaline-earth uranates. However, the formation of U(VI) colloids could potentially enhance the mobility of U(VI) under these conditions, and characterizing the potential for formation and medium-term stability of U(VI) colloids is important in underpinning our understanding of U behavior in waste disposal. Reflecting this, we applied conventional geochemical and microscopy techniques combined with synchrotron based in situ and ex situ X-ray techniques (small-angle X-ray scattering and X-ray adsorption spectroscopy (XAS)) to characterize colloidal U(VI) nanoparticles in a synthetic cement leachate (pH > 13) containing 4.2-252 μM U(VI). The results show that in cement leachates with 42 μM U(VI), colloids formed within hours and remained stable for several years. The colloids consisted of 1.5-1.8 nm nanoparticles with a proportion forming 20-60 nm aggregates. Using XAS and electron microscopy, we were able to determine that the colloidal nanoparticles had a clarkeite (sodium-uranate)-type crystallographic structure. The presented results have clear and hitherto unrecognized implications for the mobility of U(VI) in cementitious environments, in particular those associated with the geological disposal of nuclear waste.
Zonal PANS: evaluation of different treatments of the RANS-LES interface
NASA Astrophysics Data System (ADS)
Davidson, L.
2016-03-01
The partially Reynolds-averaged Navier-Stokes (PANS) model can be used to simulate turbulent flows either as RANS, large eddy simulation (LES) or DNS. Its main parameter is fk whose physical meaning is the ratio of the modelled to the total turbulent kinetic energy. In RANS fk = 1, in DNS fk = 0 and in LES fk takes values between 0 and 1. Three different ways of prescribing fk are evaluated for decaying grid turbulence and fully developed channel flow: fk = 0.4, fk = k3/2tot/ɛ and, from its definition, fk = k/ktot where ktot is the sum of the modelled, k, and resolved, kres, turbulent kinetic energy. It is found that the fk = 0.4 gives the best results. In Girimaji and Wallin, a method was proposed to include the effect of the gradient of fk. This approach is used at RANS- LES interface in the present study. Four different interface models are evaluated in fully developed channel flow and embedded LES of channel flow: in both cases, PANS is used as a zonal model with fk = 1 in the unsteady RANS (URANS) region and fk = 0.4 in the LES region. In fully developed channel flow, the RANS- LES interface is parallel to the wall (horizontal) and in embedded LES, it is parallel to the inlet (vertical). The importance of the location of the horizontal interface in fully developed channel flow is also investigated. It is found that the location - and the choice of the treatment at the interface - may be critical at low Reynolds number or if the interface is placed too close to the wall. The reason is that the modelled turbulent shear stress at the interface is large and hence the relative strength of the resolved turbulence is small. In RANS, the turbulent viscosity - and consequently also the modelled Reynolds shear stress - is only weakly dependent on Reynolds number. It is found in the present work that it also applies in the URANS region.
Epidemic modeling in complex realities.
Colizza, Vittoria; Barthélemy, Marc; Barrat, Alain; Vespignani, Alessandro
2007-04-01
In our global world, the increasing complexity of social relations and transport infrastructures are key factors in the spread of epidemics. In recent years, the increasing availability of computer power has enabled both to obtain reliable data allowing one to quantify the complexity of the networks on which epidemics may propagate and to envision computational tools able to tackle the analysis of such propagation phenomena. These advances have put in evidence the limits of homogeneous assumptions and simple spatial diffusion approaches, and stimulated the inclusion of complex features and heterogeneities relevant in the description of epidemic diffusion. In this paper, we review recent progresses that integrate complex systems and networks analysis with epidemic modelling and focus on the impact of the various complex features of real systems on the dynamics of epidemic spreading.
Fast computational scheme of image compression for 32-bit microprocessors
NASA Technical Reports Server (NTRS)
Kasperovich, Leonid
1994-01-01
This paper presents a new computational scheme of image compression based on the discrete cosine transform (DCT), underlying JPEG and MPEG International Standards. The algorithm for the 2-d DCT computation uses integer operations (register shifts and additions / subtractions only); its computational complexity is about 8 additions per image pixel. As a meaningful example of an on-board image compression application we consider the software implementation of the algorithm for the Mars Rover (Marsokhod, in Russian) imaging system being developed as a part of Mars-96 International Space Project. It's shown that fast software solution for 32-bit microprocessors may compete with the DCT-based image compression hardware.
Modeling of a Tröger’s tweezer and its complexation properties
NASA Astrophysics Data System (ADS)
Parchaňský, Václav; Matějka, Pavel; Dolenský, Bohumil; Havlík, Martin; Bouř, Petr
2009-09-01
Molecular tweezers attracted attention because of their potential to selectively bind important chemicals, which can be utilized in medicine or in pollution treatment. In this study, the aromatic binding properties of a recently synthesized tweezer based on the Tröger's base and its complex with nitrobenzene are investigated ab initio, using the DFT and MP2 computations. The predicted geometries and energies of the complex with nitrobenzene are well comparable with the experimental data. The B3LYP and BPW91 functionals did not provide a stable binding, in contrast to the observation. Only addition of the empirical Grimme correction for the van der Waals forces, not present in conventional DFT, yielded results consistent with the experiment, MP2 computations, and similar benchmark models. The correction also caused minor improvements of the Raman and infrared spectra, but not in the entire region of vibrational frequencies. The results suggest that the role of the electrostatic interaction in the investigated complex is minor and the interaction stabilization is driven by the contact area between the polarizable aromatic systems. The vdW-DFT method thus provides an efficient tool for the rational synthesis of the complexes.
Image-Processing Software For A Hypercube Computer
NASA Technical Reports Server (NTRS)
Lee, Meemong; Mazer, Alan S.; Groom, Steven L.; Williams, Winifred I.
1992-01-01
Concurrent Image Processing Executive (CIPE) is software system intended to develop and use image-processing application programs on concurrent computing environment. Designed to shield programmer from complexities of concurrent-system architecture, it provides interactive image-processing environment for end user. CIPE utilizes architectural characteristics of particular concurrent system to maximize efficiency while preserving architectural independence from user and programmer. CIPE runs on Mark-IIIfp 8-node hypercube computer and associated SUN-4 host computer.
Statistical Field Estimation for Complex Coastal Regions and Archipelagos (PREPRINT)
2011-04-09
and study the computational properties of these schemes. Specifically, we extend a multiscale Objective Analysis (OA) approach to complex coastal...computational properties of these schemes. Specifically, we extend a multiscale Objective Analysis (OA) approach to complex coastal regions and... multiscale free-surface code builds on the primitive-equation model of the Harvard Ocean Predic- tion System (HOPS, Haley et al. (2009)). Additionally
The Computational Ecologist’s Toolbox
Computational ecology, nestled in the broader field of data science, is an interdisciplinary field that attempts to improve our understanding of complex ecological systems through the use of modern computational methods. Computational ecology is based on a union of competence in...
Computed intraoperative navigation guidance--a preliminary report on a new technique.
Enislidis, G; Wagner, A; Ploder, O; Ewers, R
1997-08-01
To assess the value of a computer-assisted three-dimensional guidance system (Virtual Patient System) in maxillofacial operations. Laboratory and open clinical study. Teaching Hospital, Austria. 6 patients undergoing various procedures including removal of foreign body (n=3) and biopsy, maxillary advancement, and insertion of implants (n=1 each). Storage of computed tomographic (CT) pictures on an optical disc, and imposition of intraoperative video images on to these. The resulting display is shown to the surgeon on a micromonitor in his head-up display for guidance during the operations. To improve orientation during complex or minimally invasive maxillofacial procedures and to make such operations easier and less traumatic. Successful transferral of computed navigation technology into an operation room environment and positive evaluation of the method by the surgeons involved. Computer-assisted three-dimensional guidance systems have the potential for making complex or minimally invasive procedures easier to do, thereby reducing postoperative morbidity.
Urbina, Angel; Mahadevan, Sankaran; Paez, Thomas L.
2012-03-01
Here, performance assessment of complex systems is ideally accomplished through system-level testing, but because they are expensive, such tests are seldom performed. On the other hand, for economic reasons, data from tests on individual components that are parts of complex systems are more readily available. The lack of system-level data leads to a need to build computational models of systems and use them for performance prediction in lieu of experiments. Because their complexity, models are sometimes built in a hierarchical manner, starting with simple components, progressing to collections of components, and finally, to the full system. Quantification of uncertainty inmore » the predicted response of a system model is required in order to establish confidence in the representation of actual system behavior. This paper proposes a framework for the complex, but very practical problem of quantification of uncertainty in system-level model predictions. It is based on Bayes networks and uses the available data at multiple levels of complexity (i.e., components, subsystem, etc.). Because epistemic sources of uncertainty were shown to be secondary, in this application, aleatoric only uncertainty is included in the present uncertainty quantification. An example showing application of the techniques to uncertainty quantification of measures of response of a real, complex aerospace system is included.« less
2011-01-01
Knowledge of the relative stabilities of alane (AlH3) complexes with electron donors is essential for identifying hydrogen storage materials for vehicular applications that can be regenerated by off-board methods; however, almost no thermodynamic data are available to make this assessment. To fill this gap, we employed the G4(MP2) method to determine heats of formation, entropies, and Gibbs free energies of formation for 38 alane complexes with NH3−nRn (R = Me, Et; n = 0−3), pyridine, pyrazine, triethylenediamine (TEDA), quinuclidine, OH2−nRn (R = Me, Et; n = 0−2), dioxane, and tetrahydrofuran (THF). Monomer, bis, and selected dimer complex geometries were considered. Using these data, we computed the thermodynamics of the key formation and dehydrogenation reactions that would occur during hydrogen delivery and alane regeneration, from which trends in complex stability were identified. These predictions were tested by synthesizing six amine−alane complexes involving trimethylamine, triethylamine, dimethylethylamine, TEDA, quinuclidine, and hexamine and obtaining upper limits of ΔG° for their formation from metallic aluminum. Combining these computational and experimental results, we establish a criterion for complex stability relevant to hydrogen storage that can be used to assess potential ligands prior to attempting synthesis of the alane complex. On the basis of this, we conclude that only a subset of the tertiary amine complexes considered and none of the ether complexes can be successfully formed by direct reaction with aluminum and regenerated in an alane-based hydrogen storage system. PMID:22962624
Capturing, Codifying and Scoring Complex Data for Innovative, Computer-Based Items.
ERIC Educational Resources Information Center
Luecht, Richard M.
The Microsoft Certification Program (MCP) includes many new computer-based item types, based on complex cases involving the Windows 2000 (registered) operating system. This Innovative Item Technology (IIT) has presented challenges beyond traditional psychometric considerations such as capturing and storing the relevant response data from…
Automated Instructional Monitors for Complex Operational Tasks. Final Report.
ERIC Educational Resources Information Center
Feurzeig, Wallace
A computer-based instructional system is described which incorporates diagnosis of students difficulties in acquiring complex concepts and skills. A computer automatically generated a simulated display. It then monitored and analyzed a student's work in the performance of assigned training tasks. Two major tasks were studied. The first,…
NASA Astrophysics Data System (ADS)
Xu, Ye; Lee, Michael C.; Boroczky, Lilla; Cann, Aaron D.; Borczuk, Alain C.; Kawut, Steven M.; Powell, Charles A.
2009-02-01
Features calculated from different dimensions of images capture quantitative information of the lung nodules through one or multiple image slices. Previously published computer-aided diagnosis (CADx) systems have used either twodimensional (2D) or three-dimensional (3D) features, though there has been little systematic analysis of the relevance of the different dimensions and of the impact of combining different dimensions. The aim of this study is to determine the importance of combining features calculated in different dimensions. We have performed CADx experiments on 125 pulmonary nodules imaged using multi-detector row CT (MDCT). The CADx system computed 192 2D, 2.5D, and 3D image features of the lesions. Leave-one-out experiments were performed using five different combinations of features from different dimensions: 2D, 3D, 2.5D, 2D+3D, and 2D+3D+2.5D. The experiments were performed ten times for each group. Accuracy, sensitivity and specificity were used to evaluate the performance. Wilcoxon signed-rank tests were applied to compare the classification results from these five different combinations of features. Our results showed that 3D image features generate the best result compared with other combinations of features. This suggests one approach to potentially reducing the dimensionality of the CADx data space and the computational complexity of the system while maintaining diagnostic accuracy.
Controls for Burning Solid Wastes
ERIC Educational Resources Information Center
Toro, Richard F.; Weinstein, Norman J.
1975-01-01
Modern thermal solid waste processing systems are becoming more complex, incorporating features that require instrumentation and control systems to a degree greater than that previously required just for proper combustion control. With the advent of complex, sophisticated, thermal processing systems, TV monitoring and computer control should…
Computational Modeling and Real-Time Control of Patient-Specific Laser Treatment of Cancer
Fuentes, D.; Oden, J. T.; Diller, K. R.; Hazle, J. D.; Elliott, A.; Shetty, A.; Stafford, R. J.
2014-01-01
An adaptive feedback control system is presented which employs a computational model of bioheat transfer in living tissue to guide, in real-time, laser treatments of prostate cancer monitored by magnetic resonance thermal imaging (MRTI). The system is built on what can be referred to as cyberinfrastructure - a complex structure of high-speed network, large-scale parallel computing devices, laser optics, imaging, visualizations, inverse-analysis algorithms, mesh generation, and control systems that guide laser therapy to optimally control the ablation of cancerous tissue. The computational system has been successfully tested on in-vivo, canine prostate. Over the course of an 18 minute laser induced thermal therapy (LITT) performed at M.D. Anderson Cancer Center (MDACC) in Houston, Texas, the computational models were calibrated to intra-operative real time thermal imaging treatment data and the calibrated models controlled the bioheat transfer to within 5°C of the predetermined treatment plan. The computational arena is in Austin, Texas and managed at the Institute for Computational Engineering and Sciences (ICES). The system is designed to control the bioheat transfer remotely while simultaneously providing real-time remote visualization of the on-going treatment. Post operative histology of the canine prostate reveal that the damage region was within the targeted 1.2cm diameter treatment objective. PMID:19148754
Computational modeling and real-time control of patient-specific laser treatment of cancer.
Fuentes, D; Oden, J T; Diller, K R; Hazle, J D; Elliott, A; Shetty, A; Stafford, R J
2009-04-01
An adaptive feedback control system is presented which employs a computational model of bioheat transfer in living tissue to guide, in real-time, laser treatments of prostate cancer monitored by magnetic resonance thermal imaging. The system is built on what can be referred to as cyberinfrastructure-a complex structure of high-speed network, large-scale parallel computing devices, laser optics, imaging, visualizations, inverse-analysis algorithms, mesh generation, and control systems that guide laser therapy to optimally control the ablation of cancerous tissue. The computational system has been successfully tested on in vivo, canine prostate. Over the course of an 18 min laser-induced thermal therapy performed at M.D. Anderson Cancer Center (MDACC) in Houston, Texas, the computational models were calibrated to intra-operative real-time thermal imaging treatment data and the calibrated models controlled the bioheat transfer to within 5 degrees C of the predetermined treatment plan. The computational arena is in Austin, Texas and managed at the Institute for Computational Engineering and Sciences (ICES). The system is designed to control the bioheat transfer remotely while simultaneously providing real-time remote visualization of the on-going treatment. Post-operative histology of the canine prostate reveal that the damage region was within the targeted 1.2 cm diameter treatment objective.
Computer-assisted spinal osteotomy: a technical note and report of four cases.
Fujibayashi, Shunsuke; Neo, Masashi; Takemoto, Mitsuru; Ota, Masato; Nakayama, Tomitaka; Toguchida, Junya; Nakamura, Takashi
2010-08-15
A report of 4 cases of spinal osteotomy performed under the guidance of a computer-assisted navigation system and a technical note about the use of the navigation system for spinal osteotomy. To document the surgical technique and usefulness of computer-assisted surgery for spinal osteotomy. A computer-assisted navigation system provides accurate 3-dimensional (3D) real-time surgical information during the operation. Although there are many reports on the accuracy and usefulness of a navigation system for pedicle screw placement, there are few reports on the application for spinal osteotomy. We report on 4 complex cases including 3 solitary malignant spinal tumors and 1 spinal kyphotic deformity of ankylosing spondylitis, which were treated surgically using a computer-assisted spinal osteotomy. The surgical technique and postoperative clinical and radiologic results are presented. 3D spinal osteotomy under the guidance of a computer-assisted navigation system was performed successfully in 4 patients. All malignant tumors were resected en bloc, and the spinal deformity was corrected precisely according to the preoperative plan. Pathologic analysis confirmed the en bloc resection without tumor exposure in the 3 patients with a spinal tumor. The use of a computer-assisted navigation system will help ensure the safety and efficacy of a complex 3D spinal osteotomy.
Chaotic dynamics in nanoscale NbO2 Mott memristors for analogue computing
NASA Astrophysics Data System (ADS)
Kumar, Suhas; Strachan, John Paul; Williams, R. Stanley
2017-08-01
At present, machine learning systems use simplified neuron models that lack the rich nonlinear phenomena observed in biological systems, which display spatio-temporal cooperative dynamics. There is evidence that neurons operate in a regime called the edge of chaos that may be central to complexity, learning efficiency, adaptability and analogue (non-Boolean) computation in brains. Neural networks have exhibited enhanced computational complexity when operated at the edge of chaos, and networks of chaotic elements have been proposed for solving combinatorial or global optimization problems. Thus, a source of controllable chaotic behaviour that can be incorporated into a neural-inspired circuit may be an essential component of future computational systems. Such chaotic elements have been simulated using elaborate transistor circuits that simulate known equations of chaos, but an experimental realization of chaotic dynamics from a single scalable electronic device has been lacking. Here we describe niobium dioxide (NbO2) Mott memristors each less than 100 nanometres across that exhibit both a nonlinear-transport-driven current-controlled negative differential resistance and a Mott-transition-driven temperature-controlled negative differential resistance. Mott materials have a temperature-dependent metal-insulator transition that acts as an electronic switch, which introduces a history-dependent resistance into the device. We incorporate these memristors into a relaxation oscillator and observe a tunable range of periodic and chaotic self-oscillations. We show that the nonlinear current transport coupled with thermal fluctuations at the nanoscale generates chaotic oscillations. Such memristors could be useful in certain types of neural-inspired computation by introducing a pseudo-random signal that prevents global synchronization and could also assist in finding a global minimum during a constrained search. We specifically demonstrate that incorporating such memristors into the hardware of a Hopfield computing network can greatly improve the efficiency and accuracy of converging to a solution for computationally difficult problems.
2007-02-28
Iterative Ultrasonic Signal and Image Deconvolution for Estimation of the Complex Medium Response, International Journal of Imaging Systems and...1767-1782, 2006. 31. Z. Mu, R. Plemmons, and P. Santago. Iterative Ultrasonic Signal and Image Deconvolution for Estimation of the Complex...rigorous mathematical and computational research on inverse problems in optical imaging of direct interest to the Army and also the intelligence agencies
DOE Office of Scientific and Technical Information (OSTI.GOV)
Huang, He; Luo, Li -Shi; Li, Rui
To compute the non-oscillating mutual interaction for a systems with N points, the fast multipole method (FMM) has an efficiency that scales linearly with the number of points. Specifically, for Coulomb interaction, FMM can be constructed using either the spherical harmonic functions or the totally symmetric Cartesian tensors. In this paper, we will present that the effciency of the Cartesian tensor-based FMM for the Coulomb interaction can be significantly improved by implementing the traces of the Cartesian tensors in calculation to reduce the independent elements of the n-th rank totally symmetric Cartesian tensor from (n + 1)(n + 2)=2 tomore » 2n + 1. The computation complexity for the operations in FMM are analyzed and expressed as polynomials of the highest rank of the Cartesian tensors. For most operations, the complexity is reduced by one order. Numerical examples regarding the convergence and the effciency of the new algorithm are demonstrated. As a result, a reduction of computation time up to 50% has been observed for a moderate number of points and rank of tensors.« less
Huang, He; Luo, Li -Shi; Li, Rui; ...
2018-05-17
To compute the non-oscillating mutual interaction for a systems with N points, the fast multipole method (FMM) has an efficiency that scales linearly with the number of points. Specifically, for Coulomb interaction, FMM can be constructed using either the spherical harmonic functions or the totally symmetric Cartesian tensors. In this paper, we will present that the effciency of the Cartesian tensor-based FMM for the Coulomb interaction can be significantly improved by implementing the traces of the Cartesian tensors in calculation to reduce the independent elements of the n-th rank totally symmetric Cartesian tensor from (n + 1)(n + 2)=2 tomore » 2n + 1. The computation complexity for the operations in FMM are analyzed and expressed as polynomials of the highest rank of the Cartesian tensors. For most operations, the complexity is reduced by one order. Numerical examples regarding the convergence and the effciency of the new algorithm are demonstrated. As a result, a reduction of computation time up to 50% has been observed for a moderate number of points and rank of tensors.« less
Operating a petabyte class archive at ESO
NASA Astrophysics Data System (ADS)
Suchar, Dieter; Lockhart, John S.; Burrows, Andrew
2008-07-01
The challenges of setting up and operating a Petabyte Class Archive will be described in terms of computer systems within a complex Data Centre environment. The computer systems, including the ESO Primary and Secondary Archive and the associated computational environments such as relational databases will be explained. This encompasses the entire system project cycle, including the technical specifications, procurement process, equipment installation and all further operational phases. The ESO Data Centre construction and the complexity of managing the environment will be presented. Many factors had to be considered during the construction phase, such as power consumption, targeted cooling and the accumulated load on the building structure to enable the smooth running of a Petabyte class Archive.
Breathing Mode in Complex Plasmas
NASA Astrophysics Data System (ADS)
Fujioka, K.; Henning, C.; Ludwig, P.; Bonitz, M.; Melzer, A.; Vitkalov, S.
2007-11-01
The breathing mode is a fundamental normal mode present in Coulomb systems, and may have utility in identifying particle charge and the Debye length of certain systems. The question remains whether this mode can be extended to strongly coupled Yukawa balls [1]. These systems are characterized by particles confined within a parabolic potential well and interacting through a shielded Coulomb potential [2,3]. The breathing modes for a variety of systems in 1, 2, and 3 dimensions are computed by solving the eigenvalue problem given by the dynamical (Hesse) matrix. These results are compared to theoretical investigations that assume a strict definition for a breathing mode within the system, and an analysis is made of the most fitting model to utilize in the study of particular systems of complex plasmas [1,4]. References [1] T.E. Sheridan, Phys. of Plasmas. 13, 022106 (2006)[2] C. Henning et al., Phys. Rev. E 74, 056403 (2006)[3] M. Bonitz et al., Phys. Rev. Lett. 96, 075001 (2006)[4] C. Henning et al., submitted for publication
Hypercluster Parallel Processor
NASA Technical Reports Server (NTRS)
Blech, Richard A.; Cole, Gary L.; Milner, Edward J.; Quealy, Angela
1992-01-01
Hypercluster computer system includes multiple digital processors, operation of which coordinated through specialized software. Configurable according to various parallel-computing architectures of shared-memory or distributed-memory class, including scalar computer, vector computer, reduced-instruction-set computer, and complex-instruction-set computer. Designed as flexible, relatively inexpensive system that provides single programming and operating environment within which one can investigate effects of various parallel-computing architectures and combinations on performance in solution of complicated problems like those of three-dimensional flows in turbomachines. Hypercluster software and architectural concepts are in public domain.
GMXPBSA 2.0: A GROMACS tool to perform MM/PBSA and computational alanine scanning
NASA Astrophysics Data System (ADS)
Paissoni, C.; Spiliotopoulos, D.; Musco, G.; Spitaleri, A.
2014-11-01
GMXPBSA 2.0 is a user-friendly suite of Bash/Perl scripts for streamlining MM/PBSA calculations on structural ensembles derived from GROMACS trajectories, to automatically calculate binding free energies for protein-protein or ligand-protein complexes. GMXPBSA 2.0 is flexible and can easily be customized to specific needs. Additionally, it performs computational alanine scanning (CAS) to study the effects of ligand and/or receptor alanine mutations on the free energy of binding. Calculations require only for protein-protein or protein-ligand MD simulations. GMXPBSA 2.0 performs different comparative analysis, including a posteriori generation of alanine mutants of the wild-type complex, calculation of the binding free energy values of the mutant complexes and comparison of the results with the wild-type system. Moreover, it compares the binding free energy of different complexes trajectories, allowing the study the effects of non-alanine mutations, post-translational modifications or unnatural amino acids on the binding free energy of the system under investigation. Finally, it can calculate and rank relative affinity to the same receptor utilizing MD simulations of proteins in complex with different ligands. In order to dissect the different MM/PBSA energy contributions, including molecular mechanic (MM), electrostatic contribution to solvation (PB) and nonpolar contribution to solvation (SA), the tool combines two freely available programs: the MD simulations software GROMACS and the Poisson-Boltzmann equation solver APBS. All the calculations can be performed in single or distributed automatic fashion on a cluster facility in order to increase the calculation by dividing frames across the available processors. The program is freely available under the GPL license.
High throughput computing: a solution for scientific analysis
O'Donnell, M.
2011-01-01
handle job failures due to hardware, software, or network interruptions (obviating the need to manually resubmit the job after each stoppage); be affordable; and most importantly, allow us to complete very large, complex analyses that otherwise would not even be possible. In short, we envisioned a job-management system that would take advantage of unused FORT CPUs within a local area network (LAN) to effectively distribute and run highly complex analytical processes. What we found was a solution that uses High Throughput Computing (HTC) and High Performance Computing (HPC) systems to do exactly that (Figure 1).
Reliability/safety analysis of a fly-by-wire system
NASA Technical Reports Server (NTRS)
Brock, L. D.; Goddman, H. A.
1980-01-01
An analysis technique has been developed to estimate the reliability of a very complex, safety-critical system by constructing a diagram of the reliability equations for the total system. This diagram has many of the characteristics of a fault-tree or success-path diagram, but is much easier to construct for complex redundant systems. The diagram provides insight into system failure characteristics and identifies the most likely failure modes. A computer program aids in the construction of the diagram and the computation of reliability. Analysis of the NASA F-8 Digital Fly-by-Wire Flight Control System is used to illustrate the technique.
Cloud Computing for Complex Performance Codes.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Appel, Gordon John; Hadgu, Teklu; Klein, Brandon Thorin
This report describes the use of cloud computing services for running complex public domain performance assessment problems. The work consisted of two phases: Phase 1 was to demonstrate complex codes, on several differently configured servers, could run and compute trivial small scale problems in a commercial cloud infrastructure. Phase 2 focused on proving non-trivial large scale problems could be computed in the commercial cloud environment. The cloud computing effort was successfully applied using codes of interest to the geohydrology and nuclear waste disposal modeling community.
Theoretical investigation of rotationally inelastic collisions of CH(X2Π) with hydrogen atoms
NASA Astrophysics Data System (ADS)
Dagdigian, Paul J.
2017-06-01
We report calculations of state-to-state cross sections for collision-induced rotational transitions of CH(X2Π) with atomic hydrogen. These calculations employed the four adiabatic potential energy surfaces correlating CH(X2Π) + H(2S), computed in this work through the multi-reference configuration interaction method [MRCISD + Q(Davidson)]. Because of the presence of deep wells on three of the potential energy surfaces, the scattering calculations were carried out using the quantum statistical method of Manolopoulos and co-workers [Chem. Phys. Lett. 343, 356 (2001)]. The computed cross sections included contributions from only direct scattering since the CH2 collision complex is expected to decay predominantly to C + H2. Rotationally energy transfer rate constants were computed for this system since these are required for astrophysical modeling.
Complexity in Dynamical Systems
NASA Astrophysics Data System (ADS)
Moore, Cristopher David
The study of chaos has shown us that deterministic systems can have a kind of unpredictability, based on a limited knowledge of their initial conditions; after a finite time, the motion appears essentially random. This observation has inspired a general interest in the subject of unpredictability, and more generally, complexity; how can we characterize how "complex" a dynamical system is?. In this thesis, we attempt to answer this question with a paradigm of complexity that comes from computer science, we extract sets of symbol sequences, or languages, from a dynamical system using standard methods of symbolic dynamics; we then ask what kinds of grammars or automata are needed a generate these languages. This places them in the Chomsky heirarchy, which in turn tells us something about how subtle and complex the dynamical system's behavior is. This gives us insight into the question of unpredictability, since these automata can also be thought of as computers attempting to predict the system. In the culmination of the thesis, we find a class of smooth, two-dimensional maps which are equivalent to the highest class in the Chomsky heirarchy, the turning machine; they are capable of universal computation. Therefore, these systems possess a kind of unpredictability qualitatively different from the usual "chaos": even if the initial conditions are known exactly, questions about the system's long-term dynamics are undecidable. No algorithm exists to answer them. Although this kind of unpredictability has been discussed in the context of distributed, many-degree-of -freedom systems (for instance, cellular automata) we believe this is the first example of such phenomena in a smooth, finite-degree-of-freedom system.
CFD research and systems in Kawasaki Heavy Industries and its future prospects
NASA Astrophysics Data System (ADS)
Hiraoka, Koichi
1990-09-01
KHI Computational Fluid Dynamics (CFD) system is composed of VP100 computer and 2-D and 3-D Euler and/or Navier-Stokes (NS) analysis softwares. For KHI, this system has become a very powerful aerodynamic tool together with the Kawasaki 1 m Transonic Wind Tunnel. The 2-D Euler/NS software, developed in-house, is fully automated, requires no special skill, and was successfully applied to the design of YXX high lift devices and SST supersonic inlet, etc. The 3-D Euler/NS software, developed under joint research with NAL, has an interactively operated Multi-Block type grid generator and can effectively generate grids around complex airplane shapes. Due to the main memory size limitation, 3-D analysis of relatively simple shape, such as SST wing-body, was computed in-house on VP100, otherwise, such as detailed 3-D analyses of ASUKA and HOPE, were computed on NAL VP400, which is 10 times more powerful than VP100, under KHI-NAL joint research. These analysis results have very good correlation with experimental results. However, the present CFD system is less productive than wind tunnel and has applicability limitations.
Cyber-physical approach to the network-centric robotics control task
NASA Astrophysics Data System (ADS)
Muliukha, Vladimir; Ilyashenko, Alexander; Zaborovsky, Vladimir; Lukashin, Alexey
2016-10-01
Complex engineering tasks concerning control for groups of mobile robots are developed poorly. In our work for their formalization we use cyber-physical approach, which extends the range of engineering and physical methods for a design of complex technical objects by researching the informational aspects of communication and interaction between objects and with an external environment [1]. The paper analyzes network-centric methods for control of cyber-physical objects. Robots or cyber-physical objects interact with each other by transmitting information via computer networks using preemptive queueing system and randomized push-out mechanism [2],[3]. The main field of application for the results of our work is space robotics. The selection of cyber-physical systems as a special class of designed objects is due to the necessity of integrating various components responsible for computing, communications and control processes. Network-centric solutions allow using universal means for the organization of information exchange to integrate different technologies for the control system.
NASA Astrophysics Data System (ADS)
Li, Yong; Li, Wang; He, Kai-Yu; Li, Pei; Huang, Yan; Nie, Zhou; Yao, Shou-Zhuo
2016-04-01
In natural biological systems, proteins exploit various functional peptide motifs to exert target response and activity switch, providing a functional and logic basis for complex cellular activities. Building biomimetic peptide-based bio-logic systems is highly intriguing but remains relatively unexplored due to limited logic recognition elements and complex signal outputs. In this proof-of-principle work, we attempted to address these problems by utilizing multi-functional peptide probes and the peptide-mediated nanoparticle assembly system. Here, the rationally designed peptide probes function as the dual-target responsive element specifically responsive to metal ions and enzymes as well as the mediator regulating the assembly of gold nanoparticles (AuNPs). Taking advantage of Zn2+ ions and chymotrypsin as the model inputs of metal ions and enzymes, respectively, we constructed the peptide logic system computed by the multi-functional peptide probes and outputted by the readable colour change of AuNPs. In this way, the representative binary basic logic gates (AND, OR, INHIBIT, NAND, IMPLICATION) have been achieved by delicately coding the peptide sequence, demonstrating the versatility of our logic system. Additionally, we demonstrated that the three-input combinational logic gate (INHIBIT-OR) could also be successfully integrated and applied as a multi-tasking biosensor for colorimetric detection of dual targets. This nanoparticle-based peptide logic system presents a valid strategy to illustrate peptide information processing and provides a practical platform for executing peptide computing or peptide-related multiplexing sensing, implying that the controllable nanomaterial assembly is a promising and potent methodology for the advancement of biomimetic bio-logic computation.In natural biological systems, proteins exploit various functional peptide motifs to exert target response and activity switch, providing a functional and logic basis for complex cellular activities. Building biomimetic peptide-based bio-logic systems is highly intriguing but remains relatively unexplored due to limited logic recognition elements and complex signal outputs. In this proof-of-principle work, we attempted to address these problems by utilizing multi-functional peptide probes and the peptide-mediated nanoparticle assembly system. Here, the rationally designed peptide probes function as the dual-target responsive element specifically responsive to metal ions and enzymes as well as the mediator regulating the assembly of gold nanoparticles (AuNPs). Taking advantage of Zn2+ ions and chymotrypsin as the model inputs of metal ions and enzymes, respectively, we constructed the peptide logic system computed by the multi-functional peptide probes and outputted by the readable colour change of AuNPs. In this way, the representative binary basic logic gates (AND, OR, INHIBIT, NAND, IMPLICATION) have been achieved by delicately coding the peptide sequence, demonstrating the versatility of our logic system. Additionally, we demonstrated that the three-input combinational logic gate (INHIBIT-OR) could also be successfully integrated and applied as a multi-tasking biosensor for colorimetric detection of dual targets. This nanoparticle-based peptide logic system presents a valid strategy to illustrate peptide information processing and provides a practical platform for executing peptide computing or peptide-related multiplexing sensing, implying that the controllable nanomaterial assembly is a promising and potent methodology for the advancement of biomimetic bio-logic computation. Electronic supplementary information (ESI) available: Additional figures (Tables S1-S3 and Fig. S1-S6). See DOI: 10.1039/c6nr01072e
Computer models of complex multiloop branched pipeline systems
NASA Astrophysics Data System (ADS)
Kudinov, I. V.; Kolesnikov, S. V.; Eremin, A. V.; Branfileva, A. N.
2013-11-01
This paper describes the principal theoretical concepts of the method used for constructing computer models of complex multiloop branched pipeline networks, and this method is based on the theory of graphs and two Kirchhoff's laws applied to electrical circuits. The models make it possible to calculate velocities, flow rates, and pressures of a fluid medium in any section of pipeline networks, when the latter are considered as single hydraulic systems. On the basis of multivariant calculations the reasons for existing problems can be identified, the least costly methods of their elimination can be proposed, and recommendations for planning the modernization of pipeline systems and construction of their new sections can be made. The results obtained can be applied to complex pipeline systems intended for various purposes (water pipelines, petroleum pipelines, etc.). The operability of the model has been verified on an example of designing a unified computer model of the heat network for centralized heat supply of the city of Samara.
Formal Requirements-Based Programming for Complex Systems
NASA Technical Reports Server (NTRS)
Rash, James L.; Hinchey, Michael G.; Rouff, Christopher A.; Gracanin, Denis
2005-01-01
Computer science as a field has not yet produced a general method to mechanically transform complex computer system requirements into a provably equivalent implementation. Such a method would be one major step towards dealing with complexity in computing, yet it remains the elusive holy grail of system development. Currently available tools and methods that start with a formal model of a system and mechanically produce a provably equivalent implementation are valuable but not sufficient. The gap that such tools and methods leave unfilled is that the formal models cannot be proven to be equivalent to the system requirements as originated by the customer For the classes of complex systems whose behavior can be described as a finite (but significant) set of scenarios, we offer a method for mechanically transforming requirements (expressed in restricted natural language, or appropriate graphical notations) into a provably equivalent formal model that can be used as the basis for code generation and other transformations. While other techniques are available, this method is unique in offering full mathematical tractability while using notations and techniques that are well known and well trusted. We illustrate the application of the method to an example procedure from the Hubble Robotic Servicing Mission currently under study and preliminary formulation at NASA Goddard Space Flight Center.
Modulating of the pnicogen-bonding by a H⋯π interaction: An ab initio study.
Esrafili, Mehdi D; Sadr-Mousavi, Asma
2017-08-01
An ab initio study of the cooperativity in XH 2 P⋯NCH⋯Z and XH 2 P⋯CNH⋯Z complexes (X=F, Cl, Br, CN, NC; Z=C 2 H 2 ,C 6 H 6 ) connected by pnicogen-bonding and H⋯π interactions is carried out by means of MP2 computational method. A detailed analysis of the structures, interaction energies and bonding properties is performed on these systems. For each set of the complexes considered, a favorable cooperativity is observed, especially in X=F and CN complexes. However, for a given X or Z, the amount of cooperativity effects in XH 2 P⋯CNH⋯Z complexes are more important than XH 2 P⋯NCH⋯Z counterparts. Besides, the influence of a H⋯π interaction on a P⋯N (C) bond is more pronounced than that of a P⋯N (C) bond on a H⋯π bond. The quantum theory of atoms in molecules shows that ternary complexes have increased electron densities at their bond critical points relative to the corresponding binary systems. The results also indicate that the strength of the P⋯N(C) and H⋯π interactions increases in the presence of the solvent. Copyright © 2017 Elsevier Inc. All rights reserved.
UNH Data Cooperative: A Cyber Infrastructure for Earth System Studies
NASA Astrophysics Data System (ADS)
Braswell, B. H.; Fekete, B. M.; Prusevich, A.; Gliden, S.; Magill, A.; Vorosmarty, C. J.
2007-12-01
Earth system scientists and managers have a continuously growing demand for a wide array of earth observations derived from various data sources including (a) modern satellite retrievals, (b) "in-situ" records, (c) various simulation outputs, and (d) assimilated data products combining model results with observational records. The sheer quantity of data, and formatting inconsistencies make it difficult for users to take full advantage of this important information resource. Thus the system could benefit from a thorough retooling of our current data processing procedures and infrastructure. Emerging technologies, like OPeNDAP and OGC map services, open standard data formats (NetCDF, HDF) data cataloging systems (NASA-Echo, Global Change Master Directory, etc.) are providing the basis for a new approach in data management and processing, where web- services are increasingly designed to serve computer-to-computer communications without human interactions and complex analysis can be carried out over distributed computer resources interconnected via cyber infrastructure. The UNH Earth System Data Collaborative is designed to utilize the aforementioned emerging web technologies to offer new means of access to earth system data. While the UNH Data Collaborative serves a wide array of data ranging from weather station data (Climate Portal) to ocean buoy records and ship tracks (Portsmouth Harbor Initiative) to land cover characteristics, etc. the underlaying data architecture shares common components for data mining and data dissemination via web-services. Perhaps the most unique element of the UNH Data Cooperative's IT infrastructure is its prototype modeling environment for regional ecosystem surveillance over the Northeast corridor, which allows the integration of complex earth system model components with the Cooperative's data services. While the complexity of the IT infrastructure to perform complex computations is continuously increasing, scientists are often forced to spend considerable amount of time to solve basic data management and preprocessing tasks and deal with low level computational design problems like parallelization of model codes. Our modeling infrastructure is designed to take care the bulk of the common tasks found in complex earth system models like I/O handling, computational domain and time management, parallel execution of the modeling tasks, etc. The modeling infrastructure allows scientists to focus on the numerical implementation of the physical processes on a single computational objects(typically grid cells) while the framework takes care of the preprocessing of input data, establishing of the data exchange between computation objects and the execution of the science code. In our presentation, we will discuss the key concepts of our modeling infrastructure. We will demonstrate integration of our modeling framework with data services offered by the UNH Earth System Data Collaborative via web interfaces. We will layout the road map to turn our prototype modeling environment into a truly community framework for wide range of earth system scientists and environmental managers.
Documentation Driven Development for Complex Real-Time Systems
2004-12-01
This paper presents a novel approach for development of complex real - time systems , called the documentation-driven development (DDD) approach. This... time systems . DDD will also support automated software generation based on a computational model and some relevant techniques. DDD includes two main...stakeholders to be easily involved in development processes and, therefore, significantly improve the agility of software development for complex real
Revisiting the Quantum Brain Hypothesis: Toward Quantum (Neuro)biology?
Jedlicka, Peter
2017-01-01
The nervous system is a non-linear dynamical complex system with many feedback loops. A conventional wisdom is that in the brain the quantum fluctuations are self-averaging and thus functionally negligible. However, this intuition might be misleading in the case of non-linear complex systems. Because of an extreme sensitivity to initial conditions, in complex systems the microscopic fluctuations may be amplified and thereby affect the system's behavior. In this way quantum dynamics might influence neuronal computations. Accumulating evidence in non-neuronal systems indicates that biological evolution is able to exploit quantum stochasticity. The recent rise of quantum biology as an emerging field at the border between quantum physics and the life sciences suggests that quantum events could play a non-trivial role also in neuronal cells. Direct experimental evidence for this is still missing but future research should address the possibility that quantum events contribute to an extremely high complexity, variability and computational power of neuronal dynamics.
Complex adaptive systems: A new approach for understanding health practices.
Gomersall, Tim
2018-06-22
This article explores the potential of complex adaptive systems theory to inform behaviour change research. A complex adaptive system describes a collection of heterogeneous agents interacting within a particular context, adapting to each other's actions. In practical terms, this implies that behaviour change is 1) socially and culturally situated; 2) highly sensitive to small baseline differences in individuals, groups, and intervention components; and 3) determined by multiple components interacting "chaotically". Two approaches to studying complex adaptive systems are briefly reviewed. Agent-based modelling is a computer simulation technique that allows researchers to investigate "what if" questions in a virtual environment. Applied qualitative research techniques, on the other hand, offer a way to examine what happens when an intervention is pursued in real-time, and to identify the sorts of rules and assumptions governing social action. Although these represent very different approaches to complexity, there may be scope for mixing these methods - for example, by grounding models in insights derived from qualitative fieldwork. Finally, I will argue that the concept of complex adaptive systems offers one opportunity to gain a deepened understanding of health-related practices, and to examine the social psychological processes that produce health-promoting or damaging actions.
NASA Astrophysics Data System (ADS)
Khalili, Ashkan; Jha, Ratneshwar; Samaratunga, Dulip
2016-11-01
Wave propagation analysis in 2-D composite structures is performed efficiently and accurately through the formulation of a User-Defined Element (UEL) based on the wavelet spectral finite element (WSFE) method. The WSFE method is based on the first-order shear deformation theory which yields accurate results for wave motion at high frequencies. The 2-D WSFE model is highly efficient computationally and provides a direct relationship between system input and output in the frequency domain. The UEL is formulated and implemented in Abaqus (commercial finite element software) for wave propagation analysis in 2-D composite structures with complexities. Frequency domain formulation of WSFE leads to complex valued parameters, which are decoupled into real and imaginary parts and presented to Abaqus as real values. The final solution is obtained by forming a complex value using the real number solutions given by Abaqus. Five numerical examples are presented in this article, namely undamaged plate, impacted plate, plate with ply drop, folded plate and plate with stiffener. Wave motions predicted by the developed UEL correlate very well with Abaqus simulations. The results also show that the UEL largely retains computational efficiency of the WSFE method and extends its ability to model complex features.
DOT National Transportation Integrated Search
1974-08-01
DYNALIST, a computer program that extracts complex eigenvalues and eigenvectors for dynamic systems described in terms of matrix equations of motion, has been acquired and made operational at TSC. In this report, simple dynamic systems are used to de...
Craciun, Smaranda; Donald, Kelling J
2009-07-06
We examine the bonding possibilities of the bis(phenalenyl) MP(2) sandwich complexes of the divalent metals M = Be, Mg, Ca, Sr, Ba, Zn, Cd, and Hg, at the B3LYP level of theory. The outcome is an extraordinarily diverse class of low symmetry bis(phenalenyl)metal complexes in which bonding preferences and binding enthalpies differ dramatically. The lowest energy group 2 metal MP(2) complexes include an intriguing eta(1),eta(3) BeP(2) structure, and bent eta(6),eta(6) systems for M = Ca, Sr, and Ba. The group 12 bis(phenalenyl) complexes are thermodynamically unstable eta(1),eta(1) slip-sandwich structures. To better understand changes in the structural preferences going from the (eta(6),eta(6)) group 2 to the (eta(1),eta(1)) group 12 complexes, we explored the bonding in the bis(phenalenyl) complexes of transition metals with stable +2 oxidations states between Ca and Zn in period 4. The computed binding enthalpies are large and negative for nearly all of the minimum energy bis(phenalenyl) complexes of the group 2 and the transition metals; they are tiny for MgP(2), and are quite positive for the group 12 systems. The structural preferences and stability of the complexes is a subtle negotiation of several influences: the (un)availability of (n - 1)d and np, orbitals for bonding, the cost of the rehybridization at carbon sites in the phenalenyl rings in preparation for bonding to the metals, and the (P---P) interaction between the phenalenyl radicals.
Acetonitrile-water hydrogen-bonded interaction: Matrix-isolation infrared and ab initio computation
NASA Astrophysics Data System (ADS)
Gopi, R.; Ramanathan, N.; Sundararajan, K.
2015-08-01
The 1:1 hydrogen-bonded complex of acetonitrile (CH3CN) and water (H2O) was trapped in Ar and N2 matrices and studied using infrared technique. Ab initio computations showed two types of complexes formed between CH3CN and H2O, a linear complex A with a Ctbnd N⋯H interaction between nitrogen of CH3CN and hydrogen of H2O and a cyclic complex B, in which the interactions are between the hydrogen of CH3CN with oxygen of H2O and hydrogen of H2O with π cloud of sbnd Ctbnd N of CH3CN. Vibrational wavenumber calculations revealed that both the complexes A and B were minima on the potential energy surface. Interaction energies computed at B3LYP/6-311++G(d,p) showed that linear complex A is more stable than cyclic complex B. Computations identified a blue shift of ∼11.5 cm-1 and a red shift of ∼6.5 cm-1 in the CN stretching mode for the complexes A and B, respectively. Experimentally, we observed a blue shift of ∼15.0 and ∼8.3 cm-1 in N2 and Ar matrices, respectively, in the CN stretching mode of CH3CN, which supports the formation of complex A. The Onsager Self Consistent Reaction Field (SCRF) model was used to explain the influence of matrices on the complexes A and B. To understand the nature of the interactions, Atoms in Molecules (AIM) and Natural Bond Orbital (NBO) analyses were carried out for the complexes A and B.
The emergence of mind and brain: an evolutionary, computational, and philosophical approach.
Mainzer, Klaus
2008-01-01
Modern philosophy of mind cannot be understood without recent developments in computer science, artificial intelligence (AI), robotics, neuroscience, biology, linguistics, and psychology. Classical philosophy of formal languages as well as symbolic AI assume that all kinds of knowledge must explicitly be represented by formal or programming languages. This assumption is limited by recent insights into the biology of evolution and developmental psychology of the human organism. Most of our knowledge is implicit and unconscious. It is not formally represented, but embodied knowledge, which is learnt by doing and understood by bodily interacting with changing environments. That is true not only for low-level skills, but even for high-level domains of categorization, language, and abstract thinking. The embodied mind is considered an emergent capacity of the brain as a self-organizing complex system. Actually, self-organization has been a successful strategy of evolution to handle the increasing complexity of the world. Genetic programs are not sufficient and cannot prepare the organism for all kinds of complex situations in the future. Self-organization and emergence are fundamental concepts in the theory of complex dynamical systems. They are also applied in organic computing as a recent research field of computer science. Therefore, cognitive science, AI, and robotics try to model the embodied mind in an artificial evolution. The paper analyzes these approaches in the interdisciplinary framework of complex dynamical systems and discusses their philosophical impact.
On the writing of programming systems for spacecraft computers.
NASA Technical Reports Server (NTRS)
Mathur, F. P.; Rohr, J. A.
1972-01-01
Consideration of the systems designed to generate programs for the increasingly complex digital computers being used on board unmanned deep-space probes. Such programming systems must accommodate the special-purpose features incorporated in the hardware. The use of higher-level language facilities in the programming system can significantly simplify the task. Computers for Mariner and for the Outer Planets Grand Tour are briefly described, as well as their programming systems. Aspects of the higher level languages are considered.
Quantifying uncertainty and computational complexity for pore-scale simulations
NASA Astrophysics Data System (ADS)
Chen, C.; Yuan, Z.; Wang, P.; Yang, X.; Zhenyan, L.
2016-12-01
Pore-scale simulation is an essential tool to understand the complex physical process in many environmental problems, from multi-phase flow in the subsurface to fuel cells. However, in practice, factors such as sample heterogeneity, data sparsity and in general, our insufficient knowledge of the underlying process, render many simulation parameters and hence the prediction results uncertain. Meanwhile, most pore-scale simulations (in particular, direct numerical simulation) incur high computational cost due to finely-resolved spatio-temporal scales, which further limits our data/samples collection. To address those challenges, we propose a novel framework based on the general polynomial chaos (gPC) and build a surrogate model representing the essential features of the underlying system. To be specific, we apply the novel framework to analyze the uncertainties of the system behavior based on a series of pore-scale numerical experiments, such as flow and reactive transport in 2D heterogeneous porous media and 3D packed beds. Comparing with recent pore-scale uncertainty quantification studies using Monte Carlo techniques, our new framework requires fewer number of realizations and hence considerably reduce the overall computational cost, while maintaining the desired accuracy.
Crowd Sensing-Enabling Security Service Recommendation for Social Fog Computing Systems
Wu, Jun; Su, Zhou; Li, Jianhua
2017-01-01
Fog computing, shifting intelligence and resources from the remote cloud to edge networks, has the potential of providing low-latency for the communication from sensing data sources to users. For the objects from the Internet of Things (IoT) to the cloud, it is a new trend that the objects establish social-like relationships with each other, which efficiently brings the benefits of developed sociality to a complex environment. As fog service become more sophisticated, it will become more convenient for fog users to share their own services, resources, and data via social networks. Meanwhile, the efficient social organization can enable more flexible, secure, and collaborative networking. Aforementioned advantages make the social network a potential architecture for fog computing systems. In this paper, we design an architecture for social fog computing, in which the services of fog are provisioned based on “friend” relationships. To the best of our knowledge, this is the first attempt at an organized fog computing system-based social model. Meanwhile, social networking enhances the complexity and security risks of fog computing services, creating difficulties of security service recommendations in social fog computing. To address this, we propose a novel crowd sensing-enabling security service provisioning method to recommend security services accurately in social fog computing systems. Simulation results show the feasibilities and efficiency of the crowd sensing-enabling security service recommendation method for social fog computing systems. PMID:28758943
Crowd Sensing-Enabling Security Service Recommendation for Social Fog Computing Systems.
Wu, Jun; Su, Zhou; Wang, Shen; Li, Jianhua
2017-07-30
Fog computing, shifting intelligence and resources from the remote cloud to edge networks, has the potential of providing low-latency for the communication from sensing data sources to users. For the objects from the Internet of Things (IoT) to the cloud, it is a new trend that the objects establish social-like relationships with each other, which efficiently brings the benefits of developed sociality to a complex environment. As fog service become more sophisticated, it will become more convenient for fog users to share their own services, resources, and data via social networks. Meanwhile, the efficient social organization can enable more flexible, secure, and collaborative networking. Aforementioned advantages make the social network a potential architecture for fog computing systems. In this paper, we design an architecture for social fog computing, in which the services of fog are provisioned based on "friend" relationships. To the best of our knowledge, this is the first attempt at an organized fog computing system-based social model. Meanwhile, social networking enhances the complexity and security risks of fog computing services, creating difficulties of security service recommendations in social fog computing. To address this, we propose a novel crowd sensing-enabling security service provisioning method to recommend security services accurately in social fog computing systems. Simulation results show the feasibilities and efficiency of the crowd sensing-enabling security service recommendation method for social fog computing systems.
NASA Astrophysics Data System (ADS)
Manfredi, Sabato
2016-06-01
Large-scale dynamic systems are becoming highly pervasive in their occurrence with applications ranging from system biology, environment monitoring, sensor networks, and power systems. They are characterised by high dimensionality, complexity, and uncertainty in the node dynamic/interactions that require more and more computational demanding methods for their analysis and control design, as well as the network size and node system/interaction complexity increase. Therefore, it is a challenging problem to find scalable computational method for distributed control design of large-scale networks. In this paper, we investigate the robust distributed stabilisation problem of large-scale nonlinear multi-agent systems (briefly MASs) composed of non-identical (heterogeneous) linear dynamical systems coupled by uncertain nonlinear time-varying interconnections. By employing Lyapunov stability theory and linear matrix inequality (LMI) technique, new conditions are given for the distributed control design of large-scale MASs that can be easily solved by the toolbox of MATLAB. The stabilisability of each node dynamic is a sufficient assumption to design a global stabilising distributed control. The proposed approach improves some of the existing LMI-based results on MAS by both overcoming their computational limits and extending the applicative scenario to large-scale nonlinear heterogeneous MASs. Additionally, the proposed LMI conditions are further reduced in terms of computational requirement in the case of weakly heterogeneous MASs, which is a common scenario in real application where the network nodes and links are affected by parameter uncertainties. One of the main advantages of the proposed approach is to allow to move from a centralised towards a distributed computing architecture so that the expensive computation workload spent to solve LMIs may be shared among processors located at the networked nodes, thus increasing the scalability of the approach than the network size. Finally, a numerical example shows the applicability of the proposed method and its advantage in terms of computational complexity when compared with the existing approaches.
Automation of multi-agent control for complex dynamic systems in heterogeneous computational network
NASA Astrophysics Data System (ADS)
Oparin, Gennady; Feoktistov, Alexander; Bogdanova, Vera; Sidorov, Ivan
2017-01-01
The rapid progress of high-performance computing entails new challenges related to solving large scientific problems for various subject domains in a heterogeneous distributed computing environment (e.g., a network, Grid system, or Cloud infrastructure). The specialists in the field of parallel and distributed computing give the special attention to a scalability of applications for problem solving. An effective management of the scalable application in the heterogeneous distributed computing environment is still a non-trivial issue. Control systems that operate in networks, especially relate to this issue. We propose a new approach to the multi-agent management for the scalable applications in the heterogeneous computational network. The fundamentals of our approach are the integrated use of conceptual programming, simulation modeling, network monitoring, multi-agent management, and service-oriented programming. We developed a special framework for an automation of the problem solving. Advantages of the proposed approach are demonstrated on the parametric synthesis example of the static linear regulator for complex dynamic systems. Benefits of the scalable application for solving this problem include automation of the multi-agent control for the systems in a parallel mode with various degrees of its detailed elaboration.
Stored program concept for analog computers
NASA Technical Reports Server (NTRS)
Hannauer, G., III; Patmore, J. R.
1971-01-01
Optimization of three-stage matrices, modularization, and black boxes design techniques provides for automatically interconnecting computing component inputs and outputs in general purpose analog computer. Design also produces relatively inexpensive and less complex automatic patching system.
NASA Technical Reports Server (NTRS)
Byrne, F. (Inventor)
1981-01-01
A high speed common data buffer system is described for providing an interface and communications medium between a plurality of computers utilized in a distributed computer complex forming part of a checkout, command and control system for space vehicles and associated ground support equipment. The system includes the capability for temporarily storing data to be transferred between computers, for transferring a plurality of interrupts between computers, for monitoring and recording these transfers, and for correcting errors incurred in these transfers. Validity checks are made on each transfer and appropriate error notification is given to the computer associated with that transfer.
NASA Astrophysics Data System (ADS)
Xavier, M. P.; do Nascimento, T. M.; dos Santos, R. W.; Lobosco, M.
2014-03-01
The development of computational systems that mimics the physiological response of organs or even the entire body is a complex task. One of the issues that makes this task extremely complex is the huge computational resources needed to execute the simulations. For this reason, the use of parallel computing is mandatory. In this work, we focus on the simulation of temporal and spatial behaviour of some human innate immune system cells and molecules in a small three-dimensional section of a tissue. To perform this simulation, we use multiple Graphics Processing Units (GPUs) in a shared-memory environment. Despite of high initialization and communication costs imposed by the use of GPUs, the techniques used to implement the HIS simulator have shown to be very effective to achieve this purpose.
High temperature reaction between sea salt deposit and (U,Zr)O2 simulated corium debris
NASA Astrophysics Data System (ADS)
Takano, Masahide; Nishi, Tsuyoshi
2013-11-01
In order to clarify the possible impacts of seawater injection on the chemical and physical state of the corium debris formed in the severe accident at Fukushima Daiichi Nuclear Power Plants, the high temperature reaction between sea salt deposit and (U,Zr)O2 simulated corium debris (sim-debris) was examined in the temperature range from 1088 to 1668 K. A dense layer of calcium and sodium uranate formed on the surface of a sim-debris pellet at 1275 K under airflow, with the thickness of over 50 μm. When the oxygen partial pressure is low, calcium is likely to dissolve into the cubic sim-debris phase to form solid solution (Ca,U,Zr)O2+x. The diffusion depth was 5-6 μm from the surface, subjected to 1275 K for 12 h. The crystalline MgO remains affixed on the surface as the main residue of salt components. A part of it can also dissolve into the sim-debris.
Development of X-TOOLSS: Preliminary Design of Space Systems Using Evolutionary Computation
NASA Technical Reports Server (NTRS)
Schnell, Andrew R.; Hull, Patrick V.; Turner, Mike L.; Dozier, Gerry; Alverson, Lauren; Garrett, Aaron; Reneau, Jarred
2008-01-01
Evolutionary computational (EC) techniques such as genetic algorithms (GA) have been identified as promising methods to explore the design space of mechanical and electrical systems at the earliest stages of design. In this paper the authors summarize their research in the use of evolutionary computation to develop preliminary designs for various space systems. An evolutionary computational solver developed over the course of the research, X-TOOLSS (Exploration Toolset for the Optimization of Launch and Space Systems) is discussed. With the success of early, low-fidelity example problems, an outline of work involving more computationally complex models is discussed.
Visualizing Parallel Computer System Performance
NASA Technical Reports Server (NTRS)
Malony, Allen D.; Reed, Daniel A.
1988-01-01
Parallel computer systems are among the most complex of man's creations, making satisfactory performance characterization difficult. Despite this complexity, there are strong, indeed, almost irresistible, incentives to quantify parallel system performance using a single metric. The fallacy lies in succumbing to such temptations. A complete performance characterization requires not only an analysis of the system's constituent levels, it also requires both static and dynamic characterizations. Static or average behavior analysis may mask transients that dramatically alter system performance. Although the human visual system is remarkedly adept at interpreting and identifying anomalies in false color data, the importance of dynamic, visual scientific data presentation has only recently been recognized Large, complex parallel system pose equally vexing performance interpretation problems. Data from hardware and software performance monitors must be presented in ways that emphasize important events while eluding irrelevant details. Design approaches and tools for performance visualization are the subject of this paper.
Mono- and binuclear non-heme iron chemistry from a theoretical perspective.
Rokob, Tibor András; Chalupský, Jakub; Bím, Daniel; Andrikopoulos, Prokopis C; Srnec, Martin; Rulíšek, Lubomír
2016-09-01
In this minireview, we provide an account of the current state-of-the-art developments in the area of mono- and binuclear non-heme enzymes (NHFe and NHFe2) and the smaller NHFe(2) synthetic models, mostly from a theoretical and computational perspective. The sheer complexity, and at the same time the beauty, of the NHFe(2) world represents a challenge for experimental as well as theoretical methods. We emphasize that the concerted progress on both theoretical and experimental side is a conditio sine qua non for future understanding, exploration and utilization of the NHFe(2) systems. After briefly discussing the current challenges and advances in the computational methodology, we review the recent spectroscopic and computational studies of NHFe(2) enzymatic and inorganic systems and highlight the correlations between various experimental data (spectroscopic, kinetic, thermodynamic, electrochemical) and computations. Throughout, we attempt to keep in mind the most fascinating and attractive phenomenon in the NHFe(2) chemistry, which is the fact that despite the strong oxidative power of many reactive intermediates, the NHFe(2) enzymes perform catalysis with high selectivity. We conclude with our personal viewpoint and hope that further developments in quantum chemistry and especially in the field of multireference wave function methods are needed to have a solid theoretical basis for the NHFe(2) studies, mostly by providing benchmarking and calibration of the computationally efficient and easy-to-use DFT methods.
Computational study of C(sp3)-O bond formation at a PdIV centre.
Canty, Allan J; Ariafard, Alireza; Camasso, Nicole M; Higgs, Andrew T; Yates, Brian F; Sanford, Melanie S
2017-03-14
This report describes a computational study of C(sp 3 )-OR bond formation from Pd IV complexes of general structure Pd IV (CH 2 CMe 2 -o-C 6 H 4 -C,C')(F)(OR)(bpy-N,N') (bpy = 2,2'-bipyridine). Dissociation of - OR from the different octahedral Pd IV starting materials results in a common square-pyramidal Pd IV cation. An S N 2-type attack by - OR ( - OR = phenoxide, acetate, difluoroacetate, and nitrate) then leads to C(sp 3 )-OR bond formation. In contrast, when - OR = triflate, concerted C(sp 3 )-C(sp 2 ) bond-forming reductive elimination takes place, and the calculations indicate this outcome is the result of thermodynamic rather than kinetic control. The energy requirements for the dissociation and S N 2 steps with different - OR follow opposing trends. The S N 2 transition states exhibit "PdCO" angles in a tight range of 151.5 to 153.0°, resulting from steric interactions between the oxygen atom and the gem-dimethyl group of the ligand. Conformational effects for various OR ligands and isomerisation of the complexes were also examined as components of the solution dynamics in these systems. In all cases, the trends observed computationally agree with those observed experimentally.
High-throughput bioinformatics with the Cyrille2 pipeline system
Fiers, Mark WEJ; van der Burgt, Ate; Datema, Erwin; de Groot, Joost CW; van Ham, Roeland CHJ
2008-01-01
Background Modern omics research involves the application of high-throughput technologies that generate vast volumes of data. These data need to be pre-processed, analyzed and integrated with existing knowledge through the use of diverse sets of software tools, models and databases. The analyses are often interdependent and chained together to form complex workflows or pipelines. Given the volume of the data used and the multitude of computational resources available, specialized pipeline software is required to make high-throughput analysis of large-scale omics datasets feasible. Results We have developed a generic pipeline system called Cyrille2. The system is modular in design and consists of three functionally distinct parts: 1) a web based, graphical user interface (GUI) that enables a pipeline operator to manage the system; 2) the Scheduler, which forms the functional core of the system and which tracks what data enters the system and determines what jobs must be scheduled for execution, and; 3) the Executor, which searches for scheduled jobs and executes these on a compute cluster. Conclusion The Cyrille2 system is an extensible, modular system, implementing the stated requirements. Cyrille2 enables easy creation and execution of high throughput, flexible bioinformatics pipelines. PMID:18269742
Complex Instruction Set Quantum Computing
NASA Astrophysics Data System (ADS)
Sanders, G. D.; Kim, K. W.; Holton, W. C.
1998-03-01
In proposed quantum computers, electromagnetic pulses are used to implement logic gates on quantum bits (qubits). Gates are unitary transformations applied to coherent qubit wavefunctions and a universal computer can be created using a minimal set of gates. By applying many elementary gates in sequence, desired quantum computations can be performed. This reduced instruction set approach to quantum computing (RISC QC) is characterized by serial application of a few basic pulse shapes and a long coherence time. However, the unitary matrix of the overall computation is ultimately a unitary matrix of the same size as any of the elementary matrices. This suggests that we might replace a sequence of reduced instructions with a single complex instruction using an optimally taylored pulse. We refer to this approach as complex instruction set quantum computing (CISC QC). One trades the requirement for long coherence times for the ability to design and generate potentially more complex pulses. We consider a model system of coupled qubits interacting through nearest neighbor coupling and show that CISC QC can reduce the time required to perform quantum computations.
PROCESS FOR THE PRODUCTION OF AMMONIUM URANIUM FLUORIDE
Ellis, A.S.; Mooney, R.B.
1953-08-25
This patent relates to the preparation of ammonium uranium fluoride. The process comprises adding a water soluble fluoride to an aqueous solution of a uranous compound containing an ammonium salt, and isolating the resulting precipitate. This patent relates to the manufacture of uranium tetnafluoride from ammonium uranium fluoride, NH/sub 4/UF/sub 5/. Uranium tetrafluoride is prepared by heating the ammonium uranium fluoride to a temperature at which dissociation occurs with liberation of ammonium fluoride. Preferably the process is carried out under reduced pressure, or in a current of an inert gas.
Amabilino, Silvia; Deeth, Robert J
2017-03-06
Six-coordinate, rigorously octahedral d 4 Mn(III) spin crossover (SCO) complexes are limited by symmetry to an S = 1 (intermediate spin, IS) to S = 2 (high spin, HS) transition. In order to realize the potential S = 0 to S = 2 transition, a lower symmetry and/or change in coordination number is needed, which we explore here computationally. First, a number of complexes are analyzed to develop a reliable and relatively fast DFT protocol for reproducing known Mn(III) spin state energetics. The hybrid meta-GGA functional TPSSh with a modest split valence plus polarization basis set and an empirical dispersion correction is found to predict correctly the ground spin state of Mn(III) complexes, including true low-spin (LS) S = 0 systems, with a range of donor sets including the hexadentate [N 4 O 2 ] Schiff base ligands. The electronic structure design criteria necessary for realizing a ΔS = 2 SCO transition are described, and a number of model complexes are screened for potential SCO behavior. Five-coordinate trigonal-bipyramidal symmetry fails to yield any suitable systems. Seven-coordinate, approximately pentagonal bipyramidal symmetry is more favorable, and when a known pentadentate macrocyclic donor is combined with π-acceptor axial ligands, a novel Mn(III) complex, [Mn(PABODP)(PF 3 ) 2 ] 3+ (PABODP = 2,13-dimethyl-3,6,9,12,18-pentaazabicyclo[12.3.1]octadeca-1(18),2,12,14,16-pentaene), is predicted to have the right spin state energetics for an S = 0 to S = 2 transition. Successful synthesis of such a complex could provide the first example of a ΔS = 2 SCO transition for d 4 Mn(III). However, the combination of a rigid macrocycle and a high coordination number dilutes the stereochemical activity of the d electrons, leading to relatively small structural changes between HS and LS systems. It may therefore remain a challenge to realize strong cooperative effects in Mn(III) systems.
Language Networks as Complex Systems
ERIC Educational Resources Information Center
Lee, Max Kueiming; Ou, Sheue-Jen
2008-01-01
Starting in the late eighties, with a growing discontent with analytical methods in science and the growing power of computers, researchers began to study complex systems such as living organisms, evolution of genes, biological systems, brain neural networks, epidemics, ecology, economy, social networks, etc. In the early nineties, the research…
Learning Probabilistic Features for Robotic Navigation Using Laser Sensors
Aznar, Fidel; Pujol, Francisco A.; Pujol, Mar; Rizo, Ramón; Pujol, María-José
2014-01-01
SLAM is a popular task used by robots and autonomous vehicles to build a map of an unknown environment and, at the same time, to determine their location within the map. This paper describes a SLAM-based, probabilistic robotic system able to learn the essential features of different parts of its environment. Some previous SLAM implementations had computational complexities ranging from O(Nlog(N)) to O(N 2), where N is the number of map features. Unlike these methods, our approach reduces the computational complexity to O(N) by using a model to fuse the information from the sensors after applying the Bayesian paradigm. Once the training process is completed, the robot identifies and locates those areas that potentially match the sections that have been previously learned. After the training, the robot navigates and extracts a three-dimensional map of the environment using a single laser sensor. Thus, it perceives different sections of its world. In addition, in order to make our system able to be used in a low-cost robot, low-complexity algorithms that can be easily implemented on embedded processors or microcontrollers are used. PMID:25415377
Learning probabilistic features for robotic navigation using laser sensors.
Aznar, Fidel; Pujol, Francisco A; Pujol, Mar; Rizo, Ramón; Pujol, María-José
2014-01-01
SLAM is a popular task used by robots and autonomous vehicles to build a map of an unknown environment and, at the same time, to determine their location within the map. This paper describes a SLAM-based, probabilistic robotic system able to learn the essential features of different parts of its environment. Some previous SLAM implementations had computational complexities ranging from O(Nlog(N)) to O(N(2)), where N is the number of map features. Unlike these methods, our approach reduces the computational complexity to O(N) by using a model to fuse the information from the sensors after applying the Bayesian paradigm. Once the training process is completed, the robot identifies and locates those areas that potentially match the sections that have been previously learned. After the training, the robot navigates and extracts a three-dimensional map of the environment using a single laser sensor. Thus, it perceives different sections of its world. In addition, in order to make our system able to be used in a low-cost robot, low-complexity algorithms that can be easily implemented on embedded processors or microcontrollers are used.
NASA Astrophysics Data System (ADS)
Jakubikova, Elena; He, Sheng-Gui; Xie, Yan; Matsuda, Yoshiyuki; Bernstein, Elliot
2007-03-01
Vanadium oxide is a catalytic system that plays an important role in the conversion of SO2 to SO3. Density functional theory at the BPW91/LANL2DZ level is employed to obtain structures of VOy (y=1,,5), V2Oy (y=2,,7), V3Oy (y=4,,9), V4Oy (y=7,,12) and their complexes with SO2. BPW91/LANL2DZ is insufficient to describe properly relative V-O and S-O bond strengths of vanadium and sulfur oxides. Calibration of theoretical results with experimental data is necessary to compute enthalpies of reactions between VxOy and SO2. Theoretical results indicate SO2 to SO conversion occurs for oxygen-deficient clusters and SO2 to SO3 conversion occurs for oxygen-rich clusters. Subsequent experimental studies confirm the presence of SO in the molecular beam as well as the presence of VxOy complexes with SO2. Some possible mechanisms for SO3 formation and catalyst regeneration for solids are also suggested.
NASA Astrophysics Data System (ADS)
Alameda, J. C.
2011-12-01
Development and optimization of computational science models, particularly on high performance computers, and with the advent of ubiquitous multicore processor systems, practically on every system, has been accomplished with basic software tools, typically, command-line based compilers, debuggers, performance tools that have not changed substantially from the days of serial and early vector computers. However, model complexity, including the complexity added by modern message passing libraries such as MPI, and the need for hybrid code models (such as openMP and MPI) to be able to take full advantage of high performance computers with an increasing core count per shared memory node, has made development and optimization of such codes an increasingly arduous task. Additional architectural developments, such as many-core processors, only complicate the situation further. In this paper, we describe how our NSF-funded project, "SI2-SSI: A Productive and Accessible Development Workbench for HPC Applications Using the Eclipse Parallel Tools Platform" (WHPC) seeks to improve the Eclipse Parallel Tools Platform, an environment designed to support scientific code development targeted at a diverse set of high performance computing systems. Our WHPC project to improve Eclipse PTP takes an application-centric view to improve PTP. We are using a set of scientific applications, each with a variety of challenges, and using PTP to drive further improvements to both the scientific application, as well as to understand shortcomings in Eclipse PTP from an application developer perspective, to drive our list of improvements we seek to make. We are also partnering with performance tool providers, to drive higher quality performance tool integration. We have partnered with the Cactus group at Louisiana State University to improve Eclipse's ability to work with computational frameworks and extremely complex build systems, as well as to develop educational materials to incorporate into computational science and engineering codes. Finally, we are partnering with the lead PTP developers at IBM, to ensure we are as effective as possible within the Eclipse community development. We are also conducting training and outreach to our user community, including conference BOF sessions, monthly user calls, and an annual user meeting, so that we can best inform the improvements we make to Eclipse PTP. With these activities we endeavor to encourage use of modern software engineering practices, as enabled through the Eclipse IDE, with computational science and engineering applications. These practices include proper use of source code repositories, tracking and rectifying issues, measuring and monitoring code performance changes against both optimizations as well as ever-changing software stacks and configurations on HPC systems, as well as ultimately encouraging development and maintenance of testing suites -- things that have become commonplace in many software endeavors, but have lagged in the development of science applications. We view that the challenge with the increased complexity of both HPC systems and science applications demands the use of better software engineering methods, preferably enabled by modern tools such as Eclipse PTP, to help the computational science community thrive as we evolve the HPC landscape.
NASA Astrophysics Data System (ADS)
Delle Fratte, C.; Kennedy, J. A.; Kluth, S.; Mazzaferro, L.
2015-12-01
In a grid computing infrastructure tasks such as continuous upgrades, services installations and software deployments are part of an admins daily work. In such an environment tools to help with the management, provisioning and monitoring of the deployed systems and services have become crucial. As experiments such as the LHC increase in scale, the computing infrastructure also becomes larger and more complex. Moreover, today's admins increasingly work within teams that share responsibilities and tasks. Such a scaled up situation requires tools that not only simplify the workload on administrators but also enable them to work seamlessly in teams. In this paper will be presented our experience from managing the Max Planck Institute Tier2 using Puppet and Gitolite in a cooperative way to help the system administrator in their daily work. In addition to describing the Puppet-Gitolite system, best practices and customizations will also be shown.
Smith, Sarah J; Riley, Mark J; Noble, Christopher J; Hanson, Graeme R; Stranger, Robert; Jayaratne, Vidura; Cavigliasso, Germán; Schenk, Gerhard; Gahan, Lawrence R
2009-11-02
The binuclear heterovalent manganese model complex [Mn(II)Mn(III)(L1)(OAc)(2)] ClO(4) x H(2)O (H(2)L1 = 2-(((3-((bis(pyridin-2-ylmethyl)amino)methyl)-2-hydroxy-5-methylbenzyl)(pyridin-2-ylmethyl)amino)-methyl)phenol) has been prepared and studied structurally, spectroscopically, and computationally. The magnetic and electronic properties of the complex have been related to its structure. The complex is weakly antiferromagnetically coupled (J approximately -5 cm(-1), H = -2J S(1) x S(2)) and the electron paramagnetic resonance (EPR) and magnetic circular dichroism (MCD) spectra identify the Jahn-Teller distortion of the Mn(III) center as predominantly a tetragonal compression, with a significant rhombic component. Electronic structure calculations using density functional theory have confirmed the conclusions derived from the experimental investigations. In contrast to isostructural M(II)Fe(III) complexes (M = Fe, Mn, Zn, Ni), the Mn(II)Mn(III) system is bifunctional possessing both catalase and hydrolase activities, and only one catalytically relevant pK(a) (= 8.2) is detected. Mechanistic implications are discussed.
NASA Astrophysics Data System (ADS)
Shaat, Musbah; Bader, Faouzi
2010-12-01
Cognitive Radio (CR) systems have been proposed to increase the spectrum utilization by opportunistically access the unused spectrum. Multicarrier communication systems are promising candidates for CR systems. Due to its high spectral efficiency, filter bank multicarrier (FBMC) can be considered as an alternative to conventional orthogonal frequency division multiplexing (OFDM) for transmission over the CR networks. This paper addresses the problem of resource allocation in multicarrier-based CR networks. The objective is to maximize the downlink capacity of the network under both total power and interference introduced to the primary users (PUs) constraints. The optimal solution has high computational complexity which makes it unsuitable for practical applications and hence a low complexity suboptimal solution is proposed. The proposed algorithm utilizes the spectrum holes in PUs bands as well as active PU bands. The performance of the proposed algorithm is investigated for OFDM and FBMC based CR systems. Simulation results illustrate that the proposed resource allocation algorithm with low computational complexity achieves near optimal performance and proves the efficiency of using FBMC in CR context.
Studying the evolution of a type III radio from the Sun up to 1 AU
NASA Astrophysics Data System (ADS)
Mann, Gottfried; Breitling, Frank; Vocks, Christian; Fallows, Richard; Melnik, Valentin; Konovalenko, Alexander
2017-04-01
On March 16, 2016, a type III burst was observed with the ground-based radio telescopes LOFAR and URAN-2 as well as with the radiospectrometer aboard the spacecraft WIND.It started at 80 MHz at 06:37 UT and reached 50 kHz after 23 minutes. A type III burst are considered as the radio signature of an electron beam travelling from the corona into the interplanetary space. The energetic electrons carrying the beam excites Langmuir waves, which convert into radio waves by wave-particle interaction. The relationship between the drift rate and the frequency as derived from the dynamic radio spectra reveals that the velocity of the electrons generating the radio waves of the type III burst is increasing with increasing distance from the center of the Sun.
A joint swarm intelligence algorithm for multi-user detection in MIMO-OFDM system
NASA Astrophysics Data System (ADS)
Hu, Fengye; Du, Dakun; Zhang, Peng; Wang, Zhijun
2014-11-01
In the multi-input multi-output orthogonal frequency division multiplexing (MIMO-OFDM) system, traditional multi-user detection (MUD) algorithms that usually used to suppress multiple access interference are difficult to balance system detection performance and the complexity of the algorithm. To solve this problem, this paper proposes a joint swarm intelligence algorithm called Ant Colony and Particle Swarm Optimisation (AC-PSO) by integrating particle swarm optimisation (PSO) and ant colony optimisation (ACO) algorithms. According to simulation results, it has been shown that, with low computational complexity, the MUD for the MIMO-OFDM system based on AC-PSO algorithm gains comparable MUD performance with maximum likelihood algorithm. Thus, the proposed AC-PSO algorithm provides a satisfactory trade-off between computational complexity and detection performance.
An Efficient Model-based Diagnosis Engine for Hybrid Systems Using Structural Model Decomposition
NASA Technical Reports Server (NTRS)
Bregon, Anibal; Narasimhan, Sriram; Roychoudhury, Indranil; Daigle, Matthew; Pulido, Belarmino
2013-01-01
Complex hybrid systems are present in a large range of engineering applications, like mechanical systems, electrical circuits, or embedded computation systems. The behavior of these systems is made up of continuous and discrete event dynamics that increase the difficulties for accurate and timely online fault diagnosis. The Hybrid Diagnosis Engine (HyDE) offers flexibility to the diagnosis application designer to choose the modeling paradigm and the reasoning algorithms. The HyDE architecture supports the use of multiple modeling paradigms at the component and system level. However, HyDE faces some problems regarding performance in terms of complexity and time. Our focus in this paper is on developing efficient model-based methodologies for online fault diagnosis in complex hybrid systems. To do this, we propose a diagnosis framework where structural model decomposition is integrated within the HyDE diagnosis framework to reduce the computational complexity associated with the fault diagnosis of hybrid systems. As a case study, we apply our approach to a diagnostic testbed, the Advanced Diagnostics and Prognostics Testbed (ADAPT), using real data.
NASA Technical Reports Server (NTRS)
Hinchey, Mike
2006-01-01
The explosion of capabilities and new products within ICT (Information and Communication Technology) has fostered widespread, overly optimistic opinions regarding the industry, based on common but unjustified assumptions of quality and correctness of software. These assumptions are encouraged by software producers and vendors, who have not succeeded in finding a way to overcome the lack of an automated, mathematically sound way to develop correct systems from requirements. NASA faces this dilemma as it envisages advanced mission concepts in future exploration missions, which may well be the most ambitious computer-based systems ever developed. Such missions entail levels of complexity that beg for new methods for system development. NASA-led research in such areas as sensor networks, formal methods, autonomic computing, and requirements-based programming (to name but a few) will offer some innovative approaches to achieving correctness in complex system development.
Lin, Na; Chen, Hanning; Jing, Shikai; Liu, Fang; Liang, Xiaodan
2017-03-01
In recent years, symbiosis as a rich source of potential engineering applications and computational model has attracted more and more attentions in the adaptive complex systems and evolution computing domains. Inspired by different symbiotic coevolution forms in nature, this paper proposed a series of multi-swarm particle swarm optimizers called PS 2 Os, which extend the single population particle swarm optimization (PSO) algorithm to interacting multi-swarms model by constructing hierarchical interaction topologies and enhanced dynamical update equations. According to different symbiotic interrelationships, four versions of PS 2 O are initiated to mimic mutualism, commensalism, predation, and competition mechanism, respectively. In the experiments, with five benchmark problems, the proposed algorithms are proved to have considerable potential for solving complex optimization problems. The coevolutionary dynamics of symbiotic species in each PS 2 O version are also studied respectively to demonstrate the heterogeneity of different symbiotic interrelationships that effect on the algorithm's performance. Then PS 2 O is used for solving the radio frequency identification (RFID) network planning (RNP) problem with a mixture of discrete and continuous variables. Simulation results show that the proposed algorithm outperforms the reference algorithms for planning RFID networks, in terms of optimization accuracy and computation robustness.
1983-09-01
F.P. PX /AMPZIJ/ REFH /AMPZIJ/ REFV /AI4PZIJ/ * RHOX /AI4PZIJ/ RHOY /At4PZIJ/ RHOZ /AI4PZIJ/ S A-ZJ SA /AMPZIJ/ SALP /AMPZIJ/ 6. CALLING ROUTINE: FLDDRV...US3NG ALGORITHM 72 COMPUTE P- YES .~:*:.~~ USING* *. 1. NAME: PLAINT (GTD) ] 2. PURPOSE: To determine if a ray traveling from a given source loca...determine if a source ray reflection from plate MP occurs. If a ray traveling from the source image location in the reflected ray direction passes through
NASA Technical Reports Server (NTRS)
Hill, R. W.
1994-01-01
The integration of CLIPS into HyperCard combines the intuitive, interactive user interface of the Macintosh with the powerful symbolic computation of an expert system interpreter. HyperCard is an excellent environment for quickly developing the front end of an application with buttons, dialogs, and pictures, while the CLIPS interpreter provides a powerful inference engine for complex problem solving and analysis. In order to understand the benefit of integrating HyperCard and CLIPS, consider the following: HyperCard is an information storage and retrieval system which exploits the use of the graphics and user interface capabilities of the Apple Macintosh computer. The user can easily define buttons, dialog boxes, information templates, pictures, and graphic displays through the use of the HyperCard tools and scripting language. What is generally lacking in this environment is a powerful reasoning engine for complex problem solving, and this is where CLIPS plays a role. CLIPS 5.0 (C Language Integrated Production System, v5.0) was developed at the Johnson Space Center Software Technology Branch to allow artificial intelligence research, development, and delivery on conventional computers. CLIPS 5.0 supports forward chaining rule systems, object-oriented language, and procedural programming for the construction of expert systems. It features incremental reset, seven conflict resolution stategies, truth maintenance, and user-defined external functions. Since CLIPS is implemented in the C language it is highly portable; in addition, it is embeddable as a callable routine from a program written in another language such as Ada or Fortran. By integrating HyperCard and CLIPS the advantages and uses of both packages are made available for a wide range of applications: rapid prototyping of knowledge-based expert systems, interactive simulations of physical systems and intelligent control of hypertext processes, to name a few. HyperCLIPS 2.0 is written in C-Language (54%) and Pascal (46%) for Apple Macintosh computers running Macintosh System 6.0.2 or greater. HyperCLIPS requires HyperCard 1.2 or higher and at least 2Mb of RAM are recommended to run. An executable is provided. To compile the source code, the Macintosh Programmer's Workshop (MPW) version 3.0, CLIPS 5.0 (MSC-21927), and the MPW C-Language compiler are also required. NOTE: Installing this program under Macintosh System 7 requires HyperCard v2.1. This program is distributed on a 3.5 inch Macintosh format diskette. A copy of the program documentation is included on the diskette, but may be purchased separately. HyperCLIPS was developed in 1990 and version 2.0 was released in 1991. HyperCLIPS is a copyrighted work with all copyright vested in NASA. Apple, Macintosh, MPW, and HyperCard are registered trademarks of Apple Computer, Inc.
NASA Technical Reports Server (NTRS)
Vinci, Samuel, J.
2012-01-01
This report is the third part of a three-part final report of research performed under an NRA cooperative Agreement contract. The first part was published as NASA/CR-2012-217415. The second part was published as NASA/CR-2012-217416. The study of the very high lift low-pressure turbine airfoil L1A in the presence of unsteady wakes was performed computationally and compared against experimental results. The experiments were conducted in a low speed wind tunnel under high (4.9%) and then low (0.6%) freestream turbulence intensity for Reynolds number equal to 25,000 and 50,000. The experimental and computational data have shown that in cases without wakes, the boundary layer separated without reattachment. The CFD was done with LES and URANS utilizing the finite-volume code ANSYS Fluent (ANSYS, Inc.) under the same freestream turbulence and Reynolds number conditions as the experiment but only at a rod to blade spacing of 1. With wakes, separation was largely suppressed, particularly if the wake passing frequency was sufficiently high. This was validated in the 3D CFD efforts by comparing the experimental results for the pressure coefficients and velocity profiles, which were reasonable for all cases examined. The 2D CFD efforts failed to capture the three dimensionality effects of the wake and thus were less consistent with the experimental data. The effect of the freestream turbulence intensity levels also showed a little more consistency with the experimental data at higher intensities when compared with the low intensity cases. Additional cases with higher wake passing frequencies which were not run experimentally were simulated. The results showed that an initial 25% increase from the experimental wake passing greatly reduced the size of the separation bubble, nearly completely suppressing it.
Lee, J D; Caven, B; Haake, S; Brown, T L
2001-01-01
As computer applications for cars emerge, a speech-based interface offers an appealing alternative to the visually demanding direct manipulation interface. However, speech-based systems may pose cognitive demands that could undermine driving safety. This study used a car-following task to evaluate how a speech-based e-mail system affects drivers' response to the periodic braking of a lead vehicle. The study included 24 drivers between the ages of 18 and 24 years. A baseline condition with no e-mail system was compared with a simple and a complex e-mail system in both simple and complex driving environments. The results show a 30% (310 ms) increase in reaction time when the speech-based system is used. Subjective workload ratings and probe questions also indicate that speech-based interaction introduces a significant cognitive load, which was highest for the complex e-mail system. These data show that a speech-based interface is not a panacea that eliminates the potential distraction of in-vehicle computers. Actual or potential applications of this research include design of in-vehicle information systems and evaluation of their contributions to driver distraction.
Computational Aspects of Heat Transfer in Structures
NASA Technical Reports Server (NTRS)
Adelman, H. M. (Compiler)
1982-01-01
Techniques for the computation of heat transfer and associated phenomena in complex structures are examined with an emphasis on reentry flight vehicle structures. Analysis methods, computer programs, thermal analysis of large space structures and high speed vehicles, and the impact of computer systems are addressed.
Computing health quality measures using Informatics for Integrating Biology and the Bedside.
Klann, Jeffrey G; Murphy, Shawn N
2013-04-19
The Health Quality Measures Format (HQMF) is a Health Level 7 (HL7) standard for expressing computable Clinical Quality Measures (CQMs). Creating tools to process HQMF queries in clinical databases will become increasingly important as the United States moves forward with its Health Information Technology Strategic Plan to Stages 2 and 3 of the Meaningful Use incentive program (MU2 and MU3). Informatics for Integrating Biology and the Bedside (i2b2) is one of the analytical databases used as part of the Office of the National Coordinator (ONC)'s Query Health platform to move toward this goal. Our goal is to integrate i2b2 with the Query Health HQMF architecture, to prepare for other HQMF use-cases (such as MU2 and MU3), and to articulate the functional overlap between i2b2 and HQMF. Therefore, we analyze the structure of HQMF, and then we apply this understanding to HQMF computation on the i2b2 clinical analytical database platform. Specifically, we develop a translator between two query languages, HQMF and i2b2, so that the i2b2 platform can compute HQMF queries. We use the HQMF structure of queries for aggregate reporting, which define clinical data elements and the temporal and logical relationships between them. We use the i2b2 XML format, which allows flexible querying of a complex clinical data repository in an easy-to-understand domain-specific language. The translator can represent nearly any i2b2-XML query as HQMF and execute in i2b2 nearly any HQMF query expressible in i2b2-XML. This translator is part of the freely available reference implementation of the QueryHealth initiative. We analyze limitations of the conversion and find it covers many, but not all, of the complex temporal and logical operators required by quality measures. HQMF is an expressive language for defining quality measures, and it will be important to understand and implement for CQM computation, in both meaningful use and population health. However, its current form might allow complexity that is intractable for current database systems (both in terms of implementation and computation). Our translator, which supports the subset of HQMF currently expressible in i2b2-XML, may represent the beginnings of a practical compromise. It is being pilot-tested in two Query Health demonstration projects, and it can be further expanded to balance computational tractability with the advanced features needed by measure developers.
Computing Health Quality Measures Using Informatics for Integrating Biology and the Bedside
Murphy, Shawn N
2013-01-01
Background The Health Quality Measures Format (HQMF) is a Health Level 7 (HL7) standard for expressing computable Clinical Quality Measures (CQMs). Creating tools to process HQMF queries in clinical databases will become increasingly important as the United States moves forward with its Health Information Technology Strategic Plan to Stages 2 and 3 of the Meaningful Use incentive program (MU2 and MU3). Informatics for Integrating Biology and the Bedside (i2b2) is one of the analytical databases used as part of the Office of the National Coordinator (ONC)’s Query Health platform to move toward this goal. Objective Our goal is to integrate i2b2 with the Query Health HQMF architecture, to prepare for other HQMF use-cases (such as MU2 and MU3), and to articulate the functional overlap between i2b2 and HQMF. Therefore, we analyze the structure of HQMF, and then we apply this understanding to HQMF computation on the i2b2 clinical analytical database platform. Specifically, we develop a translator between two query languages, HQMF and i2b2, so that the i2b2 platform can compute HQMF queries. Methods We use the HQMF structure of queries for aggregate reporting, which define clinical data elements and the temporal and logical relationships between them. We use the i2b2 XML format, which allows flexible querying of a complex clinical data repository in an easy-to-understand domain-specific language. Results The translator can represent nearly any i2b2-XML query as HQMF and execute in i2b2 nearly any HQMF query expressible in i2b2-XML. This translator is part of the freely available reference implementation of the QueryHealth initiative. We analyze limitations of the conversion and find it covers many, but not all, of the complex temporal and logical operators required by quality measures. Conclusions HQMF is an expressive language for defining quality measures, and it will be important to understand and implement for CQM computation, in both meaningful use and population health. However, its current form might allow complexity that is intractable for current database systems (both in terms of implementation and computation). Our translator, which supports the subset of HQMF currently expressible in i2b2-XML, may represent the beginnings of a practical compromise. It is being pilot-tested in two Query Health demonstration projects, and it can be further expanded to balance computational tractability with the advanced features needed by measure developers. PMID:23603227
Drawert, Brian; Trogdon, Michael; Toor, Salman; Petzold, Linda; Hellander, Andreas
2017-01-01
Computational experiments using spatial stochastic simulations have led to important new biological insights, but they require specialized tools and a complex software stack, as well as large and scalable compute and data analysis resources due to the large computational cost associated with Monte Carlo computational workflows. The complexity of setting up and managing a large-scale distributed computation environment to support productive and reproducible modeling can be prohibitive for practitioners in systems biology. This results in a barrier to the adoption of spatial stochastic simulation tools, effectively limiting the type of biological questions addressed by quantitative modeling. In this paper, we present PyURDME, a new, user-friendly spatial modeling and simulation package, and MOLNs, a cloud computing appliance for distributed simulation of stochastic reaction-diffusion models. MOLNs is based on IPython and provides an interactive programming platform for development of sharable and reproducible distributed parallel computational experiments. PMID:28190948
Computational modelling of oxygenation processes in enzymes and biomimetic model complexes.
de Visser, Sam P; Quesne, Matthew G; Martin, Bodo; Comba, Peter; Ryde, Ulf
2014-01-11
With computational resources becoming more efficient and more powerful and at the same time cheaper, computational methods have become more and more popular for studies on biochemical and biomimetic systems. Although large efforts from the scientific community have gone into exploring the possibilities of computational methods for studies on large biochemical systems, such studies are not without pitfalls and often cannot be routinely done but require expert execution. In this review we summarize and highlight advances in computational methodology and its application to enzymatic and biomimetic model complexes. In particular, we emphasize on topical and state-of-the-art methodologies that are able to either reproduce experimental findings, e.g., spectroscopic parameters and rate constants, accurately or make predictions of short-lived intermediates and fast reaction processes in nature. Moreover, we give examples of processes where certain computational methods dramatically fail.
Synthetic biology: insights into biological computation.
Manzoni, Romilde; Urrios, Arturo; Velazquez-Garcia, Silvia; de Nadal, Eulàlia; Posas, Francesc
2016-04-18
Organisms have evolved a broad array of complex signaling mechanisms that allow them to survive in a wide range of environmental conditions. They are able to sense external inputs and produce an output response by computing the information. Synthetic biology attempts to rationally engineer biological systems in order to perform desired functions. Our increasing understanding of biological systems guides this rational design, while the huge background in electronics for building circuits defines the methodology. In this context, biocomputation is the branch of synthetic biology aimed at implementing artificial computational devices using engineered biological motifs as building blocks. Biocomputational devices are defined as biological systems that are able to integrate inputs and return outputs following pre-determined rules. Over the last decade the number of available synthetic engineered devices has increased exponentially; simple and complex circuits have been built in bacteria, yeast and mammalian cells. These devices can manage and store information, take decisions based on past and present inputs, and even convert a transient signal into a sustained response. The field is experiencing a fast growth and every day it is easier to implement more complex biological functions. This is mainly due to advances in in vitro DNA synthesis, new genome editing tools, novel molecular cloning techniques, continuously growing part libraries as well as other technological advances. This allows that digital computation can now be engineered and implemented in biological systems. Simple logic gates can be implemented and connected to perform novel desired functions or to better understand and redesign biological processes. Synthetic biological digital circuits could lead to new therapeutic approaches, as well as new and efficient ways to produce complex molecules such as antibiotics, bioplastics or biofuels. Biological computation not only provides possible biomedical and biotechnological applications, but also affords a greater understanding of biological systems.
Second-order relativistic corrections for the S(L=0) states in one- and two-electron atomic systems
NASA Astrophysics Data System (ADS)
Frolov, A. M.; Mitelut, C. C.; Zhong, Z.
2005-01-01
An analytical approach is developed to compute the first- (similar to alpha(2)) and second-order (similar to alpha(4)) relativistic corrections in one- and two-electron atomic systems. The approach is based on the reduction of all operators to divergent (singular) and nondivergent (regular) parts. Then, we show that all the divergent parts from the differentmatrix elements cancel each other. The remaining expression contains only regular operators and its expectation value can be easily computed. Analysis of the S(L = 0) states in such systems is of specific interest since the corresponding operators for these states contain a large number of singularities. For one-electron systems the computed relativistic corrections coincide exactly with the appropriate result that follows from the Taylor expansion of the relativistic (i.e., Dirac) energy. We also discuss an alternative approach that allows one to cancel all singularities by using the so-called operator-compensation technique. This second approach is found to be very effective in applications of more complex systems, such as helium-like atoms and ions, H-2(+)-like ions, and some exotic three-body systems.
Revisiting the Quantum Brain Hypothesis: Toward Quantum (Neuro)biology?
Jedlicka, Peter
2017-01-01
The nervous system is a non-linear dynamical complex system with many feedback loops. A conventional wisdom is that in the brain the quantum fluctuations are self-averaging and thus functionally negligible. However, this intuition might be misleading in the case of non-linear complex systems. Because of an extreme sensitivity to initial conditions, in complex systems the microscopic fluctuations may be amplified and thereby affect the system’s behavior. In this way quantum dynamics might influence neuronal computations. Accumulating evidence in non-neuronal systems indicates that biological evolution is able to exploit quantum stochasticity. The recent rise of quantum biology as an emerging field at the border between quantum physics and the life sciences suggests that quantum events could play a non-trivial role also in neuronal cells. Direct experimental evidence for this is still missing but future research should address the possibility that quantum events contribute to an extremely high complexity, variability and computational power of neuronal dynamics. PMID:29163041
ERIC Educational Resources Information Center
Mitchell, Christine M.; Govindaraj, T.
1990-01-01
Discusses the use of intelligent tutoring systems as opposed to traditional on-the-job training for training operators of complex dynamic systems and describes the computer architecture for a system for operators of a NASA (National Aeronautics and Space Administration) satellite control system. An experimental evaluation with college students is…
O'Brien, Kieran T P; Kaltsoyannis, Nikolas
2017-01-17
A systematic computational study of organoactinide complexes of the form [LAnX] n+ has been carried out using density functional theory, the quantum theory of atoms in molecules (QTAIM) and Ziegler-Rauk energy decomposition analysis (EDA) methods. The systems studied feature L = trans-calix[2]benzene[2]pyrrolide, An = Th(iv), Th(iii), U(iii) and X = BH 4 , BO 2 C 2 H 4 , Me, N(SiH 3 ) 2 , OPh, CH 3 , NH 2 , OH, F, SiH 3 , PH 2 , SH, Cl, CH 2 Ph, NHPh, OPh, SiH 2 Ph, PHPh 2 , SPh, CPh 3 , NPh 2 , OPh, SiPh 3 PPh 2 , SPh. The PBE0 hybrid functional proved most suitable for geometry optimisations based on comparisons with available experimental data. An-X bond critical point electron densities, energy densities and An-X delocalisation indices, calculated with the PBE functional at the PBE0 geometries, are correlated with An-X bond energies, enthalpies and with the terms in the EDA. Good correlations are found between energies and QTAIM metrics, particularly for the orbital interaction term, provided the X ligand is part of an isoelectronic series and the number of open shell electrons is low (i.e. for the present Th(iv) and Th(iii) systems).
Mentat: An object-oriented macro data flow system
NASA Technical Reports Server (NTRS)
Grimshaw, Andrew S.; Liu, Jane W. S.
1988-01-01
Mentat, an object-oriented macro data flow system designed to facilitate parallelism in distributed systems, is presented. The macro data flow model is a model of computation similar to the data flow model with two principal differences: the computational complexity of the actors is much greater than in traditional data flow systems, and there are persistent actors that maintain state information between executions. Mentat is a system that combines the object-oriented programming paradigm and the macro data flow model of computation. Mentat programs use a dynamic structure called a future list to represent the future of computations.
Biswas, Amitava; Liu, Chen; Monga, Inder; ...
2016-01-01
For last few years, there has been a tremendous growth in data traffic due to high adoption rate of mobile devices and cloud computing. Internet of things (IoT) will stimulate even further growth. This is increasing scale and complexity of telecom/internet service provider (SP) and enterprise data centre (DC) compute and network infrastructures. As a result, managing these large network-compute converged infrastructures is becoming complex and cumbersome. To cope up, network and DC operators are trying to automate network and system operations, administrations and management (OAM) functions. OAM includes all non-functional mechanisms which keep the network running.
Paraskevopoulou, Sivylla E; Barsakcioglu, Deren Y; Saberi, Mohammed R; Eftekhar, Amir; Constandinou, Timothy G
2013-04-30
Next generation neural interfaces aspire to achieve real-time multi-channel systems by integrating spike sorting on chip to overcome limitations in communication channel capacity. The feasibility of this approach relies on developing highly efficient algorithms for feature extraction and clustering with the potential of low-power hardware implementation. We are proposing a feature extraction method, not requiring any calibration, based on first and second derivative features of the spike waveform. The accuracy and computational complexity of the proposed method are quantified and compared against commonly used feature extraction methods, through simulation across four datasets (with different single units) at multiple noise levels (ranging from 5 to 20% of the signal amplitude). The average classification error is shown to be below 7% with a computational complexity of 2N-3, where N is the number of sample points of each spike. Overall, this method presents a good trade-off between accuracy and computational complexity and is thus particularly well-suited for hardware-efficient implementation. Copyright © 2013 Elsevier B.V. All rights reserved.
Combining high performance simulation, data acquisition, and graphics display computers
NASA Technical Reports Server (NTRS)
Hickman, Robert J.
1989-01-01
Issues involved in the continuing development of an advanced simulation complex are discussed. This approach provides the capability to perform the majority of tests on advanced systems, non-destructively. The controlled test environments can be replicated to examine the response of the systems under test to alternative treatments of the system control design, or test the function and qualification of specific hardware. Field tests verify that the elements simulated in the laboratories are sufficient. The digital computer is hosted by a Digital Equipment Corp. MicroVAX computer with an Aptec Computer Systems Model 24 I/O computer performing the communication function. An Applied Dynamics International AD100 performs the high speed simulation computing and an Evans and Sutherland PS350 performs on-line graphics display. A Scientific Computer Systems SCS40 acts as a high performance FORTRAN program processor to support the complex, by generating numerous large files from programs coded in FORTRAN that are required for the real time processing. Four programming languages are involved in the process, FORTRAN, ADSIM, ADRIO, and STAPLE. FORTRAN is employed on the MicroVAX host to initialize and terminate the simulation runs on the system. The generation of the data files on the SCS40 also is performed with FORTRAN programs. ADSIM and ADIRO are used to program the processing elements of the AD100 and its IOCP processor. STAPLE is used to program the Aptec DIP and DIA processors.
The complexity of proving chaoticity and the Church-Turing thesis
NASA Astrophysics Data System (ADS)
Calude, Cristian S.; Calude, Elena; Svozil, Karl
2010-09-01
Proving the chaoticity of some dynamical systems is equivalent to solving the hardest problems in mathematics. Conversely, classical physical systems may "compute the hard or even the incomputable" by measuring observables which correspond to computationally hard or even incomputable problems.
User’s Guide for the VTRPE (Variable Terrain Radio Parabolic Equation) Computer Model
1991-10-01
propagation effects and antenna characteristics in radar system performance calculations. the radar transmission equation is oiten employed. Fol- lowing Kerr.2...electromagnetic wave equations for the complex electric and magnetic radiation fields. The model accounts for the effects of nonuniform atmospheric refractivity...mission equation, that is used in the performance prediction and analysis of radar and communication systems. Optimized fast Fourier transform (FFT
Advanced in Visualization of 3D Time-Dependent CFD Solutions
NASA Technical Reports Server (NTRS)
Lane, David A.; Lasinski, T. A. (Technical Monitor)
1995-01-01
Numerical simulations of complex 3D time-dependent (unsteady) flows are becoming increasingly feasible because of the progress in computing systems. Unfortunately, many existing flow visualization systems were developed for time-independent (steady) solutions and do not adequately depict solutions from unsteady flow simulations. Furthermore, most systems only handle one time step of the solutions individually and do not consider the time-dependent nature of the solutions. For example, instantaneous streamlines are computed by tracking the particles using one time step of the solution. However, for streaklines and timelines, particles need to be tracked through all time steps. Streaklines can reveal quite different information about the flow than those revealed by instantaneous streamlines. Comparisons of instantaneous streamlines with dynamic streaklines are shown. For a complex 3D flow simulation, it is common to generate a grid system with several millions of grid points and to have tens of thousands of time steps. The disk requirement for storing the flow data can easily be tens of gigabytes. Visualizing solutions of this magnitude is a challenging problem with today's computer hardware technology. Even interactive visualization of one time step of the flow data can be a problem for some existing flow visualization systems because of the size of the grid. Current approaches for visualizing complex 3D time-dependent CFD solutions are described. The flow visualization system developed at NASA Ames Research Center to compute time-dependent particle traces from unsteady CFD solutions is described. The system computes particle traces (streaklines) by integrating through the time steps. This system has been used by several NASA scientists to visualize their CFD time-dependent solutions. The flow visualization capabilities of this system are described, and visualization results are shown.
Computer modeling and simulation of human movement. Applications in sport and rehabilitation.
Neptune, R R
2000-05-01
Computer modeling and simulation of human movement plays an increasingly important role in sport and rehabilitation, with applications ranging from sport equipment design to understanding pathologic gait. The complex dynamic interactions within the musculoskeletal and neuromuscular systems make analyzing human movement with existing experimental techniques difficult but computer modeling and simulation allows for the identification of these complex interactions and causal relationships between input and output variables. This article provides an overview of computer modeling and simulation and presents an example application in the field of rehabilitation.
ADAM: Analysis of Discrete Models of Biological Systems Using Computer Algebra
2011-01-01
Background Many biological systems are modeled qualitatively with discrete models, such as probabilistic Boolean networks, logical models, Petri nets, and agent-based models, to gain a better understanding of them. The computational complexity to analyze the complete dynamics of these models grows exponentially in the number of variables, which impedes working with complex models. There exist software tools to analyze discrete models, but they either lack the algorithmic functionality to analyze complex models deterministically or they are inaccessible to many users as they require understanding the underlying algorithm and implementation, do not have a graphical user interface, or are hard to install. Efficient analysis methods that are accessible to modelers and easy to use are needed. Results We propose a method for efficiently identifying attractors and introduce the web-based tool Analysis of Dynamic Algebraic Models (ADAM), which provides this and other analysis methods for discrete models. ADAM converts several discrete model types automatically into polynomial dynamical systems and analyzes their dynamics using tools from computer algebra. Specifically, we propose a method to identify attractors of a discrete model that is equivalent to solving a system of polynomial equations, a long-studied problem in computer algebra. Based on extensive experimentation with both discrete models arising in systems biology and randomly generated networks, we found that the algebraic algorithms presented in this manuscript are fast for systems with the structure maintained by most biological systems, namely sparseness and robustness. For a large set of published complex discrete models, ADAM identified the attractors in less than one second. Conclusions Discrete modeling techniques are a useful tool for analyzing complex biological systems and there is a need in the biological community for accessible efficient analysis tools. ADAM provides analysis methods based on mathematical algorithms as a web-based tool for several different input formats, and it makes analysis of complex models accessible to a larger community, as it is platform independent as a web-service and does not require understanding of the underlying mathematics. PMID:21774817
A new order-theoretic characterisation of the polytime computable functions☆
Avanzini, Martin; Eguchi, Naohi; Moser, Georg
2015-01-01
We propose a new order-theoretic characterisation of the class of polytime computable functions. To this avail we define the small polynomial path order (sPOP⁎ for short). This termination order entails a new syntactic method to analyse the innermost runtime complexity of term rewrite systems fully automatically: for any rewrite system compatible with sPOP⁎ that employs recursion up to depth d, the (innermost) runtime complexity is polynomially bounded of degree d. This bound is tight. Thus we obtain a direct correspondence between a syntactic (and easily verifiable) condition of a program and the asymptotic worst-case complexity of the program. PMID:26412933
NASA Technical Reports Server (NTRS)
Steger, J. L.; Dougherty, F. C.; Benek, J. A.
1983-01-01
A mesh system composed of multiple overset body-conforming grids is described for adapting finite-difference procedures to complex aircraft configurations. In this so-called 'chimera mesh,' a major grid is generated about a main component of the configuration and overset minor grids are used to resolve all other features. Methods for connecting overset multiple grids and modifications of flow-simulation algorithms are discussed. Computational tests in two dimensions indicate that the use of multiple overset grids can simplify the task of grid generation without an adverse effect on flow-field algorithms and computer code complexity.
A Study of Complex Deep Learning Networks on High Performance, Neuromorphic, and Quantum Computers
DOE Office of Scientific and Technical Information (OSTI.GOV)
Potok, Thomas E; Schuman, Catherine D; Young, Steven R
Current Deep Learning models use highly optimized convolutional neural networks (CNN) trained on large graphical processing units (GPU)-based computers with a fairly simple layered network topology, i.e., highly connected layers, without intra-layer connections. Complex topologies have been proposed, but are intractable to train on current systems. Building the topologies of the deep learning network requires hand tuning, and implementing the network in hardware is expensive in both cost and power. In this paper, we evaluate deep learning models using three different computing architectures to address these problems: quantum computing to train complex topologies, high performance computing (HPC) to automatically determinemore » network topology, and neuromorphic computing for a low-power hardware implementation. Due to input size limitations of current quantum computers we use the MNIST dataset for our evaluation. The results show the possibility of using the three architectures in tandem to explore complex deep learning networks that are untrainable using a von Neumann architecture. We show that a quantum computer can find high quality values of intra-layer connections and weights, while yielding a tractable time result as the complexity of the network increases; a high performance computer can find optimal layer-based topologies; and a neuromorphic computer can represent the complex topology and weights derived from the other architectures in low power memristive hardware. This represents a new capability that is not feasible with current von Neumann architecture. It potentially enables the ability to solve very complicated problems unsolvable with current computing technologies.« less
Computer vision-based classification of hand grip variations in neurorehabilitation.
Zariffa, José; Steeves, John D
2011-01-01
The complexity of hand function is such that most existing upper limb rehabilitation robotic devices use only simplified hand interfaces. This is in contrast to the importance of the hand in regaining function after neurological injury. Computer vision technology has been used to identify hand posture in the field of Human Computer Interaction, but this approach has not been translated to the rehabilitation context. We describe a computer vision-based classifier that can be used to discriminate rehabilitation-relevant hand postures, and could be integrated into a virtual reality-based upper limb rehabilitation system. The proposed system was tested on a set of video recordings from able-bodied individuals performing cylindrical grasps, lateral key grips, and tip-to-tip pinches. The overall classification success rate was 91.2%, and was above 98% for 6 out of the 10 subjects. © 2011 IEEE
Intrinsic evolution of controllable oscillators in FPTA-2
NASA Technical Reports Server (NTRS)
Sekanina, Lukas; Zebulum, Ricardo S.
2005-01-01
Simple one- and two-bit controllable oscillators were intrinsically evolved using only four cells of Field Programmable Transistor Array (FPTA-2). These oscillators can produce different oscillations for different setting of control signals. Therefore, they could be used, in principle, to compose complex networks of oscillators that could exhibit rich dynamical behavior in order to perform a computation or to model a desired system.
Some Observations on the Current Status of Performing Finite Element Analyses
NASA Technical Reports Server (NTRS)
Raju, Ivatury S.; Knight, Norman F., Jr; Shivakumar, Kunigal N.
2015-01-01
Aerospace structures are complex high-performance structures. Advances in reliable and efficient computing and modeling tools are enabling analysts to consider complex configurations, build complex finite element models, and perform analysis rapidly. Many of the early career engineers of today are very proficient in the usage of modern computers, computing engines, complex software systems, and visualization tools. These young engineers are becoming increasingly efficient in building complex 3D models of complicated aerospace components. However, the current trends demonstrate blind acceptance of the results of the finite element analysis results. This paper is aimed at raising an awareness of this situation. Examples of the common encounters are presented. To overcome the current trends, some guidelines and suggestions for analysts, senior engineers, and educators are offered.
Graphics Flutter Analysis Methods, an interactive computing system at Lockheed-California Company
NASA Technical Reports Server (NTRS)
Radovcich, N. A.
1975-01-01
An interactive computer graphics system, Graphics Flutter Analysis Methods (GFAM), was developed to complement FAMAS, a matrix-oriented batch computing system, and other computer programs in performing complex numerical calculations using a fully integrated data management system. GFAM has many of the matrix operation capabilities found in FAMAS, but on a smaller scale, and is utilized when the analysis requires a high degree of interaction between the engineer and computer, and schedule constraints exclude the use of batch entry programs. Applications of GFAM to a variety of preliminary design, development design, and project modification programs suggest that interactive flutter analysis using matrix representations is a feasible and cost effective computing tool.
NASA Astrophysics Data System (ADS)
Bhardwaj, Jyotirmoy; Gupta, Karunesh K.; Gupta, Rajiv
2018-02-01
New concepts and techniques are replacing traditional methods of water quality parameter measurement systems. This paper introduces a cyber-physical system (CPS) approach for water quality assessment in a distribution network. Cyber-physical systems with embedded sensors, processors and actuators can be designed to sense and interact with the water environment. The proposed CPS is comprised of sensing framework integrated with five different water quality parameter sensor nodes and soft computing framework for computational modelling. Soft computing framework utilizes the applications of Python for user interface and fuzzy sciences for decision making. Introduction of multiple sensors in a water distribution network generates a huge number of data matrices, which are sometimes highly complex, difficult to understand and convoluted for effective decision making. Therefore, the proposed system framework also intends to simplify the complexity of obtained sensor data matrices and to support decision making for water engineers through a soft computing framework. The target of this proposed research is to provide a simple and efficient method to identify and detect presence of contamination in a water distribution network using applications of CPS.
Design and Development of a Web-Based Interactive Software Tool for Teaching Operating Systems
ERIC Educational Resources Information Center
Garmpis, Aristogiannis
2011-01-01
Operating Systems (OS) is an important and mandatory discipline in many Computer Science, Information Systems and Computer Engineering curricula. Some of its topics require a careful and detailed explanation from the instructor as they often involve theoretical concepts and somewhat complex mechanisms, demanding a certain degree of abstraction…
ERIC Educational Resources Information Center
Gardner, William L., III; Gundersen, David E.
1995-01-01
Discusses use of a mail survey to study the extent to which the hospitality industry employs various information technologies, including computer-mediated communication systems. Finds that hotel/motel size and chain affiliation are related to information system complexity, and that chain-affiliated hotels provide less computer and…
Dickie, Courtney M; Laughlin, Alexander L; Wofford, Joshua D; Bhuvanesh, Nattamai S; Nippe, Michael
2017-12-01
Single-molecule magnets (SMMs) are considered viable candidates for next-generation data storage and quantum computing. Systems featuring switchability of their magnetization dynamics are particularly interesting with respect to accessing more complex logic gates and device architectures. Here we show that transition metal based redox events can be exploited to enable reversible switchability of slow magnetic relaxation of magnetically anisotropic lanthanide ions. Specifically, we report anionic homoleptic bis-diamidoferrocene complexes of Dy 3+ (oblate) and Er 3+ (prolate) which can be reversibly oxidized by one electron to yield their respective charge neutral redox partners (Dy: [1] - , 1 ; Er: [2] - , 2 ). Importantly, compounds 1 and 2 are thermally stable which allowed for detailed studies of their magnetization dynamics. We show that the Dy 3+ [1] - / 1 system can function as an "on"/"off" or a "slow"/"fast" redox switchable SMM system in the absence or presence of applied dc fields, respectively. The Er 3+ based [2] - / 2 system features "on"/"off" switchability of SMM properties in the presence of applied fields. Results from electrochemical investigations, UV-vis-NIR spectroscopy, and 57 Fe Mössbauer spectroscopy indicate the presence of significant electronic communication between the mixed-valent Fe ions in 1 and 2 in both solution and solid state. This comparative evaluation of redox-switchable magnetization dynamics in low coordinate lanthanide complexes may be used as a potential blueprint toward the development of future switchable magnetic materials.
Reduced circuit implementation of encoder and syndrome generator
DOE Office of Scientific and Technical Information (OSTI.GOV)
Trager, Barry M; Winograd, Shmuel
An error correction method and system includes an Encoder and Syndrome-generator that operate in parallel to reduce the amount of circuitry used to compute check symbols and syndromes for error correcting codes. The system and method computes the contributions to the syndromes and check symbols 1 bit at a time instead of 1 symbol at a time. As a result, the even syndromes can be computed as powers of the odd syndromes. Further, the system assigns symbol addresses so that there are, for an example GF(2.sup.8) which has 72 symbols, three (3) blocks of addresses which differ by a cubemore » root of unity to allow the data symbols to be combined for reducing size and complexity of odd syndrome circuits. Further, the implementation circuit for generating check symbols is derived from syndrome circuit using the inverse of the part of the syndrome matrix for check locations.« less
An Integrated Crustal Dynamics Simulator
NASA Astrophysics Data System (ADS)
Xing, H. L.; Mora, P.
2007-12-01
Numerical modelling offers an outstanding opportunity to gain an understanding of the crustal dynamics and complex crustal system behaviour. This presentation provides our long-term and ongoing effort on finite element based computational model and software development to simulate the interacting fault system for earthquake forecasting. A R-minimum strategy based finite-element computational model and software tool, PANDAS, for modelling 3-dimensional nonlinear frictional contact behaviour between multiple deformable bodies with the arbitrarily-shaped contact element strategy has been developed by the authors, which builds up a virtual laboratory to simulate interacting fault systems including crustal boundary conditions and various nonlinearities (e.g. from frictional contact, materials, geometry and thermal coupling). It has been successfully applied to large scale computing of the complex nonlinear phenomena in the non-continuum media involving the nonlinear frictional instability, multiple material properties and complex geometries on supercomputers, such as the South Australia (SA) interacting fault system, South California fault model and Sumatra subduction model. It has been also extended and to simulate the hot fractured rock (HFR) geothermal reservoir system in collaboration of Geodynamics Ltd which is constructing the first geothermal reservoir system in Australia and to model the tsunami generation induced by earthquakes. Both are supported by Australian Research Council.
A CFD study of complex missile and store configurations in relative motion
NASA Technical Reports Server (NTRS)
Baysal, Oktay
1995-01-01
An investigation was conducted from May 16, 1990 to August 31, 1994 on the development of computational fluid dynamics (CFD) methodologies for complex missiles and the store separation problem. These flowfields involved multiple-component configurations, where at least one of the objects was engaged in relative motion. The two most important issues that had to be addressed were: (1) the unsteadiness of the flowfields (time-accurate and efficient CFD algorithms for the unsteady equations), and (2) the generation of grid systems which would permit multiple and moving bodies in the computational domain (dynamic domain decomposition). The study produced two competing and promising methodologies, and their proof-of-concept cases, which have been reported in the open literature: (1) Unsteady solutions on dynamic, overlapped grids, which may also be perceived as moving, locally-structured grids, and (2) Unsteady solutions on dynamic, unstructured grids.
Computationally efficient algorithm for high sampling-frequency operation of active noise control
NASA Astrophysics Data System (ADS)
Rout, Nirmal Kumar; Das, Debi Prasad; Panda, Ganapati
2015-05-01
In high sampling-frequency operation of active noise control (ANC) system the length of the secondary path estimate and the ANC filter are very long. This increases the computational complexity of the conventional filtered-x least mean square (FXLMS) algorithm. To reduce the computational complexity of long order ANC system using FXLMS algorithm, frequency domain block ANC algorithms have been proposed in past. These full block frequency domain ANC algorithms are associated with some disadvantages such as large block delay, quantization error due to computation of large size transforms and implementation difficulties in existing low-end DSP hardware. To overcome these shortcomings, the partitioned block ANC algorithm is newly proposed where the long length filters in ANC are divided into a number of equal partitions and suitably assembled to perform the FXLMS algorithm in the frequency domain. The complexity of this proposed frequency domain partitioned block FXLMS (FPBFXLMS) algorithm is quite reduced compared to the conventional FXLMS algorithm. It is further reduced by merging one fast Fourier transform (FFT)-inverse fast Fourier transform (IFFT) combination to derive the reduced structure FPBFXLMS (RFPBFXLMS) algorithm. Computational complexity analysis for different orders of filter and partition size are presented. Systematic computer simulations are carried out for both the proposed partitioned block ANC algorithms to show its accuracy compared to the time domain FXLMS algorithm.
Butler, Samuel D; Nauyoks, Stephen E; Marciniak, Michael A
2015-06-01
Of the many classes of bidirectional reflectance distribution function (BRDF) models, two popular classes of models are the microfacet model and the linear systems diffraction model. The microfacet model has the benefit of speed and simplicity, as it uses geometric optics approximations, while linear systems theory uses a diffraction approach to compute the BRDF, at the expense of greater computational complexity. In this Letter, nongrazing BRDF measurements of rough and polished surface-reflecting materials at multiple incident angles are scaled by the microfacet cross section conversion term, but in the linear systems direction cosine space, resulting in great alignment of BRDF data at various incident angles in this space. This results in a predictive BRDF model for surface-reflecting materials at nongrazing angles, while avoiding some of the computational complexities in the linear systems diffraction model.
Hasson, Uri; Skipper, Jeremy I; Wilde, Michael J; Nusbaum, Howard C; Small, Steven L
2008-01-15
The increasingly complex research questions addressed by neuroimaging research impose substantial demands on computational infrastructures. These infrastructures need to support management of massive amounts of data in a way that affords rapid and precise data analysis, to allow collaborative research, and to achieve these aims securely and with minimum management overhead. Here we present an approach that overcomes many current limitations in data analysis and data sharing. This approach is based on open source database management systems that support complex data queries as an integral part of data analysis, flexible data sharing, and parallel and distributed data processing using cluster computing and Grid computing resources. We assess the strengths of these approaches as compared to current frameworks based on storage of binary or text files. We then describe in detail the implementation of such a system and provide a concrete description of how it was used to enable a complex analysis of fMRI time series data.
Hasson, Uri; Skipper, Jeremy I.; Wilde, Michael J.; Nusbaum, Howard C.; Small, Steven L.
2007-01-01
The increasingly complex research questions addressed by neuroimaging research impose substantial demands on computational infrastructures. These infrastructures need to support management of massive amounts of data in a way that affords rapid and precise data analysis, to allow collaborative research, and to achieve these aims securely and with minimum management overhead. Here we present an approach that overcomes many current limitations in data analysis and data sharing. This approach is based on open source database management systems that support complex data queries as an integral part of data analysis, flexible data sharing, and parallel and distributed data processing using cluster computing and Grid computing resources. We assess the strengths of these approaches as compared to current frameworks based on storage of binary or text files. We then describe in detail the implementation of such a system and provide a concrete description of how it was used to enable a complex analysis of fMRI time series data. PMID:17964812
Nutrition and the science of disease prevention: a systems approach to support metabolic health
Bennett, Brian J.; Hall, Kevin D.; Hu, Frank B.; McCartney, Anne L.; Roberto, Christina
2017-01-01
Progress in nutritional science, genetics, computer science, and behavioral economics can be leveraged to address the challenge of noncommunicable disease. This report highlights the connection between nutrition and the complex science of preventing disease and discusses the promotion of optimal metabolic health, building on input from several complementary disciplines. The discussion focuses on (1) the basic science of optimal metabolic health, including data from gene–diet interactions, microbiome, and epidemiological research in nutrition, with the goal of defining better targets and interventions, and (2) how nutrition, from pharma to lifestyle, can build on systems science to address complex issues. PMID:26415028
NASA Astrophysics Data System (ADS)
Bérczi, Sz.; Hegyi, S.; Hudoba, Gy.; Hargitai, H.; Kokiny, A.; Drommer, B.; Gucsik, A.; Pintér, A.; Kovács, Zs.
Several teachers and students had the possibility to visit International Space Camp in the vicinity of the MSFC NASA in Huntsville Alabama USA where they learned the success of simulators in space science education To apply these results in universities and colleges in Hungary we began a unified complex modelling in planetary geology robotics electronics and complex environmental analysis by constructing an experimental space probe model system First a university experimental lander HUNVEYOR Hungarian UNiversity surVEYOR then a rover named HUSAR Hungarian University Surface Analyser Rover has been built For Hunveyor the idea and example was the historical Surveyor program of NASA in the 1960-ies for the Husar the idea and example was the Pathfinder s rover Sojouner rover The first step was the construction of the lander a year later the rover followed The main goals are 1 to build the lander structure and basic electronics from cheap everyday PC compatible elements 2 to construct basic experiments and their instruments 3 to use the system as a space activity simulator 4 this simulator contains lander with on board computer for works on a test planetary surface and a terrestrial control computer 5 to harmonize the assemblage of the electronic system and instruments in various levels of autonomy from the power and communication circuits 6 to use the complex system in education for in situ understanding complex planetary environmental problems 7 to build various planetary environments for application of the
76 FR 64330 - Advanced Scientific Computing Advisory Committee
Federal Register 2010, 2011, 2012, 2013, 2014
2011-10-18
... talks on HPC Reliability, Diffusion on Complex Networks, and Reversible Software Execution Systems Report from Applied Math Workshop on Mathematics for the Analysis, Simulation, and Optimization of Complex Systems Report from ASCR-BES Workshop on Data Challenges from Next Generation Facilities Public...
Nagula, Narsimha; Kunche, Sudeepa; Jaheer, Mohmed; Mudavath, Ravi; Sivan, Sreekanth; Ch, Sarala Devi
2018-01-01
Some novel transition metal [Cu (II), Ni (II) and Co (II)] complexes of nalidixic acid hydrazone have been prepared and characterized by employing spectro-analytical techniques viz: elemental analysis, 1 H-NMR, Mass, UV-Vis, IR, TGA-DTA, SEM-EDX, ESR and Spectrophotometry studies. The HyperChem 7.5 software was used for geometry optimization of title compound in its molecular and ionic forms. Quantum mechanical parameters, contour maps of highest occupied molecular orbitals (HOMO) and lowest unoccupied molecular orbitals (LUMO) and corresponding binding energy values were computed using semi empirical single point PM3 method. The stoichiometric equilibrium studies of metal complexes carried out spectrophotometrically using Job's continuous variation and mole ratio methods inferred formation of 1:2 (ML 2 ) metal complexes in respective systems. The title compound and its metal complexes screened for antibacterial and antifungal properties, exemplified improved activity in metal complexes. The studies of nuclease activity for the cleavage of CT- DNA and MTT assay for in vitro cytotoxic properties involving metal complexes exhibited high activity. In addition, the DNA binding properties of Cu (II), Ni (II) and Co (II) complexes investigated by electronic absorption and fluorescence measurements revealed their good binding ability and commended agreement of K b values obtained from both the techniques. Molecular docking studies were also performed to find the binding affinity of synthesized compounds with DNA (PDB ID: 1N37) and "Thymidine phosphorylase from E.coli" (PDB ID: 4EAF) protein targets.
Human systems dynamics: Toward a computational model
NASA Astrophysics Data System (ADS)
Eoyang, Glenda H.
2012-09-01
A robust and reliable computational model of complex human systems dynamics could support advancements in theory and practice for social systems at all levels, from intrapersonal experience to global politics and economics. Models of human interactions have evolved from traditional, Newtonian systems assumptions, which served a variety of practical and theoretical needs of the past. Another class of models has been inspired and informed by models and methods from nonlinear dynamics, chaos, and complexity science. None of the existing models, however, is able to represent the open, high dimension, and nonlinear self-organizing dynamics of social systems. An effective model will represent interactions at multiple levels to generate emergent patterns of social and political life of individuals and groups. Existing models and modeling methods are considered and assessed against characteristic pattern-forming processes in observed and experienced phenomena of human systems. A conceptual model, CDE Model, based on the conditions for self-organizing in human systems, is explored as an alternative to existing models and methods. While the new model overcomes the limitations of previous models, it also provides an explanatory base and foundation for prospective analysis to inform real-time meaning making and action taking in response to complex conditions in the real world. An invitation is extended to readers to engage in developing a computational model that incorporates the assumptions, meta-variables, and relationships of this open, high dimension, and nonlinear conceptual model of the complex dynamics of human systems.
NASA Astrophysics Data System (ADS)
Mallaekeh, Hassan; Shams, Alireza; Shaker, Mohammad; Bahramzadeh, Ehsan; Arefi, Donya
2014-12-01
In this paper the complexation reaction of the 5,5'-(2 E,2' E)-2,2'-(ethane-1,2-diylidene)bis(hydrazine-1-yl-2-ylidene)bis(4-amino-4H-1,2,4-triazole-3-thiol) ligand (HYT) with Cd2+ education was studied in some binary mixtures of methanol (MeOH), n-propanol (PrOH) and dimethyl-formamide (DMF) at different temperatures using the conductometry and spectrophotometry. The stability constants of the complex was determined using a GENPLOT computer program. The conductance data and absorbance-mole ratio plots show that in all solvent systems, the stoichiometry of the complex formed between (HYT) and Cd2+ cation is 1: 1. The obtained results show that the stability of (HYT)-Cd complex is sensitive to the mixed solvents composition. The values of thermodynamic parameters (Δ G ∘, Δ H ∘, and Δ S ∘) for formation of (HYT)-Cd complex were obtained from temperature dependence of the stability constant using the van't Hoff plots. The results show that in most cases, the complex are enthalpy destabilized but entropy stabilized and the complex formation is affected by pH, time, temperature and the nature of the solvent.
The application of artificial intelligence in the optimal design of mechanical systems
NASA Astrophysics Data System (ADS)
Poteralski, A.; Szczepanik, M.
2016-11-01
The paper is devoted to new computational techniques in mechanical optimization where one tries to study, model, analyze and optimize very complex phenomena, for which more precise scientific tools of the past were incapable of giving low cost and complete solution. Soft computing methods differ from conventional (hard) computing in that, unlike hard computing, they are tolerant of imprecision, uncertainty, partial truth and approximation. The paper deals with an application of the bio-inspired methods, like the evolutionary algorithms (EA), the artificial immune systems (AIS) and the particle swarm optimizers (PSO) to optimization problems. Structures considered in this work are analyzed by the finite element method (FEM), the boundary element method (BEM) and by the method of fundamental solutions (MFS). The bio-inspired methods are applied to optimize shape, topology and material properties of 2D, 3D and coupled 2D/3D structures, to optimize the termomechanical structures, to optimize parameters of composites structures modeled by the FEM, to optimize the elastic vibrating systems to identify the material constants for piezoelectric materials modeled by the BEM and to identify parameters in acoustics problem modeled by the MFS.
Numerical propulsion system simulation: An interdisciplinary approach
NASA Technical Reports Server (NTRS)
Nichols, Lester D.; Chamis, Christos C.
1991-01-01
The tremendous progress being made in computational engineering and the rapid growth in computing power that is resulting from parallel processing now make it feasible to consider the use of computer simulations to gain insights into the complex interactions in aerospace propulsion systems and to evaluate new concepts early in the design process before a commitment to hardware is made. Described here is a NASA initiative to develop a Numerical Propulsion System Simulation (NPSS) capability.
Numerical propulsion system simulation - An interdisciplinary approach
NASA Technical Reports Server (NTRS)
Nichols, Lester D.; Chamis, Christos C.
1991-01-01
The tremendous progress being made in computational engineering and the rapid growth in computing power that is resulting from parallel processing now make it feasible to consider the use of computer simulations to gain insights into the complex interactions in aerospace propulsion systems and to evaluate new concepts early in the design process before a commitment to hardware is made. Described here is a NASA initiative to develop a Numerical Propulsion System Simulation (NPSS) capability.
Rosenthal, L E
1986-10-01
Software is the component in a computer system that permits the hardware to perform the various functions that a computer system is capable of doing. The history of software and its development can be traced to the early nineteenth century. All computer systems are designed to utilize the "stored program concept" as first developed by Charles Babbage in the 1850s. The concept was lost until the mid-1940s, when modern computers made their appearance. Today, because of the complex and myriad tasks that a computer system can perform, there has been a differentiation of types of software. There is software designed to perform specific business applications. There is software that controls the overall operation of a computer system. And there is software that is designed to carry out specialized tasks. Regardless of types, software is the most critical component of any computer system. Without it, all one has is a collection of circuits, transistors, and silicone chips.
Anharmonic Vibrational Spectroscopy on Metal Transition Complexes
NASA Astrophysics Data System (ADS)
Latouche, Camille; Bloino, Julien; Barone, Vincenzo
2014-06-01
Advances in hardware performance and the availability of efficient and reliable computational models have made possible the application of computational spectroscopy to ever larger molecular systems. The systematic interpretation of experimental data and the full characterization of complex molecules can then be facilitated. Focusing on vibrational spectroscopy, several approaches have been proposed to simulate spectra beyond the double harmonic approximation, so that more details become available. However, a routine use of such tools requires the preliminary definition of a valid protocol with the most appropriate combination of electronic structure and nuclear calculation models. Several benchmark of anharmonic calculations frequency have been realized on organic molecules. Nevertheless, benchmarks of organometallics or inorganic metal complexes at this level are strongly lacking despite the interest of these systems due to their strong emission and vibrational properties. Herein we report the benchmark study realized with anharmonic calculations on simple metal complexes, along with some pilot applications on systems of direct technological or biological interest.
Experimentally modeling stochastic processes with less memory by the use of a quantum processor
Palsson, Matthew S.; Gu, Mile; Ho, Joseph; Wiseman, Howard M.; Pryde, Geoff J.
2017-01-01
Computer simulation of observable phenomena is an indispensable tool for engineering new technology, understanding the natural world, and studying human society. However, the most interesting systems are often so complex that simulating their future behavior demands storing immense amounts of information regarding how they have behaved in the past. For increasingly complex systems, simulation becomes increasingly difficult and is ultimately constrained by resources such as computer memory. Recent theoretical work shows that quantum theory can reduce this memory requirement beyond ultimate classical limits, as measured by a process’ statistical complexity, C. We experimentally demonstrate this quantum advantage in simulating stochastic processes. Our quantum implementation observes a memory requirement of Cq = 0.05 ± 0.01, far below the ultimate classical limit of C = 1. Scaling up this technique would substantially reduce the memory required in simulations of more complex systems. PMID:28168218
Optical solver of combinatorial problems: nanotechnological approach.
Cohen, Eyal; Dolev, Shlomi; Frenkel, Sergey; Kryzhanovsky, Boris; Palagushkin, Alexandr; Rosenblit, Michael; Zakharov, Victor
2013-09-01
We present an optical computing system to solve NP-hard problems. As nano-optical computing is a promising venue for the next generation of computers performing parallel computations, we investigate the application of submicron, or even subwavelength, computing device designs. The system utilizes a setup of exponential sized masks with exponential space complexity produced in polynomial time preprocessing. The masks are later used to solve the problem in polynomial time. The size of the masks is reduced to nanoscaled density. Simulations were done to choose a proper design, and actual implementations show the feasibility of such a system.
Lewis hybrid computing system, users manual
NASA Technical Reports Server (NTRS)
Bruton, W. M.; Cwynar, D. S.
1979-01-01
The Lewis Research Center's Hybrid Simulation Lab contains a collection of analog, digital, and hybrid (combined analog and digital) computing equipment suitable for the dynamic simulation and analysis of complex systems. This report is intended as a guide to users of these computing systems. The report describes the available equipment' and outlines procedures for its use. Particular is given to the operation of the PACER 100 digital processor. System software to accomplish the usual digital tasks such as compiling, editing, etc. and Lewis-developed special purpose software are described.
Structural system reliability calculation using a probabilistic fault tree analysis method
NASA Technical Reports Server (NTRS)
Torng, T. Y.; Wu, Y.-T.; Millwater, H. R.
1992-01-01
The development of a new probabilistic fault tree analysis (PFTA) method for calculating structural system reliability is summarized. The proposed PFTA procedure includes: developing a fault tree to represent the complex structural system, constructing an approximation function for each bottom event, determining a dominant sampling sequence for all bottom events, and calculating the system reliability using an adaptive importance sampling method. PFTA is suitable for complicated structural problems that require computer-intensive computer calculations. A computer program has been developed to implement the PFTA.
Electronic Structure of Transition Metal Clusters, Actinide Complexes and Their Reactivities
DOE Office of Scientific and Technical Information (OSTI.GOV)
Krishnan Balasubramanian
2009-07-18
This is a continuing DOE-BES funded project on transition metal and actinide containing species, aimed at the electronic structure and spectroscopy of transition metal and actinide containing species. While a long term connection of these species is to catalysis and environmental management of high-level nuclear wastes, the immediate relevance is directly to other DOE-BES funded experimental projects at DOE-National labs and universities. There are a number of ongoing gas-phase spectroscopic studies of these species at various places, and our computational work has been inspired by these experimental studies and we have also inspired other experimental and theoretical studies. Thus ourmore » studies have varied from spectroscopy of diatomic transition metal carbides to large complexes containing transition metals, and actinide complexes that are critical to the environment. In addition, we are continuing to make code enhancements and modernization of ALCHEMY II set of codes and its interface with relativistic configuration interaction (RCI). At present these codes can carry out multi-reference computations that included up to 60 million configurations and multiple states from each such CI expansion. ALCHEMY II codes have been modernized and converted to a variety of platforms such as Windows XP, and Linux. We have revamped the symbolic CI code to automate the MRSDCI technique so that the references are automatically chosen with a given cutoff from the CASSCF and thus we are doing accurate MRSDCI computations with 10,000 or larger reference space of configurations. The RCI code can also handle a large number of reference configurations, which include up to 10,000 reference configurations. Another major progress is in routinely including larger basis sets up to 5g functions in thee computations. Of course higher angular momenta functions can also be handled using Gaussian and other codes with other methods such as DFT, MP2, CCSD(T), etc. We have also calibrated our RECP methods with all-electron Douglas-Kroll relativistic methods. We have the capabilities for computing full CI extrapolations including spin-orbit effects and several one-electron properties and electron density maps including spin-orbit effects. We are continuously collaborating with several experimental groups around the country and at National Labs to carry out computational studies on the DOE-BES funded projects. The past work in the last 3 years was primarily motivated and driven by the concurrent or recent experimental studies on these systems. We were thus significantly benefited by coordinating our computational efforts with experimental studies. The interaction between theory and experiment has resulted in some unique and exciting opportunities. For example, for the very first time ever, the upper spin-orbit component of a heavy trimer such as Au{sub 3} was experimentally observed as a result of our accurate computational study on the upper electronic states of gold trimer. Likewise for the first time AuH{sub 2} could be observed and interpreted clearly due to our computed potential energy surfaces that revealed the existence of a large barrier to convert the isolated AuH{sub 2} back to Au and H{sub 2}. We have also worked on yet to be observed systems and have made predictions for future experiments. We have computed the spectroscopic and thermodynamic properties of transition metal carbides transition metal clusters and compared our electronic states to the anion photodetachment spectra of Lai Sheng Wang. Prof Mike Morse and coworkers(funded also by DOE-BES) and Prof Stimle and coworkers(also funded by DOE-BES) are working on the spectroscopic properties of transition metal carbides and nitrides. Our predictions on the excited states of transition metal clusters such as Hf{sub 3}, Nb{sub 2}{sup +} etc., have been confirmed experimentally by Prof. Lombardi and coworkers using resonance Raman spectroscopy. We have also been studying larger complexes critical to the environmental management of high-level nuclear wastes. In collaboration with experimental colleague Prof Hieno Nitsche (Berkeley) and Dr. Pat Allen (Livermore, EXAFS) we have studied the uranyl complexes with silicates and carbonates. It should be stressed that although our computed ionization potential of uranium oxide was in conflict with the existing experimental data at the time, a subsequent gas-phase experimental work by Prof Mike Haven and coworkers published as communication in JACS confirmed our computed result to within 0.1 eV. This provides considerable confidence that the computed results in large basis sets with highly-correlated wave functions have excellent accuracies and they have the capabilities to predict the excited states also with great accuracy. Computations of actinide complexes (Uranyl and plutonyl complexes) are critical to management of high-level nuclear wastes.« less
High Performance Computing Based Parallel HIearchical Modal Association Clustering (HPAR HMAC)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Patlolla, Dilip R; Surendran Nair, Sujithkumar; Graves, Daniel A.
For many applications, clustering is a crucial step in order to gain insight into the makeup of a dataset. The best approach to a given problem often depends on a variety of factors, such as the size of the dataset, time restrictions, and soft clustering requirements. The HMAC algorithm seeks to combine the strengths of 2 particular clustering approaches: model-based and linkage-based clustering. One particular weakness of HMAC is its computational complexity. HMAC is not practical for mega-scale data clustering. For high-definition imagery, a user would have to wait months or years for a result; for a 16-megapixel image, themore » estimated runtime skyrockets to over a decade! To improve the execution time of HMAC, it is reasonable to consider an multi-core implementation that utilizes available system resources. An existing imple-mentation (Ray and Cheng 2014) divides the dataset into N partitions - one for each thread prior to executing the HMAC algorithm. This implementation benefits from 2 types of optimization: parallelization and divide-and-conquer. By running each partition in parallel, the program is able to accelerate computation by utilizing more system resources. Although the parallel implementation provides considerable improvement over the serial HMAC, it still suffers from poor computational complexity, O(N2). Once the maximum number of cores on a system is exhausted, the program exhibits slower behavior. We now consider a modification to HMAC that involves a recursive partitioning scheme. Our modification aims to exploit divide-and-conquer benefits seen by the parallel HMAC implementation. At each level in the recursion tree, partitions are divided into 2 sub-partitions until a threshold size is reached. When the partition can no longer be divided without falling below threshold size, the base HMAC algorithm is applied. This results in a significant speedup over the parallel HMAC.« less
Sequential Test Strategies for Multiple Fault Isolation
NASA Technical Reports Server (NTRS)
Shakeri, M.; Pattipati, Krishna R.; Raghavan, V.; Patterson-Hine, Ann; Kell, T.
1997-01-01
In this paper, we consider the problem of constructing near optimal test sequencing algorithms for diagnosing multiple faults in redundant (fault-tolerant) systems. The computational complexity of solving the optimal multiple-fault isolation problem is super-exponential, that is, it is much more difficult than the single-fault isolation problem, which, by itself, is NP-hard. By employing concepts from information theory and Lagrangian relaxation, we present several static and dynamic (on-line or interactive) test sequencing algorithms for the multiple fault isolation problem that provide a trade-off between the degree of suboptimality and computational complexity. Furthermore, we present novel diagnostic strategies that generate a static diagnostic directed graph (digraph), instead of a static diagnostic tree, for multiple fault diagnosis. Using this approach, the storage complexity of the overall diagnostic strategy reduces substantially. Computational results based on real-world systems indicate that the size of a static multiple fault strategy is strictly related to the structure of the system, and that the use of an on-line multiple fault strategy can diagnose faults in systems with as many as 10,000 failure sources.
NASA Astrophysics Data System (ADS)
Miao, Sha; Hendrickson, Kelli; Liu, Yuming
2017-12-01
This work presents a Fully-Coupled Immersed Flow (FCIF) solver for the three-dimensional simulation of fluid-fluid interaction by coupling two distinct flow solvers using an Immersed Boundary (IB) method. The FCIF solver captures dynamic interactions between two fluids with disparate flow properties, while retaining the desirable simplicity of non-boundary-conforming grids. For illustration, we couple an IB-based unsteady Reynolds Averaged Navier Stokes (uRANS) simulator with a depth-integrated (long-wave) solver for the application of slug development with turbulent gas and laminar liquid. We perform a series of validations including turbulent/laminar flows over prescribed wavy boundaries and freely-evolving viscous fluids. These confirm the effectiveness and accuracy of both one-way and two-way coupling in the FCIF solver. Finally, we present a simulation example of the evolution from a stratified turbulent/laminar flow through the initiation of a slug that nearly bridges the channel. The results show both the interfacial wave dynamics excited by the turbulent gas forcing and the influence of the liquid on the gas turbulence. These results demonstrate that the FCIF solver effectively captures the essential physics of gas-liquid interaction and can serve as a useful tool for the mechanistic study of slug generation in two-phase gas/liquid flows in channels and pipes.
NASA Astrophysics Data System (ADS)
Veiga Rodrigues, C.; Palma, J. M. L. M.
2014-06-01
Mesoscale results using the WRF model were downscaled from 3 km to 250 m resolution in a one-way coupling with VENTOS®/M. The results were compared against field measurements at one site comprising 4 meteorological masts, each with two sets of cup anemometers and wind vanes. The results showed that the addition of VENTOS®/M to the model chain improved the wind speed RMSE. Regarding the prediction of wind direction ambivalent results were obtained. Special attention was given to the prediction of turbulence intensity, particularly in reproducing its inverse proportionality with increasing wind speed (cf. IEC 61400-1 standard). The typical use of computational models in wind resource assessment, i.e., relying on decoupled methodologies and neutrally-stratified regimes, does not allow the representation of turbulence intensity for all wind speeds. The results obtained with VENTOS®/M were in agreement with the measured turbulence characteristics at both high and low wind speeds. Such was achieved without the coupling of any turbulence related field, relying solely on the turbulence model embedded in VENTOS®/M and its respective wall boundary conditions, based on Monin-Obukhov similarity theory. The behaviour under different stratification regimes was verified by analysing diurnal and nocturnal events separately.
NASA Astrophysics Data System (ADS)
Cheng, Sheng-Yi; Liu, Wen-Jin; Chen, Shan-Qiu; Dong, Li-Zhi; Yang, Ping; Xu, Bing
2015-08-01
Among all kinds of wavefront control algorithms in adaptive optics systems, the direct gradient wavefront control algorithm is the most widespread and common method. This control algorithm obtains the actuator voltages directly from wavefront slopes through pre-measuring the relational matrix between deformable mirror actuators and Hartmann wavefront sensor with perfect real-time characteristic and stability. However, with increasing the number of sub-apertures in wavefront sensor and deformable mirror actuators of adaptive optics systems, the matrix operation in direct gradient algorithm takes too much time, which becomes a major factor influencing control effect of adaptive optics systems. In this paper we apply an iterative wavefront control algorithm to high-resolution adaptive optics systems, in which the voltages of each actuator are obtained through iteration arithmetic, which gains great advantage in calculation and storage. For AO system with thousands of actuators, the computational complexity estimate is about O(n2) ˜ O(n3) in direct gradient wavefront control algorithm, while the computational complexity estimate in iterative wavefront control algorithm is about O(n) ˜ (O(n)3/2), in which n is the number of actuators of AO system. And the more the numbers of sub-apertures and deformable mirror actuators, the more significant advantage the iterative wavefront control algorithm exhibits. Project supported by the National Key Scientific and Research Equipment Development Project of China (Grant No. ZDYZ2013-2), the National Natural Science Foundation of China (Grant No. 11173008), and the Sichuan Provincial Outstanding Youth Academic Technology Leaders Program, China (Grant No. 2012JQ0012).
Some new traveling wave exact solutions of the (2+1)-dimensional Boiti-Leon-Pempinelli equations.
Qi, Jian-ming; Zhang, Fu; Yuan, Wen-jun; Huang, Zi-feng
2014-01-01
We employ the complex method to obtain all meromorphic exact solutions of complex (2+1)-dimensional Boiti-Leon-Pempinelli equations (BLP system of equations). The idea introduced in this paper can be applied to other nonlinear evolution equations. Our results show that all rational and simply periodic traveling wave exact solutions of the equations (BLP) are solitary wave solutions, the complex method is simpler than other methods, and there exist some rational solutions ur,2 (z) and simply periodic solutions us,2-6(z) which are not only new but also not degenerated successively by the elliptic function solutions. We believe that this method should play an important role for finding exact solutions in the mathematical physics. For these new traveling wave solutions, we give some computer simulations to illustrate our main results.
Substituent Effects on the [N-I-N](+) Halogen Bond.
Carlsson, Anna-Carin C; Mehmeti, Krenare; Uhrbom, Martin; Karim, Alavi; Bedin, Michele; Puttreddy, Rakesh; Kleinmaier, Roland; Neverov, Alexei A; Nekoueishahraki, Bijan; Gräfenstein, Jürgen; Rissanen, Kari; Erdélyi, Máté
2016-08-10
We have investigated the influence of electron density on the three-center [N-I-N](+) halogen bond. A series of [bis(pyridine)iodine](+) and [1,2-bis((pyridine-2-ylethynyl)benzene)iodine](+) BF4(-) complexes substituted with electron withdrawing and donating functionalities in the para-position of their pyridine nitrogen were synthesized and studied by spectroscopic and computational methods. The systematic change of electron density of the pyridine nitrogens upon alteration of the para-substituent (NO2, CF3, H, F, Me, OMe, NMe2) was confirmed by (15)N NMR and by computation of the natural atomic population and the π electron population of the nitrogen atoms. Formation of the [N-I-N](+) halogen bond resulted in >100 ppm (15)N NMR coordination shifts. Substituent effects on the (15)N NMR chemical shift are governed by the π population rather than the total electron population at the nitrogens. Isotopic perturbation of equilibrium NMR studies along with computation on the DFT level indicate that all studied systems possess static, symmetric [N-I-N](+) halogen bonds, independent of their electron density. This was further confirmed by single crystal X-ray diffraction data of 4-substituted [bis(pyridine)iodine](+) complexes. An increased electron density of the halogen bond acceptor stabilizes the [N···I···N](+) bond, whereas electron deficiency reduces the stability of the complexes, as demonstrated by UV-kinetics and computation. In contrast, the N-I bond length is virtually unaffected by changes of the electron density. The understanding of electronic effects on the [N-X-N](+) halogen bond is expected to provide a useful handle for the modulation of the reactivity of [bis(pyridine)halogen](+)-type synthetic reagents.
Substituent Effects on the [N–I–N]+ Halogen Bond
2016-01-01
We have investigated the influence of electron density on the three-center [N–I–N]+ halogen bond. A series of [bis(pyridine)iodine]+ and [1,2-bis((pyridine-2-ylethynyl)benzene)iodine]+ BF4– complexes substituted with electron withdrawing and donating functionalities in the para-position of their pyridine nitrogen were synthesized and studied by spectroscopic and computational methods. The systematic change of electron density of the pyridine nitrogens upon alteration of the para-substituent (NO2, CF3, H, F, Me, OMe, NMe2) was confirmed by 15N NMR and by computation of the natural atomic population and the π electron population of the nitrogen atoms. Formation of the [N–I–N]+ halogen bond resulted in >100 ppm 15N NMR coordination shifts. Substituent effects on the 15N NMR chemical shift are governed by the π population rather than the total electron population at the nitrogens. Isotopic perturbation of equilibrium NMR studies along with computation on the DFT level indicate that all studied systems possess static, symmetric [N–I–N]+ halogen bonds, independent of their electron density. This was further confirmed by single crystal X-ray diffraction data of 4-substituted [bis(pyridine)iodine]+ complexes. An increased electron density of the halogen bond acceptor stabilizes the [N···I···N]+ bond, whereas electron deficiency reduces the stability of the complexes, as demonstrated by UV-kinetics and computation. In contrast, the N–I bond length is virtually unaffected by changes of the electron density. The understanding of electronic effects on the [N–X–N]+ halogen bond is expected to provide a useful handle for the modulation of the reactivity of [bis(pyridine)halogen]+-type synthetic reagents. PMID:27265247
NASA Astrophysics Data System (ADS)
Yeşilkaynak, Tuncay; Özpınar, Celal; Emen, Fatih Mehmet; Ateş, Burhan; Kaya, Kerem
2017-02-01
N-((5-chloropyridin-2-yl)carbamothioyl)furan-2-carboxamide (HL: C11H8ClN3O2S) and its Co(II), Ni(II) and Cu(II) complexes have been synthesized and characterized by elemental analysis, FT-IR,1H NMR and HR-MS methods. The HL was characterized by single crystal X-ray diffraction technique. It crystallizes in the monoclinic system. The HL has the space group P 1 21/c 1, Z = 4, and its unit cell parameters are a = 4.5437(5) Å, b = 22.4550(3) Å, c = 11.8947(14) Å. The ligand coordinates the metal ions as bidentate and thus essentially yields neutral complexes of the [ML2] type. ML2 complex structures were optimized using B97D/TZVP level. Molecular orbitals of both HL ligand were calculated at the same level. Thermal decomposition of the complexes has been investigated by thermogravimetry. The complexes were screened for their anticancer and antioxidant activities. Antioxidant activity of the complexes was determined by using the DPPH and ABTS assays. The anticancer activity of the complexes was studied by using MTT assay in MCF-7 breast cancer cells.
Applications of Computer Technology in Complex Craniofacial Reconstruction.
Day, Kristopher M; Gabrick, Kyle S; Sargent, Larry A
2018-03-01
To demonstrate our use of advanced 3-dimensional (3D) computer technology in the analysis, virtual surgical planning (VSP), 3D modeling (3DM), and treatment of complex congenital and acquired craniofacial deformities. We present a series of craniofacial defects treated at a tertiary craniofacial referral center utilizing state-of-the-art 3D computer technology. All patients treated at our center using computer-assisted VSP, prefabricated custom-designed 3DMs, and/or 3D printed custom implants (3DPCI) in the reconstruction of craniofacial defects were included in this analysis. We describe the use of 3D computer technology to precisely analyze, plan, and reconstruct 31 craniofacial deformities/syndromes caused by: Pierre-Robin (7), Treacher Collins (5), Apert's (2), Pfeiffer (2), Crouzon (1) Syndromes, craniosynostosis (6), hemifacial microsomia (2), micrognathia (2), multiple facial clefts (1), and trauma (3). In select cases where the available bone was insufficient for skeletal reconstruction, 3DPCIs were fabricated using 3D printing. We used VSP in 30, 3DMs in all 31, distraction osteogenesis in 16, and 3DPCIs in 13 cases. Utilizing these technologies, the above complex craniofacial defects were corrected without significant complications and with excellent aesthetic results. Modern 3D technology allows the surgeon to better analyze complex craniofacial deformities, precisely plan surgical correction with computer simulation of results, customize osteotomies, plan distractions, and print 3DPCI, as needed. The use of advanced 3D computer technology can be applied safely and potentially improve aesthetic and functional outcomes after complex craniofacial reconstruction. These techniques warrant further study and may be reproducible in various centers of care.
Bai, Shuming; Song, Kai; Shi, Qiang
2015-05-21
Observations of oscillatory features in the 2D spectra of several photosynthetic complexes have led to diverged opinions on their origins, including electronic coherence, vibrational coherence, and vibronic coherence. In this work, effects of these different types of quantum coherence on ultrafast pump-probe polarization anisotropy are investigated and distinguished. We first simulate the isotropic pump-probe signal and anisotropy decay of the Fenna-Matthews-Olson (FMO) complex using a model with only electronic coherence at low temperature and obtain the same coherence time as in the previous experiment. Then, three model dimer systems with different prespecified quantum coherence are simulated, and the results show that their different spectral characteristics can be used to determine the type of coherence during the spectral process. Finally, we simulate model systems with different electronic-vibrational couplings and reveal the condition in which long time vibronic coherence can be observed in systems like the FMO complex.
Space Station Simulation Computer System (SCS) study for NASA/MSFC. Phased development plan
NASA Technical Reports Server (NTRS)
1990-01-01
NASA's Space Station Freedom Program (SSFP) planning efforts have identified a need for a payload training simulator system to serve as both a training facility and as a demonstrator to validate operational concepts. The envisioned MSFC Payload Training Complex (PTC) required to meet this need will train the Space Station payload scientists, station scientists and ground controllers to operate the wide variety of experiments that will be onboard the Space Station Freedom. The Simulation Computer System (SCS) is made up of computer hardware, software, and workstations that will support the Payload Training Complex at MSFC. The purpose of this SCS Study is to investigate issues related to the SCS, alternative requirements, simulator approaches, and state-of-the-art technologies to develop candidate concepts and designs.
NASA Technical Reports Server (NTRS)
1990-01-01
NASA's Space Station Freedom Program (SSFP) planning efforts have identified a need for a payload training simulator system to serve as both a training facility and as a demonstrator to validate operational concepts. The envisioned MSFC Payload Training Complex (PTC) required to meet this need will train the Space Station payload scientists, station scientists, and ground controllers to operate the wide variety of experiments that will be onboard the Space Station Freedom. The Simulation Computer System (SCS) is made up of the computer hardware, software, and workstations that will support the Payload Training Complex at MSFC. The purpose of this SCS Study is to investigate issues related to the SCS, alternative requirements, simulator approaches, and state-of-the-art technologies to develop candidate concepts and designs.
Space Station Simulation Computer System (SCS) study for NASA/MSFC. Operations concept report
NASA Technical Reports Server (NTRS)
1990-01-01
NASA's Space Station Freedom Program (SSFP) planning efforts have identified a need for a payload training simulator system to serve as both a training facility and as a demonstrator to validate operational concepts. The envisioned MSFC Payload Training Complex (PTC) required to meet this need will train the Space Station payload scientists, station scientists, and ground controllers to operate the wide variety of experiments that will be onboard the Space Station Freedom. The Simulation Computer System (SCS) is made up of computer hardware, software, and workstations that will support the Payload Training Complex at MSFC. The purpose of this SCS Study is to investigate issues related to the SCS, alternative requirements, simulator approaches, and state-of-the-art technologies to develop candidate concepts and designs.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ray, Manisha; Kafader, Jared O.; Topolski, Josey E.
The electronic structures of several small Ce–Pt oxide complexes were explored using a combination of anion photoelectron (PE) spectroscopy and density functional theory calculations. Pt and Pt{sub 2} both accept electron density from CeO diatomic molecules, in which the cerium atom is in a lower-than-bulk oxidation state (+2 versus bulk +4). Neutral [CeO]Pt and [CeO]Pt{sub 2} complexes are therefore ionic, with electronic structures described qualitatively as [CeO{sup +2}]Pt{sup −2} and [CeO{sup +}]Pt{sub 2}{sup −}, respectively. The associated anions are described qualitatively as [CeO{sup +}]Pt{sup −2} and [CeO{sup +}]Pt{sub 2}{sup −2}, respectively. In both neutrals and anions, the most stable molecularmore » structures determined by calculations feature a distinct CeO moiety, with the positively charged Ce center pointing toward the electron rich Pt or Pt{sub 2} moiety. Spectral simulations based on calculated spectroscopic parameters are in fair agreement with the spectra, validating the computationally determined structures. In contrast, when Pt is coupled with CeO{sub 2}, which has no Ce-localized electrons that can readily be donated to Pt, the anion is described as [CeO{sub 2}]Pt{sup −}. The molecular structure predicted computationally suggests that it is governed by charge-dipole interactions. The neutral [CeO{sub 2}]Pt complex lacks charge-dipole stabilizing interactions, and is predicted to be structurally very different from the anion, featuring a single Pt–O–Ce bridge bond. The PE spectra of several of the complexes exhibit evidence of photodissociation with Pt{sup −} daughter ion formation. The electronic structures of these complexes are related to local interactions in Pt-ceria catalyst-support systems.« less
Care 3 phase 2 report, maintenance manual
NASA Technical Reports Server (NTRS)
Bryant, L. A.; Stiffler, J. J.
1982-01-01
CARE 3 (Computer-Aided Reliability Estimation, version three) is a computer program designed to help estimate the reliability of complex, redundant systems. Although the program can model a wide variety of redundant structures, it was developed specifically for fault-tolerant avionics systems--systems distinguished by the need for extremely reliable performance since a system failure could well result in the loss of human life. It substantially generalizes the class of redundant configurations that could be accommodated, and includes a coverage model to determine the various coverage probabilities as a function of the applicable fault recovery mechanisms (detection delay, diagnostic scheduling interval, isolation and recovery delay, etc.). CARE 3 further generalizes the class of system structures that can be modeled and greatly expands the coverage model to take into account such effects as intermittent and transient faults, latent faults, error propagation, etc.
NASA Technical Reports Server (NTRS)
Migneault, Gerard E.
1987-01-01
Emulation techniques can be a solution to a difficulty that arises in the analysis of the reliability of guidance and control computer systems for future commercial aircraft. Described here is the difficulty, the lack of credibility of reliability estimates obtained by analytical modeling techniques. The difficulty is an unavoidable consequence of the following: (1) a reliability requirement so demanding as to make system evaluation by use testing infeasible; (2) a complex system design technique, fault tolerance; (3) system reliability dominated by errors due to flaws in the system definition; and (4) elaborate analytical modeling techniques whose precision outputs are quite sensitive to errors of approximation in their input data. Use of emulation techniques for pseudo-testing systems to evaluate bounds on the parameter values needed for the analytical techniques is then discussed. Finally several examples of the application of emulation techniques are described.
Computer Systems for Teaching Complex Concepts.
ERIC Educational Resources Information Center
Feurzeig, Wallace
Four Programing systems--Mentor, Stringcomp, Simon, and Logo--were designed and implemented as integral parts of research into the various ways computers may be used for teaching problem-solving concepts and skills. Various instructional contexts, among them medicine, mathematics, physics, and basic problem-solving for elementary school children,…
Computational Systems for Multidisciplinary Applications
NASA Technical Reports Server (NTRS)
Soni, Bharat; Haupt, Tomasz; Koomullil, Roy; Luke, Edward; Thompson, David
2002-01-01
In this paper, we briefly describe our efforts to develop complex simulation systems. We focus first on four key infrastructure items: enterprise computational services, simulation synthesis, geometry modeling and mesh generation, and a fluid flow solver for arbitrary meshes. We conclude by presenting three diverse applications developed using these technologies.
van Weerdenburg, Bram J A; Engwerda, Anthonius H J; Eshuis, Nan; Longo, Alessandro; Banerjee, Dipanjan; Tessari, Marco; Guerra, Célia Fonseca; Rutjes, Floris P J T; Bickelhaupt, F Matthias; Feiters, Martin C
2015-07-13
Signal amplification by reversible exchange (SABRE) is an emerging hyperpolarization method in NMR spectroscopy, in which hyperpolarization is transferred through the scalar coupling network of para-hydrogen derived hydrides in a metal complex to a reversibly bound substrate. Substrates can even be hyperpolarized at concentrations below that of the metal complex by addition of a suitable co-substrate. Here we investigate the catalytic system used for trace detection in NMR spectroscopy with [Ir(IMes)(H)2 (L)3 ](+) (IMes=1,3-dimesitylimidazol-2-ylidene) as catalyst, pyridine as a substrate and 1-methyl-1,2,3-triazole as co-substrate in great detail. With density functional theory (DFT), validated by extended X-ray absorption fine structure (EXAFS) experiments, we provide explanations for the relative abundance of the observed metal complexes, as well as their contribution to SABRE. We have established that the interaction between iridium and ligands cis to IMes is weaker than that with the trans ligand, and that in mixed complexes with pyridine and triazole, the latter preferentially takes up the trans position. © 2015 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.