Sample records for calculation grid size

  1. SU-E-T-374: Evaluation and Verification of Dose Calculation Accuracy with Different Dose Grid Sizes for Intracranial Stereotactic Radiosurgery

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Han, C; Schultheiss, T

    Purpose: In this study, we aim to evaluate the effect of dose grid size on the accuracy of calculated dose for small lesions in intracranial stereotactic radiosurgery (SRS), and to verify dose calculation accuracy with radiochromic film dosimetry. Methods: 15 intracranial lesions from previous SRS patients were retrospectively selected for this study. The planning target volume (PTV) ranged from 0.17 to 2.3 cm{sup 3}. A commercial treatment planning system was used to generate SRS plans using the volumetric modulated arc therapy (VMAT) technique using two arc fields. Two convolution-superposition-based dose calculation algorithms (Anisotropic Analytical Algorithm and Acuros XB algorithm) weremore » used to calculate volume dose distribution with dose grid size ranging from 1 mm to 3 mm with 0.5 mm step size. First, while the plan monitor units (MU) were kept constant, PTV dose variations were analyzed. Second, with 95% of the PTV covered by the prescription dose, variations of the plan MUs as a function of dose grid size were analyzed. Radiochomic films were used to compare the delivered dose and profile with the calculated dose distribution with different dose grid sizes. Results: The dose to the PTV, in terms of the mean dose, maximum, and minimum dose, showed steady decrease with increasing dose grid size using both algorithms. With 95% of the PTV covered by the prescription dose, the total MU increased with increasing dose grid size in most of the plans. Radiochromic film measurements showed better agreement with dose distributions calculated with 1-mm dose grid size. Conclusion: Dose grid size has significant impact on calculated dose distribution in intracranial SRS treatment planning with small target volumes. Using the default dose grid size could lead to under-estimation of delivered dose. A small dose grid size should be used to ensure calculation accuracy and agreement with QA measurements.« less

  2. Impact of grid size on uniform scanning and IMPT plans in XiO treatment planning system for brain cancer

    PubMed Central

    Zheng, Yuanshui

    2015-01-01

    The main purposes of this study are to: 1) evaluate the accuracy of XiO treatment planning system (TPS) for different dose calculation grid size based on head phantom measurements in uniform scanning proton therapy (USPT); and 2) compare the dosimetric results for various dose calculation grid sizes based on real computed tomography (CT) dataset of pediatric brain cancer treatment plans generated by USPT and intensity‐modulated proton therapy (IMPT) techniques. For phantom study, we have utilized the anthropomorphic head proton phantom provided by Imaging and Radiation Oncology Core (IROC). The imaging, treatment planning, and beam delivery were carried out following the guidelines provided by the IROC. The USPT proton plan was generated in the XiO TPS, and dose calculations were performed for grid size ranged from 1 to 3 mm. The phantom containing thermoluminescent dosimeter (TLDs) and films was irradiated using uniform scanning proton beam. The irradiated TLDs were read by the IROC. The calculated doses from the XiO for different grid sizes were compared to the measured TLD doses provided by the IROC. Gamma evaluation was done by comparing calculated planar dose distribution of 3 mm grid size with measured planar dose distribution. Additionally, IMPT plan was generated based on the same CT dataset of the IROC phantom, and IMPT dose calculations were performed for grid size ranged from 1 to 3 mm. For comparative purpose, additional gamma analysis was done by comparing the planar dose distributions of standard grid size (3 mm) with that of other grid sizes (1, 1.5, 2, and 2.5 mm) for both the USPT and IMPT plans. For patient study, USPT plans of three pediatric brain cancer cases were selected. IMPT plans were generated for each of three pediatric cases. All patient treatment plans (USPT and IMPT) were generated in the XiO TPS for a total dose of 54 Gy (relative biological effectiveness [RBE]). Treatment plans (USPT and IMPT) of each case was recalculated for grid sizes of 1, 1.5, 2, and 2.5 mm; these dosimetric results were then compared with that of 3 mm grid size. Phantom study results: There was no distinct trend exhibiting the dependence of grid size on dose calculation accuracy when calculated point dose of different grid sizes were compared to the measured point (TLD) doses. On average, the calculated point dose was higher than the measured dose by 1.49% and 2.63% for the right and left TLDs, respectively. The gamma analysis showed very minimal differences among planar dose distributions of various grid sizes, with percentage of points meeting gamma index criteria 1% and 1 mm to be from 97.92% to 99.97%. The gamma evaluation using 2% and 2 mm criteria showed both the IMPT and USPT plans have 100% points meeting the criteria. Patient study results: In USPT, there was no very distinct relationship between the absolute difference in mean planning target volume (PTV) dose and grid size, whereas in IMPT, it was found that the decrease in grid size slightly increased the PTV maximum dose and decreased the PTV mean dose and PTV D50%. For the PTV doses, the average differences were up to 0.35 Gy (RBE) and 1.47 Gy (RBE) in the USPT and IMPT plans, respectively. Dependency on grid size was not very clear for the organs at risk (OARs), with average difference ranged from −0.61 Gy (RBE) to 0.53 Gy (RBE) in the USPT plans and from −0.83 Gy (RBE) to 1.39 Gy (RBE) in the IMPT plans. In conclusion, the difference in the calculated point dose between the smallest grid size (1 mm) and the largest grid size (3 mm) in phantom for USPT was typically less than 0.1%. Patient study results showed that the decrease in grid size slightly increased the PTV maximum dose in both the USPT and IMPT plans. However, no distinct trend was obtained between the absolute difference in dosimetric parameter and dose calculation grid size for the OARs. Grid size has a large effect on dose calculation efficiency, and use of 2 mm or less grid size can increase the dose calculation time significantly. It is recommended to use grid size either 2.5 or 3 mm for dose calculations of pediatric brain cancer plans generated by USPT and IMPT techniques in XiO TPS. PACS numbers: 87.55.D‐, 87.55.ne, 87.55.dk PMID:26699310

  3. SU-E-T-454: Impact of Calculation Grid Size On Dosimetry and Radiobiological Parameters for Head and Neck IMRT

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Srivastava, S; Das, I; Indiana University Health Methodist Hospital, Indianapolis, IN

    2014-06-01

    Purpose: IMRT has become standard of care for complex treatments to optimize dose to target and spare normal tissues. However, the impact of calculation grid size is not widely known especially dose distribution, tumor control probability (TCP) and normal tissue complication probability (NTCP) which is investigated in this study. Methods: Ten head and neck IMRT patients treated with 6 MV photons were chosen for this study. Using Eclipse TPS, treatment plans were generated for different grid sizes in the range 1–5 mm for the same optimization criterion with specific dose-volume constraints. The dose volume histogram (DVH) was calculated for allmore » IMRT plans and dosimetric data were compared. ICRU-83 dose points such as D2%, D50%, D98%, as well as the homogeneity and conformity indices (HI, CI) were calculated. In addition, TCP and NTCP were calculated from DVH data. Results: The PTV mean dose and TCP decreases with increasing grid size with an average decrease in mean dose by 2% and TCP by 3% respectively. Increasing grid size from 1–5 mm grid size, the average mean dose and NTCP for left parotid was increased by 6.0% and 8.0% respectively. Similar patterns were observed for other OARs such as cochlea, parotids and spinal cord. The HI increases up to 60% and CI decreases on average by 3.5% between 1 and 5 mm grid that resulted in decreased TCP and increased NTCP values. The number of points meeting the gamma criteria of ±3% dose difference and ±3mm DTA was higher with a 1 mm on average (97.2%) than with a 5 mm grid (91.3%). Conclusion: A smaller calculation grid provides superior dosimetry with improved TCP and reduced NTCP values. The effect is more pronounced for smaller OARs. Thus, the smallest possible grid size should be used for accurate dose calculation especially in H and N planning.« less

  4. Adaptive grid generation in a patient-specific cerebral aneurysm

    NASA Astrophysics Data System (ADS)

    Hodis, Simona; Kallmes, David F.; Dragomir-Daescu, Dan

    2013-11-01

    Adapting grid density to flow behavior provides the advantage of increasing solution accuracy while decreasing the number of grid elements in the simulation domain, therefore reducing the computational time. One method for grid adaptation requires successive refinement of grid density based on observed solution behavior until the numerical errors between successive grids are negligible. However, such an approach is time consuming and it is often neglected by the researchers. We present a technique to calculate the grid size distribution of an adaptive grid for computational fluid dynamics (CFD) simulations in a complex cerebral aneurysm geometry based on the kinematic curvature and torsion calculated from the velocity field. The relationship between the kinematic characteristics of the flow and the element size of the adaptive grid leads to a mathematical equation to calculate the grid size in different regions of the flow. The adaptive grid density is obtained such that it captures the more complex details of the flow with locally smaller grid size, while less complex flow characteristics are calculated on locally larger grid size. The current study shows that kinematic curvature and torsion calculated from the velocity field in a cerebral aneurysm can be used to find the locations of complex flow where the computational grid needs to be refined in order to obtain an accurate solution. We found that the complexity of the flow can be adequately described by velocity and vorticity and the angle between the two vectors. For example, inside the aneurysm bleb, at the bifurcation, and at the major arterial turns the element size in the lumen needs to be less than 10% of the artery radius, while at the boundary layer, the element size should be smaller than 1% of the artery radius, for accurate results within a 0.5% relative approximation error. This technique of quantifying flow complexity and adaptive remeshing has the potential to improve results accuracy and reduce computational time for patient-specific hemodynamics simulations, which are used to help assess the likelihood of aneurysm rupture using CFD calculated flow patterns.

  5. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Pokharel, S; Rana, S

    Purpose: purpose of this study is to evaluate the effect of grid size in Eclipse AcurosXB dose calculation algorithm for SBRT lung. Methods: Five cases of SBRT lung previously treated have been chosen for present study. Four of the plans were 5 fields conventional IMRT and one was Rapid Arc plan. All five cases have been calculated with five grid sizes (1, 1.5, 2, 2.5 and 3mm) available for AXB algorithm with same plan normalization. Dosimetric indices relevant to SBRT along with MUs and time have been recorded for different grid sizes. The maximum difference was calculated as a percentagemore » of mean of all five values. All the plans were IMRT QAed with portal dosimetry. Results: The maximum difference of MUs was within 2%. The time increased was as high as 7 times from highest 3mm to lowest 1mm grid size. The largest difference of PTV minimum, maximum and mean dose were 7.7%, 1.5% and 1.6% respectively. The highest D2-Max difference was 6.1%. The highest difference in ipsilateral lung mean, V5Gy, V10Gy and V20Gy were 2.6%, 2.4%, 1.9% and 3.8% respectively. The maximum difference of heart, cord and esophagus dose were 6.5%, 7.8% and 4.02% respectively. The IMRT Gamma passing rate at 2%/2mm remains within 1.5% with at least 98% points passing with all grid sizes. Conclusion: This work indicates the lowest grid size of 1mm available in AXB is not necessarily required for accurate dose calculation. The IMRT passing rate was insignificant or not observed with the reduction of grid size less than 2mm. Although the maximum percentage difference of some of the dosimetric indices appear large, most of them are clinically insignificant in absolute dose values. So we conclude that 2mm grid size calculation is best compromise in light of dose calculation accuracy and time it takes to calculate dose.« less

  6. The influence of the dose calculation resolution of VMAT plans on the calculated dose for eye lens and optic pathway.

    PubMed

    Park, Jong Min; Park, So-Yeon; Kim, Jung-In; Carlson, Joel; Kim, Jin Ho

    2017-03-01

    To investigate the effect of dose calculation grid on calculated dose-volumetric parameters for eye lenses and optic pathways. A total of 30 patients treated using the volumetric modulated arc therapy (VMAT) technique, were retrospectively selected. For each patient, dose distributions were calculated with calculation grids ranging from 1 to 5 mm at 1 mm intervals. Identical structures were used for VMAT planning. The changes in dose-volumetric parameters according to the size of the calculation grid were investigated. Compared to dose calculation with 1 mm grid, the maximum doses to the eye lens with calculation grids of 2, 3, 4 and 5 mm increased by 0.2 ± 0.2 Gy, 0.5 ± 0.5 Gy, 0.9 ± 0.8 Gy and 1.7 ± 1.5 Gy on average, respectively. The Spearman's correlation coefficient between dose gradients near structures vs. the differences between the calculated doses with 1 mm grid and those with 5 mm grid, were 0.380 (p < 0.001). For the accurate calculation of dose distributions, as well as efficiency, using a grid size of 2 mm appears to be the most appropriate choice.

  7. The Impact of the Grid Size on TomoTherapy for Prostate Cancer

    PubMed Central

    Kawashima, Motohiro; Kawamura, Hidemasa; Onishi, Masahiro; Takakusagi, Yosuke; Okonogi, Noriyuki; Okazaki, Atsushi; Sekihara, Tetsuo; Ando, Yoshitaka; Nakano, Takashi

    2017-01-01

    Discretization errors due to the digitization of computed tomography images and the calculation grid are a significant issue in radiation therapy. Such errors have been quantitatively reported for a fixed multifield intensity-modulated radiation therapy using traditional linear accelerators. The aim of this study is to quantify the influence of the calculation grid size on the dose distribution in TomoTherapy. This study used ten treatment plans for prostate cancer. The final dose calculation was performed with “fine” (2.73 mm) and “normal” (5.46 mm) grid sizes. The dose distributions were compared from different points of view: the dose-volume histogram (DVH) parameters for planning target volume (PTV) and organ at risk (OAR), the various indices, and dose differences. The DVH parameters were used Dmax, D2%, D2cc, Dmean, D95%, D98%, and Dmin for PTV and Dmax, D2%, and D2cc for OARs. The various indices used were homogeneity index and equivalent uniform dose for plan evaluation. Almost all of DVH parameters for the “fine” calculations tended to be higher than those for the “normal” calculations. The largest difference of DVH parameters for PTV was Dmax and that for OARs was rectal D2cc. The mean difference of Dmax was 3.5%, and the rectal D2cc was increased up to 6% at the maximum and 2.9% on average. The mean difference of D95% for PTV was the smallest among the differences of the other DVH parameters. For each index, whether there was a significant difference between the two grid sizes was determined through a paired t-test. There were significant differences for most of the indices. The dose difference between the “fine” and “normal” calculations was evaluated. Some points around high-dose regions had differences exceeding 5% of the prescription dose. The influence of the calculation grid size in TomoTherapy is smaller than traditional linear accelerators. However, there was a significant difference. We recommend calculating the final dose using the “fine” grid size. PMID:28974860

  8. Verification of the grid size and angular increment effects in lung stereotactic body radiation therapy using the dynamic conformal arc technique

    NASA Astrophysics Data System (ADS)

    Park, Hae-Jin; Suh, Tae-Suk; Park, Ji-Yeon; Lee, Jeong-Woo; Kim, Mi-Hwa; Oh, Young-Taek; Chun, Mison; Noh, O. Kyu; Suh, Susie

    2013-06-01

    The dosimetric effects of variable grid size and angular increment were systematically evaluated in the measured dose distributions of dynamic conformal arc therapy (DCAT) for lung stereotactic body radiation therapy (SBRT). Dose variations with different grid sizes (2, 3, and 4 mm) and angular increments (2, 4, 6, and 10°) for spherical planning target volumes (PTVs) were verified in a thorax phantom by using EBT2 films. Although the doses for identical PTVs were predicted for the different grid sizes, the dose discrepancy was evaluated using one measured dose distribution with the gamma tool because the beam was delivered in the same set-up for DCAT. The dosimetric effect of the angular increment was verified by comparing the measured dose area histograms of organs at risk (OARs) at each angular increment. When the difference in the OAR doses is higher than the uncertainty of the film dosimetry, the error is regarded as the angular increment effect in discretely calculated doses. In the results, even when a 2-mm grid size was used with an elaborate dose calculation, 4-mm grid size led to a higher gamma pass ratio due to underdosage, a steep-dose descent gradient, and lower estimated PTV doses caused by the smoothing effect in the calculated dose distribution. An undulating dose distribution and a difference in the maximum contralateral lung dose of up to 14% were observed in dose calculation using a 10° angular increment. The DCAT can be effectively applied for an approximately spherical PTV in a relatively uniform geometry, which is less affected by inhomogeneous materials and differences in the beam path length.

  9. SU-E-T-196: Comparative Analysis of Surface Dose Measurements Using MOSFET Detector and Dose Predicted by Eclipse - AAA with Varying Dose Calculation Grid Size

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Badkul, R; Nejaiman, S; Pokhrel, D

    2015-06-15

    Purpose: Skin dose can be the limiting factor and fairly common reason to interrupt the treatment, especially for treating head-and-neck with Intensity-modulated-radiation-therapy(IMRT) or Volumetrically-modulated - arc-therapy (VMAT) and breast with tangentially-directed-beams. Aim of this study was to investigate accuracy of near-surface dose predicted by Eclipse treatment-planning-system (TPS) using Anisotropic-Analytic Algorithm (AAA)with varying calculation grid-size and comparing with metal-oxide-semiconductor-field-effect-transistors(MOSFETs)measurements for a range of clinical-conditions (open-field,dynamic-wedge, physical-wedge, IMRT,VMAT). Methods: QUASAR™-Body-Phantom was used in this study with oval curved-surfaces to mimic breast, chest wall and head-and-neck sites.A CT-scan was obtained with five radio-opaque markers(ROM) placed on the surface of phantom to mimic themore » range of incident angles for measurements and dose prediction using 2mm slice thickness.At each ROM, small structure(1mmx2mm) were contoured to obtain mean-doses from TPS.Calculations were performed for open-field,dynamic-wedge,physical-wedge,IMRT and VMAT using Varian-21EX,6&15MV photons using twogrid-sizes:2.5mm and 1mm.Calibration checks were performed to ensure that MOSFETs response were within ±5%.Surface-doses were measured at five locations and compared with TPS calculations. Results: For 6MV: 2.5mm grid-size,mean calculated doses(MCD)were higher by 10%(±7.6),10%(±7.6),20%(±8.5),40%(±7.5),30%(±6.9) and for 1mm grid-size MCD were higher by 0%(±5.7),0%(±4.2),0%(±5.5),1.2%(±5.0),1.1% (±7.8) for open-field,dynamic-wedge,physical-wedge,IMRT,VMAT respectively.For 15MV: 2.5mm grid-size,MCD were higher by 30%(±14.6),30%(±14.6),30%(±14.0),40%(±11.0),30%(±3.5)and for 1mm grid-size MCD were higher by 10% (±10.6), 10%(±9.8),10%(±8.0),30%(±7.8),10%(±3.8) for open-field, dynamic-wedge, physical-wedge, IMRT, VMAT respectively.For 6MV, 86% and 56% of all measured values agreed better than ±20% for 1mm and 2.5mm grid-sizes respectively. For 18MV, 56% and 18% of all measured-values agreed better than ±20% for 1mm and 2.5mm grid-sizes respectively. Conclusion: Reliable Skin-dose calculations by TPS can be very difficult due to steep dose-gradient and inaccurate beam-modelling in buildup region.Our results showed that Eclipse over-estimates surface-dose.Impact of grid-size is also significant,surface-dose increased up to 40% from 1mm to 2.5mm,however, 1mm calculated-values closely agrees with measurements. Due to large uncertnities in skin-dose predictions from TPS, outmost caution must be exercised when skin dose is evaluated,a sufficiently smaller grid-size(1mm)can improve the accuracy and MOSFETs can be used for verification.« less

  10. SU-E-T-538: Evaluation of IMRT Dose Calculation Based on Pencil-Beam and AAA Algorithms.

    PubMed

    Yuan, Y; Duan, J; Popple, R; Brezovich, I

    2012-06-01

    To evaluate the accuracy of dose calculation for intensity modulated radiation therapy (IMRT) based on Pencil Beam (PB) and Analytical Anisotropic Algorithm (AAA) computation algorithms. IMRT plans of twelve patients with different treatment sites, including head/neck, lung and pelvis, were investigated. For each patient, dose calculation with PB and AAA algorithms using dose grid sizes of 0.5 mm, 0.25 mm, and 0.125 mm, were compared with composite-beam ion chamber and film measurements in patient specific QA. Discrepancies between the calculation and the measurement were evaluated by percentage error for ion chamber dose and γ〉l failure rate in gamma analysis (3%/3mm) for film dosimetry. For 9 patients, ion chamber dose calculated with AAA-algorithms is closer to ion chamber measurement than that calculated with PB algorithm with grid size of 2.5 mm, though all calculated ion chamber doses are within 3% of the measurements. For head/neck patients and other patients with large treatment volumes, γ〉l failure rate is significantly reduced (within 5%) with AAA-based treatment planning compared to generally more than 10% with PB-based treatment planning (grid size=2.5 mm). For lung and brain cancer patients with medium and small treatment volumes, γ〉l failure rates are typically within 5% for both AAA and PB-based treatment planning (grid size=2.5 mm). For both PB and AAA-based treatment planning, improvements of dose calculation accuracy with finer dose grids were observed in film dosimetry of 11 patients and in ion chamber measurements for 3 patients. AAA-based treatment planning provides more accurate dose calculation for head/neck patients and other patients with large treatment volumes. Compared with film dosimetry, a γ〉l failure rate within 5% can be achieved for AAA-based treatment planning. © 2012 American Association of Physicists in Medicine.

  11. Grid-size dependence of Cauchy boundary conditions used to simulate stream-aquifer interactions

    USGS Publications Warehouse

    Mehl, S.; Hill, M.C.

    2010-01-01

    This work examines the simulation of stream–aquifer interactions as grids are refined vertically and horizontally and suggests that traditional methods for calculating conductance can produce inappropriate values when the grid size is changed. Instead, different grid resolutions require different estimated values. Grid refinement strategies considered include global refinement of the entire model and local refinement of part of the stream. Three methods of calculating the conductance of the Cauchy boundary conditions are investigated. Single- and multi-layer models with narrow and wide streams produced stream leakages that differ by as much as 122% as the grid is refined. Similar results occur for globally and locally refined grids, but the latter required as little as one-quarter the computer execution time and memory and thus are useful for addressing some scale issues of stream–aquifer interactions. Results suggest that existing grid-size criteria for simulating stream–aquifer interactions are useful for one-layer models, but inadequate for three-dimensional models. The grid dependence of the conductance terms suggests that values for refined models using, for example, finite difference or finite-element methods, cannot be determined from previous coarse-grid models or field measurements. Our examples demonstrate the need for a method of obtaining conductances that can be translated to different grid resolutions and provide definitive test cases for investigating alternative conductance formulations.

  12. Fast and accurate 3D tensor calculation of the Fock operator in a general basis

    NASA Astrophysics Data System (ADS)

    Khoromskaia, V.; Andrae, D.; Khoromskij, B. N.

    2012-11-01

    The present paper contributes to the construction of a “black-box” 3D solver for the Hartree-Fock equation by the grid-based tensor-structured methods. It focuses on the calculation of the Galerkin matrices for the Laplace and the nuclear potential operators by tensor operations using the generic set of basis functions with low separation rank, discretized on a fine N×N×N Cartesian grid. We prove the Ch2 error estimate in terms of mesh parameter, h=O(1/N), that allows to gain a guaranteed accuracy of the core Hamiltonian part in the Fock operator as h→0. However, the commonly used problem adapted basis functions have low regularity yielding a considerable increase of the constant C, hence, demanding a rather large grid-size N of about several tens of thousands to ensure the high resolution. Modern tensor-formatted arithmetics of complexity O(N), or even O(logN), practically relaxes the limitations on the grid-size. Our tensor-based approach allows to improve significantly the standard basis sets in quantum chemistry by including simple combinations of Slater-type, local finite element and other basis functions. Numerical experiments for moderate size organic molecules show efficiency and accuracy of grid-based calculations to the core Hamiltonian in the range of grid parameter N3˜1015.

  13. Application of a multi-level grid method to transonic flow calculations

    NASA Technical Reports Server (NTRS)

    South, J. C., Jr.; Brandt, A.

    1976-01-01

    A multi-level grid method was studied as a possible means of accelerating convergence in relaxation calculations for transonic flows. The method employs a hierarchy of grids, ranging from very coarse to fine. The coarser grids are used to diminish the magnitude of the smooth part of the residuals. The method was applied to the solution of the transonic small disturbance equation for the velocity potential in conservation form. Nonlifting transonic flow past a parabolic arc airfoil is studied with meshes of both constant and variable step size.

  14. Assessment of an Euler-Interacting Boundary Layer Method Using High Reynolds Number Transonic Flight Data

    NASA Technical Reports Server (NTRS)

    Bonhaus, Daryl L.; Maddalon, Dal V.

    1998-01-01

    Flight-measured high Reynolds number turbulent-flow pressure distributions on a transport wing in transonic flow are compared to unstructured-grid calculations to assess the predictive ability of a three-dimensional Euler code (USM3D) coupled to an interacting boundary layer module. The two experimental pressure distributions selected for comparative analysis with the calculations are complex and turbulent but typical of an advanced technology laminar flow wing. An advancing front method (VGRID) was used to generate several tetrahedral grids for each test case. Initial calculations left considerable room for improvement in accuracy. Studies were then made of experimental errors, transition location, viscous effects, nacelle flow modeling, number and placement of spanwise boundary layer stations, and grid resolution. The most significant improvements in the accuracy of the calculations were gained by improvement of the nacelle flow model and by refinement of the computational grid. Final calculations yield results in close agreement with the experiment. Indications are that further grid refinement would produce additional improvement but would require more computer memory than is available. The appendix data compare the experimental attachment line location with calculations for different grid sizes. Good agreement is obtained between the experimental and calculated attachment line locations.

  15. Sensitivity of LES results from turbine rim seals to changes in grid resolution and sector size

    NASA Astrophysics Data System (ADS)

    O'Mahoney, T.; Hills, N.; Chew, J.

    2012-07-01

    Large-Eddy Simulations (LES) were carried out for a turbine rim seal and the sensitivity of the results to changes in grid resolution and the size of the computational domain are investigated. Ingestion of hot annulus gas into the rotor-stator cavity is compared between LES results and against experiments and Unsteady Reynolds-Averaged Navier-Stokes (URANS) calculations. The LES calculations show greater ingestion than the URANS calculation and show better agreement with experiments. Increased grid resolution shows a small improvement in ingestion predictions whereas increasing the sector model size has little effect on the results. The contrast between the different CFD models is most stark in the inner cavity, where the URANS shows almost no ingestion. Particular attention is also paid to the presence of low frequency oscillations in the disc cavity. URANS calculations show such low frequency oscillations at different frequencies than the LES. The oscillations also take a very long time to develop in the LES. The results show that the difficult problem of estimating ingestion through rim seals could be overcome by using LES but that the computational requirements were still restrictive.

  16. A method of selecting grid size to account for Hertz deformation in finite element analysis of spur gears

    NASA Technical Reports Server (NTRS)

    Coy, J. J.; Chao, C. H. C.

    1981-01-01

    A method of selecting grid size for the finite element analysis of gear tooth deflection is presented. The method is based on a finite element study of two cylinders in line contact, where the criterion for establishing element size was that there be agreement with the classical Hertzian solution for deflection. The results are applied to calculate deflection for the gear specimen used in the NASA spur gear test rig. Comparisons are made between the present results and the results of two other methods of calculation. The results have application in design of gear tooth profile modifications to reduce noise and dynamic loads.

  17. Diffraction Analysis of Antennas With Mesh Surfaces

    NASA Technical Reports Server (NTRS)

    Rahmat-Samii, Yahya

    1987-01-01

    Strip-aperture model replaces wire-grid model. Far-field radiation pattern of antenna with mesh reflector calculated more accurately with new strip-aperture model than with wire-grid model of reflector surface. More adaptable than wire-grid model to variety of practical configurations and decidedly superior for reflectors in which mesh-cell width exceeds mesh thickness. Satisfies reciprocity theorem. Applied where mesh cells are no larger than tenth of wavelength. Small cell size permits use of simplifying approximation that reflector-surface current induced by electromagnetic field is present even in apertures. Approximation useful in calculating far field.

  18. Influence of Terraced area DEM Resolution on RUSLE LS Factor

    NASA Astrophysics Data System (ADS)

    Zhang, Hongming; Baartman, Jantiene E. M.; Yang, Xiaomei; Gai, Lingtong; Geissen, Viollette

    2017-04-01

    Topography has a large impact on the erosion of soil by water. Slope steepness and slope length are combined (the LS factor) in the universal soil-loss equation (USLE) and its revised version (RUSLE) for predicting soil erosion. The LS factor is usually extracted from a digital elevation model (DEM). The grid size of the DEM will thus influence the LS factor and the subsequent calculation of soil loss. Terracing is considered as a support practice factor (P) in the USLE/RUSLE equations, which is multiplied with the other USLE/RUSLE factors. However, as terraces change the slope length and steepness, they also affect the LS factor. The effect of DEM grid size on the LS factor has not been investigated for a terraced area. We obtained a high-resolution DEM by unmanned aerial vehicles (UAVs) photogrammetry, from which the slope steepness, slope length, and LS factor were extracted. The changes in these parameters at various DEM resolutions were then analysed. The DEM produced detailed LS-factor maps, particularly for low LS factors. High (small valleys, gullies, and terrace ridges) and low (flats and terrace fields) spatial frequencies were both sensitive to changes in resolution, so the areas of higher and lower slope steepness both decreased with increasing grid size. Average slope steepness decreased and average slope length increased with grid size. Slope length, however, had a larger effect than slope steepness on the LS factor as the grid size varied. The LS factor increased when the grid size increased from 0.5 to 30-m and increased significantly at grid sizes >5-m. The LS factor was increasingly overestimated as grid size decreased. The LS factor decreased from grid sizes of 30 to 100-m, because the details of the terraced terrain were gradually lost, but the factor was still overestimated.

  19. Numerical Nuclear Second Derivatives on a Computing Grid: Enabling and Accelerating Frequency Calculations on Complex Molecular Systems.

    PubMed

    Yang, Tzuhsiung; Berry, John F

    2018-06-04

    The computation of nuclear second derivatives of energy, or the nuclear Hessian, is an essential routine in quantum chemical investigations of ground and transition states, thermodynamic calculations, and molecular vibrations. Analytic nuclear Hessian computations require the resolution of costly coupled-perturbed self-consistent field (CP-SCF) equations, while numerical differentiation of analytic first derivatives has an unfavorable 6 N ( N = number of atoms) prefactor. Herein, we present a new method in which grid computing is used to accelerate and/or enable the evaluation of the nuclear Hessian via numerical differentiation: NUMFREQ@Grid. Nuclear Hessians were successfully evaluated by NUMFREQ@Grid at the DFT level as well as using RIJCOSX-ZORA-MP2 or RIJCOSX-ZORA-B2PLYP for a set of linear polyacenes with systematically increasing size. For the larger members of this group, NUMFREQ@Grid was found to outperform the wall clock time of analytic Hessian evaluation; at the MP2 or B2LYP levels, these Hessians cannot even be evaluated analytically. We also evaluated a 156-atom catalytically relevant open-shell transition metal complex and found that NUMFREQ@Grid is faster (7.7 times shorter wall clock time) and less demanding (4.4 times less memory requirement) than an analytic Hessian. Capitalizing on the capabilities of parallel grid computing, NUMFREQ@Grid can outperform analytic methods in terms of wall time, memory requirements, and treatable system size. The NUMFREQ@Grid method presented herein demonstrates how grid computing can be used to facilitate embarrassingly parallel computational procedures and is a pioneer for future implementations.

  20. Effect of particle size distribution of maize and soybean meal on the precaecal amino acid digestibility in broiler chickens.

    PubMed

    Siegert, W; Ganzer, C; Kluth, H; Rodehutscord, M

    2018-02-01

    1. Herein, it was investigated whether different particle size distributions of feed ingredients achieved by grinding through a 2- or 3-mm grid would have an effect on precaecal (pc) amino acid (AA) digestibility. Maize and soybean meal were used as the test ingredients. 2. Maize and soybean meal was ground with grid sizes of 2 or 3 mm. Nine diets were prepared. The basal diet contained 500 g/kg of maize starch. The other experimental diets contained maize or soybean meal samples at concentrations of 250 and 500, and 150 and 300 g/kg, respectively, instead of maize starch. Each diet was tested using 6 replicate groups of 10 birds each. The regression approach was applied to calculate the pc AA digestibility of the test ingredients. 3. The reduction of the grid size from 3 to 2 mm reduced the average particle size of both maize and soybean meal, mainly by reducing the proportion of coarse particles. Reducing the grid size significantly (P < 0.050) increased the pc digestibility of all AA in the soybean meal. In maize, reducing the grid size decreased the pc digestibility of all AA numerically, but not significantly (P > 0.050). The mean numerical differences in pc AA digestibility between the grid sizes were 0.045 and 0.055 in maize and soybean meal, respectively. 4. Future studies investigating the pc AA digestibility should specify the particle size distribution and should investigate the test ingredients ground similarly for practical applications.

  1. An FPGA Architecture for Extracting Real-Time Zernike Coefficients from Measured Phase Gradients

    NASA Astrophysics Data System (ADS)

    Moser, Steven; Lee, Peter; Podoleanu, Adrian

    2015-04-01

    Zernike modes are commonly used in adaptive optics systems to represent optical wavefronts. However, real-time calculation of Zernike modes is time consuming due to two factors: the large factorial components in the radial polynomials used to define them and the large inverse matrix calculation needed for the linear fit. This paper presents an efficient parallel method for calculating Zernike coefficients from phase gradients produced by a Shack-Hartman sensor and its real-time implementation using an FPGA by pre-calculation and storage of subsections of the large inverse matrix. The architecture exploits symmetries within the Zernike modes to achieve a significant reduction in memory requirements and a speed-up of 2.9 when compared to published results utilising a 2D-FFT method for a grid size of 8×8. Analysis of processor element internal word length requirements show that 24-bit precision in precalculated values of the Zernike mode partial derivatives ensures less than 0.5% error per Zernike coefficient and an overall error of <1%. The design has been synthesized on a Xilinx Spartan-6 XC6SLX45 FPGA. The resource utilisation on this device is <3% of slice registers, <15% of slice LUTs, and approximately 48% of available DSP blocks independent of the Shack-Hartmann grid size. Block RAM usage is <16% for Shack-Hartmann grid sizes up to 32×32.

  2. Massive parallel 3D PIC simulation of negative ion extraction

    NASA Astrophysics Data System (ADS)

    Revel, Adrien; Mochalskyy, Serhiy; Montellano, Ivar Mauricio; Wünderlich, Dirk; Fantz, Ursel; Minea, Tiberiu

    2017-09-01

    The 3D PIC-MCC code ONIX is dedicated to modeling Negative hydrogen/deuterium Ion (NI) extraction and co-extraction of electrons from radio-frequency driven, low pressure plasma sources. It provides valuable insight on the complex phenomena involved in the extraction process. In previous calculations, a mesh size larger than the Debye length was used, implying numerical electron heating. Important steps have been achieved in terms of computation performance and parallelization efficiency allowing successful massive parallel calculations (4096 cores), imperative to resolve the Debye length. In addition, the numerical algorithms have been improved in terms of grid treatment, i.e., the electric field near the complex geometry boundaries (plasma grid) is calculated more accurately. The revised model preserves the full 3D treatment, but can take advantage of a highly refined mesh. ONIX was used to investigate the role of the mesh size, the re-injection scheme for lost particles (extracted or wall absorbed), and the electron thermalization process on the calculated extracted current and plasma characteristics. It is demonstrated that all numerical schemes give the same NI current distribution for extracted ions. Concerning the electrons, the pair-injection technique is found well-adapted to simulate the sheath in front of the plasma grid.

  3. Catching ghosts with a coarse net: use and abuse of spatial sampling data in detecting synchronization

    PubMed Central

    2017-01-01

    Synchronization of population dynamics in different habitats is a frequently observed phenomenon. A common mathematical tool to reveal synchronization is the (cross)correlation coefficient between time courses of values of the population size of a given species where the population size is evaluated from spatial sampling data. The corresponding sampling net or grid is often coarse, i.e. it does not resolve all details of the spatial configuration, and the evaluation error—i.e. the difference between the true value of the population size and its estimated value—can be considerable. We show that this estimation error can make the value of the correlation coefficient very inaccurate or even irrelevant. We consider several population models to show that the value of the correlation coefficient calculated on a coarse sampling grid rarely exceeds 0.5, even if the true value is close to 1, so that the synchronization is effectively lost. We also observe ‘ghost synchronization’ when the correlation coefficient calculated on a coarse sampling grid is close to 1 but in reality the dynamics are not correlated. Finally, we suggest a simple test to check the sampling grid coarseness and hence to distinguish between the true and artifactual values of the correlation coefficient. PMID:28202589

  4. SU-F-T-365: Clinical Commissioning of the Monaco Treatment Planning System for the Novalis Tx to Deliver VMAT, SRS and SBRT Treatments

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Adnani, N

    Purpose: To commission the Monaco Treatment Planning System for the Novalis Tx machine. Methods: The commissioning of Monte-Carlo (MC), Collapsed Cone (CC) and electron Monte-Carlo (eMC) beam models was performed through a series of measurements and calculations in medium and in water. In medium measurements relied Octavius 4D QA system with the 1000 SRS detector array for field sizes less than 4 cm × 4 cm and the 1500 detector array for larger field sizes. Heterogeneity corrections were validated using a custom built phantom. Prior to clinical implementation, an end to end testing of a Prostate and H&N VMAT plansmore » was performed. Results: Using a 0.5% uncertainty and 2 mm grid sizes, Tables I and II summarize the MC validation at 6 MV and 18 MV in both medium and water. Tables III and IV show similar comparisons for CC. Using the custom heterogeneity phantom setup of Figure 1 and IGRT guidance summarized in Figure 2, Table V lists the percent pass rate for a 2%, 2 mm gamma criteria at 6 and 18 MV for both MC and CC. The relationship between MC calculations settings of uncertainty and grid size and the gamma passing rate for a prostate and H&N case is shown in Table VI. Table VII lists the results of the eMC calculations compared to measured data for clinically available applicators and Table VIII for small field cutouts. Conclusion: MU calculations using MC are highly sensitive to uncertainty and grid size settings. The difference can be of the order of several per cents. MC is superior to CC for small fields and when using heterogeneity corrections, regardless of field size, making it more suitable for SRS, SBRT and VMAT deliveries. eMC showed good agreement with measurements down to 2 cm − 2 cm field size.« less

  5. Divide-and-conquer density functional theory on hierarchical real-space grids: Parallel implementation and applications

    NASA Astrophysics Data System (ADS)

    Shimojo, Fuyuki; Kalia, Rajiv K.; Nakano, Aiichiro; Vashishta, Priya

    2008-02-01

    A linear-scaling algorithm based on a divide-and-conquer (DC) scheme has been designed to perform large-scale molecular-dynamics (MD) simulations, in which interatomic forces are computed quantum mechanically in the framework of the density functional theory (DFT). Electronic wave functions are represented on a real-space grid, which is augmented with a coarse multigrid to accelerate the convergence of iterative solutions and with adaptive fine grids around atoms to accurately calculate ionic pseudopotentials. Spatial decomposition is employed to implement the hierarchical-grid DC-DFT algorithm on massively parallel computers. The largest benchmark tests include 11.8×106 -atom ( 1.04×1012 electronic degrees of freedom) calculation on 131 072 IBM BlueGene/L processors. The DC-DFT algorithm has well-defined parameters to control the data locality, with which the solutions converge rapidly. Also, the total energy is well conserved during the MD simulation. We perform first-principles MD simulations based on the DC-DFT algorithm, in which large system sizes bring in excellent agreement with x-ray scattering measurements for the pair-distribution function of liquid Rb and allow the description of low-frequency vibrational modes of graphene. The band gap of a CdSe nanorod calculated by the DC-DFT algorithm agrees well with the available conventional DFT results. With the DC-DFT algorithm, the band gap is calculated for larger system sizes until the result reaches the asymptotic value.

  6. Ostracod Body Size Change Across Space and Time

    NASA Astrophysics Data System (ADS)

    Nolen, L.; Llarena, L. A.; Saux, J.; Heim, N. A.; Payne, J.

    2014-12-01

    Many factors drive evolution, although it is not always clear which factors are more influential. Miller et al. (2009) found that there is a change in geographic disparity in diversity in marine biotas over time. We tested if there was also geographic disparity in body size during different epochs. We used marine ostracods, which are tiny crustaceans, as a study group for this analysis. We also studied which factor is more influential in body size change: distance or time. We compared the mean body size from different geologic time intervals as well as the mean body size from different locations for each epoch. We grouped ostracod occurrences from the Paleobiology Database into 10º x 10º grid cells on a paleogeographic map. Then we calculated the difference in mean size and the distance between the grid cells containing specimens. Our size data came from the Ellis & Messina"Catalogue of Ostracod" as well as the"Treatise on Invertebrate Paleontology". Sizes were calculated by applying the formula for the volume of an ellipsoid to three linear dimensions of the ostracod carapace (anteroposterior, dorsoventral, and right-left lengths). Throughout this analysis we have come to the realization that there is a trend in ostracods towards smaller size over time. Therefore there is also a trend through time of decreasing difference in size between occurrences in different grid cells. However, if time is not taken into account, there is no correlation between size and geographic distance. This may be attributed to the fact that one might not expect a big size difference between locations that are far apart but still at a similar latitude (for example, at the equator). This analysis suggests that distance alone is not the main factor in driving changes in ostracod size over time.

  7. Simulating ground water-lake interactions: Approaches and insights

    USGS Publications Warehouse

    Hunt, R.J.; Haitjema, H.M.; Krohelski, J.T.; Feinstein, D.T.

    2003-01-01

    Approaches for modeling lake-ground water interactions have evolved significantly from early simulations that used fixed lake stages specified as constant head to sophisticated LAK packages for MODFLOW. Although model input can be complex, the LAK package capabilities and output are superior to methods that rely on a fixed lake stage and compare well to other simple methods where lake stage can be calculated. Regardless of the approach, guidelines presented here for model grid size, location of three-dimensional flow, and extent of vertical capture can facilitate the construction of appropriately detailed models that simulate important lake-ground water interactions without adding unnecessary complexity. In addition to MODFLOW approaches, lake simulation has been formulated in terms of analytic elements. The analytic element lake package had acceptable agreement with a published LAK1 problem, even though there were differences in the total lake conductance and number of layers used in the two models. The grid size used in the original LAK1 problem, however, violated a grid size guideline presented in this paper. Grid sensitivity analyses demonstrated that an appreciable discrepancy in the distribution of stream and lake flux was related to the large grid size used in the original LAK1 problem. This artifact is expected regardless of MODFLOW LAK package used. When the grid size was reduced, a finite-difference formulation approached the analytic element results. These insights and guidelines can help ensure that the proper lake simulation tool is being selected and applied.

  8. Optimal system sizing in grid-connected photovoltaic applications

    NASA Astrophysics Data System (ADS)

    Simoens, H. M.; Baert, D. H.; de Mey, G.

    A costs/benefits analysis for optimizing the combination of photovoltaic (PV) panels, batteries and an inverter for grid interconnected systems at a 500 W/day Belgian residence is presented. It is assumed that some power purchases from the grid will always be necessary, and that excess PV power can be fed into the grid. A minimal value for the cost divided by the performance is defined for economic optimization. Shortages and excesses are calculated for PV panels of 0.5-10 kWp output, with consideration given to the advantages of a battery back-up. The minimal economic value is found to increase with the magnitude of PV output, and an inverter should never be rated at more than half the array maximum output. A maximum panel size for the Belgian residence is projected to be 6 kWp.

  9. Influence of model grid size on the simulation of PM2.5 and the related excess mortality in Japan

    NASA Astrophysics Data System (ADS)

    Goto, D.; Ueda, K.; Ng, C. F.; Takami, A.; Ariga, T.; Matsuhashi, K.; Nakajima, T.

    2016-12-01

    Aerosols, especially PM2.5, can affect air pollution, climate change, and human health. The estimation of health impacts due to PM2.5 is often performed using global and regional aerosol transport models with various horizontal resolutions. To investigate the dependence of the simulated PM2.5 on model grid sizes, we executed two simulations using a high-resolution model ( 10km; HRM) and a low-resolution model ( 100km; LRM, which is a typical value for general circulation models). In this study, we used a global-to-regional atmospheric transport model to simulate PM2.5 in Japan with a stretched grid system in HRM and a uniform grid system in LRM for the present (the 2000) and the future (the 2030, as proposed by the Representative Concentrations Pathway 4.5, RCP4.5). These calculations were performed by nudging meteorological fields obtained from an atmosphere-ocean coupled model and providing emission inventories used in the coupled model. After correcting for bias, we calculated the excess mortality due to long-term exposure to PM2.5 for the elderly. Results showed the LRM underestimated by approximately 30 % (of PM2.5 concentrations in the 2000 and 2030), approximately 60 % (excess mortality in the 2000) and approximately 90 % (excess mortality in 2030) compared to the HRM results. The estimation of excess mortality therefore performed better with high-resolution grid sizes. In addition, we also found that our nesting method could be a useful tool to obtain better estimation results.

  10. Adaptive Correlation Space Adjusted Open-Loop Tracking Approach for Vehicle Positioning with Global Navigation Satellite System in Urban Areas

    PubMed Central

    Ruan, Hang; Li, Jian; Zhang, Lei; Long, Teng

    2015-01-01

    For vehicle positioning with Global Navigation Satellite System (GNSS) in urban areas, open-loop tracking shows better performance because of its high sensitivity and superior robustness against multipath. However, no previous study has focused on the effects of the code search grid size on the code phase measurement accuracy of open-loop tracking. Traditional open-loop tracking methods are performed by the batch correlators with fixed correlation space. The code search grid size, which is the correlation space, is a constant empirical value and the code phase measuring accuracy will be largely degraded due to the improper grid size, especially when the signal carrier-to-noise density ratio (C/N0) varies. In this study, the Adaptive Correlation Space Adjusted Open-Loop Tracking Approach (ACSA-OLTA) is proposed to improve the code phase measurement dependent pseudo range accuracy. In ACSA-OLTA, the correlation space is adjusted according to the signal C/N0. The novel Equivalent Weighted Pseudo Range Error (EWPRE) is raised to obtain the optimal code search grid sizes for different C/N0. The code phase measuring errors of different measurement calculation methods are analyzed for the first time. The measurement calculation strategy of ACSA-OLTA is derived from the analysis to further improve the accuracy but reduce the correlator consumption. Performance simulation and real tests confirm that the pseudo range and positioning accuracy of ASCA-OLTA are better than the traditional open-loop tracking methods in the usual scenarios of urban area. PMID:26343683

  11. Effect of grid transparency and finite collector size on determining ion temperature and density by the retarding potential analyzer

    NASA Technical Reports Server (NTRS)

    Troy, B. E., Jr.; Maier, E. J.

    1975-01-01

    The effects of the grid transparency and finite collector size on the values of thermal ion density and temperature determined by the standard RPA (retarding potential analyzer) analysis method are investigated. The current-voltage curves calculated for varying RPA parameters and a given ion mass, temperature, and density are analyzed by the standard RPA method. It is found that only small errors in temperature and density are introduced for an RPA with typical dimensions, and that even when the density error is substantial for nontypical dimensions, the temperature error remains minimum.

  12. Fast Computation of Ground Motion Shaking Map base on the Modified Stochastic Finite Fault Modeling

    NASA Astrophysics Data System (ADS)

    Shen, W.; Zhong, Q.; Shi, B.

    2012-12-01

    Rapidly regional MMI mapping soon after a moderate-large earthquake is crucial to loss estimation, emergency services and planning of emergency action by the government. In fact, many countries show different degrees of attention on the technology of rapid estimation of MMI , and this technology has made significant progress in earthquake-prone countries. In recent years, numerical modeling of strong ground motion has been well developed with the advances of computation technology and earthquake science. The computational simulation of strong ground motion caused by earthquake faulting has become an efficient way to estimate the regional MMI distribution soon after earthquake. In China, due to the lack of strong motion observation in network sparse or even completely missing areas, the development of strong ground motion simulation method has become an important means of quantitative estimation of strong motion intensity. In many of the simulation models, stochastic finite fault model is preferred to rapid MMI estimating for its time-effectiveness and accuracy. In finite fault model, a large fault is divided into N subfaults, and each subfault is considered as a small point source. The ground motions contributed by each subfault are calculated by the stochastic point source method which is developed by Boore, and then summed at the observation point to obtain the ground motion from the entire fault with a proper time delay. Further, Motazedian and Atkinson proposed the concept of Dynamic Corner Frequency, with the new approach, the total radiated energy from the fault and the total seismic moment are conserved independent of subfault size over a wide range of subfault sizes. In current study, the program EXSIM developed by Motazedian and Atkinson has been modified for local or regional computations of strong motion parameters such as PGA, PGV and PGD, which are essential for MMI estimating. To make the results more reasonable, we consider the impact of V30 for the ground shaking intensity, and the results of the comparisons between the simulated and observed MMI for the 2004 Mw 6.0 Parkfield earthquake, the 2008 Mw 7.9Wenchuan earthquake and the 1976 Mw 7.6Tangshan earthquake is fairly well. Take Parkfield earthquake as example, the simulative result reflect the directivity effect and the influence of the shallow velocity structure well. On the other hand, the simulative data is in good agreement with the network data and NGA (Next Generation Attenuation). The consumed time depends on the number of the subfaults and the number of the grid point. For the 2004 Mw 6.0 Parkfield earthquake, the grid size we calculated is 2.5° × 2.5°, the grid space is 0.025°, and the total time consumed is about 1.3hours. For the 2008 Mw 7.9 Wenchuan earthquake, the grid size calculated is 10° × 10°, the grid space is 0.05°, the total number of grid point is more than 40,000, and the total time consumed is about 7.5 hours. For t the 1976 Mw 7.6 Tangshan earthquake, the grid size we calculated is 4° × 6°, the grid space is 0.05°, and the total time consumed is about 2.1 hours. The CPU we used is 3.40GHz, and such computational time could further reduce by using GPU computing technique and other parallel computing technique. This is also our next focus.

  13. A stand-alone tidal prediction application for mobile devices

    NASA Astrophysics Data System (ADS)

    Tsai, Cheng-Han; Fan, Ren-Ye; Yang, Yi-Chung

    2017-04-01

    It is essential for people conducting fishing, leisure, or research activities at the coasts to have timely and handy tidal information. Although tidal information can be found easily on the internet or using mobile device applications, this information is all applicable for only certain specific locations, not anywhere on the coast, and they need an internet connection. We have developed an application for Android devices, which allows the user to obtain hourly tidal height anywhere on the coast for the next 24 hours without having to have any internet connection. All the necessary information needed for the tidal height calculation is stored in the application. To develop this application, we first simulate tides in the Taiwan Sea using the hydrodynamic model (MIKE21 HD) developed by the DHI. The simulation domain covers the whole coast of Taiwan and the surrounding seas with a grid size of 1 km by 1 km. This grid size allows us to calculate tides with high spatial resolution. The boundary conditions for the simulation domain were obtained from the Tidal Model Driver of the Oregon State University, using its tidal constants of eight constituents: M2, S2, N2, K2, K1, O1, P1, and Q1. The simulation calculates tides for 183 days so that the tidal constants for the above eight constituents of each water grid can be extracted by harmonic analysis. Using the calculated tidal constants, we can predict the tides in each grid of our simulation domain, which is useful when one needs the tidal information for any location in the Taiwan Sea. However, for the mobile application, we only store the eight tidal constants for the water grids on the coast. Once the user activates the application, it reads the longitude and latitude from the GPS sensor in the mobile device and finds the nearest coastal grid which has our tidal constants. Then, the application calculates tidal height variation based on the harmonic analysis. The application also allows the user to input location and time to obtain tides for any historic or future dates for the input location. The predicted tides have been verified with the historic tidal records of certain tidal stations. The verification shows that the tides predicted by the application match the measured record well.

  14. Enhancement of a 2D front-tracking algorithm with a non-uniform distribution of Lagrangian markers

    NASA Astrophysics Data System (ADS)

    Febres, Mijail; Legendre, Dominique

    2018-04-01

    The 2D front tracking method is enhanced to control the development of spurious velocities for non-uniform distributions of markers. The hybrid formulation of Shin et al. (2005) [7] is considered. A new tangent calculation is proposed for the calculation of the tension force at markers. A new reconstruction method is also proposed to manage non-uniform distributions of markers. We show that for both the static and the translating spherical drop test case the spurious currents are reduced to the machine precision. We also show that the ratio of the Lagrangian grid size Δs over the Eulerian grid size Δx has to satisfy Δs / Δx > 0.2 for ensuring such low level of spurious velocity. The method is found to provide very good agreement with benchmark test cases from the literature.

  15. Research on Grid Size Suitability of Gridded Population Distribution in Urban Area: A Case Study in Urban Area of Xuanzhou District, China.

    PubMed

    Dong, Nan; Yang, Xiaohuan; Cai, Hongyan; Xu, Fengjiao

    2017-01-01

    The research on the grid size suitability is important to provide improvement in accuracies of gridded population distribution. It contributes to reveal the actual spatial distribution of population. However, currently little research has been done in this area. Many well-modeled gridded population dataset are basically built at a single grid scale. If the grid cell size is not appropriate, it will result in spatial information loss or data redundancy. Therefore, in order to capture the desired spatial variation of population within the area of interest, it is necessary to conduct research on grid size suitability. This study summarized three expressed levels to analyze grid size suitability, which include location expressed level, numeric information expressed level, and spatial relationship expressed level. This study elaborated the reasons for choosing the five indexes to explore expression suitability. These five indexes are consistency measure, shape index rate, standard deviation of population density, patches diversity index, and the average local variance. The suitable grid size was determined by constructing grid size-indicator value curves and suitable grid size scheme. Results revealed that the three expressed levels on 10m grid scale are satisfying. And the population distribution raster data with 10m grid size provide excellent accuracy without loss. The 10m grid size is recommended as the appropriate scale for generating a high-quality gridded population distribution in our study area. Based on this preliminary study, it indicates the five indexes are coordinated with each other and reasonable and effective to assess grid size suitability. We also suggest choosing these five indexes in three perspectives of expressed level to carry out the research on grid size suitability of gridded population distribution.

  16. Research on Grid Size Suitability of Gridded Population Distribution in Urban Area: A Case Study in Urban Area of Xuanzhou District, China

    PubMed Central

    Dong, Nan; Yang, Xiaohuan; Cai, Hongyan; Xu, Fengjiao

    2017-01-01

    The research on the grid size suitability is important to provide improvement in accuracies of gridded population distribution. It contributes to reveal the actual spatial distribution of population. However, currently little research has been done in this area. Many well-modeled gridded population dataset are basically built at a single grid scale. If the grid cell size is not appropriate, it will result in spatial information loss or data redundancy. Therefore, in order to capture the desired spatial variation of population within the area of interest, it is necessary to conduct research on grid size suitability. This study summarized three expressed levels to analyze grid size suitability, which include location expressed level, numeric information expressed level, and spatial relationship expressed level. This study elaborated the reasons for choosing the five indexes to explore expression suitability. These five indexes are consistency measure, shape index rate, standard deviation of population density, patches diversity index, and the average local variance. The suitable grid size was determined by constructing grid size-indicator value curves and suitable grid size scheme. Results revealed that the three expressed levels on 10m grid scale are satisfying. And the population distribution raster data with 10m grid size provide excellent accuracy without loss. The 10m grid size is recommended as the appropriate scale for generating a high-quality gridded population distribution in our study area. Based on this preliminary study, it indicates the five indexes are coordinated with each other and reasonable and effective to assess grid size suitability. We also suggest choosing these five indexes in three perspectives of expressed level to carry out the research on grid size suitability of gridded population distribution. PMID:28122050

  17. The effect of grid transparency and finite collector size on determining ion temperature and density by the retarding potential analyzer

    NASA Technical Reports Server (NTRS)

    Troy, B. E., Jr.; Maier, E. J.

    1973-01-01

    The analysis of ion data from retarding potential analyzers (RPA's) is generally done under the planar approximation, which assumes that the grid transparency is constant with angle of incidence and that all ions reaching the plane of the collectors are collected. These approximations are not valid for situations in which the ion thermal velocity is comparable to the vehicle velocity, causing ions to enter the RPA with high average transverse velocity. To investigate these effects, the current-voltage curves for H+ at 4000 K were calculated, taking into account the finite collector size and the variation of grid transparency with angle. These curves are then analyzed under the planar approximation. The results show that only small errors in temperature and density are introduced for an RPA with typical dimensions; and that even when the density error is substantial for non-typical dimensions, the temperature error remains minimal.

  18. Grid-Based Surface Generalized Born Model for Calculation of Electrostatic Binding Free Energies.

    PubMed

    Forouzesh, Negin; Izadi, Saeed; Onufriev, Alexey V

    2017-10-23

    Fast and accurate calculation of solvation free energies is central to many applications, such as rational drug design. In this study, we present a grid-based molecular surface implementation of "R6" flavor of the generalized Born (GB) implicit solvent model, named GBNSR6. The speed, accuracy relative to numerical Poisson-Boltzmann treatment, and sensitivity to grid surface parameters are tested on a set of 15 small protein-ligand complexes and a set of biomolecules in the range of 268 to 25099 atoms. Our results demonstrate that the proposed model provides a relatively successful compromise between the speed and accuracy of computing polar components of the solvation free energies (ΔG pol ) and binding free energies (ΔΔG pol ). The model tolerates a relatively coarse grid size h = 0.5 Å, where the grid artifact error in computing ΔΔG pol remains in the range of k B T ∼ 0.6 kcal/mol. The estimated ΔΔG pol s are well correlated (r 2 = 0.97) with the numerical Poisson-Boltzmann reference, while showing virtually no systematic bias and RMSE = 1.43 kcal/mol. The grid-based GBNSR6 model is available in Amber (AmberTools) package of molecular simulation programs.

  19. Water equivalent thickness of immobilization devices in proton therapy planning - Modelling at treatment planning and validation by measurements with a multi-layer ionization chamber.

    PubMed

    Fellin, Francesco; Righetto, Roberto; Fava, Giovanni; Trevisan, Diego; Amelio, Dante; Farace, Paolo

    2017-03-01

    To investigate the range errors made in treatment planning due to the presence of the immobilization devices along the proton beam path. The measured water equivalent thickness (WET) of selected devices was measured by a high-energy spot and a multi-layer ionization chamber and compared with that predicted by treatment planning system (TPS). Two treatment couches, two thermoplastic masks (both un-stretched and stretched) and one headrest were selected. At TPS, every immobilization device was modelled as being part of the patient. The following parameters were assessed: CT acquisition protocol, dose-calculation grid-sizes (1.5 and 3.0mm) and beam-entrance with respect to the devices (coplanar and non-coplanar). Finally, the potential errors produced by a wrong manual separation between treatment couch and the CT table (not present during treatment) were investigated. In the thermoplastic mask, there was a clear effect due to beam entrance, a moderate effect due to the CT protocols and almost no effect due to TPS grid-size, with 1mm errors observed only when thick un-stretched portions were crossed by non-coplanar beams. In the treatment couches the WET errors were negligible (<0.3mm) regardless of the grid-size and CT protocol. The potential range errors produced in the manual separation between treatment couch and CT table were small with 1.5mm grid-size, but could be >0.5mm with a 3.0mm grid-size. In the headrest, WET errors were negligible (0.2mm). With only one exception (un-stretched mask, non-coplanar beams), the WET of all the immobilization devices was properly modelled by the TPS. Copyright © 2017 Associazione Italiana di Fisica Medica. Published by Elsevier Ltd. All rights reserved.

  20. The Effect of DEM Source and Grid Size on the Index of Connectivity in Savanna Catchments

    NASA Astrophysics Data System (ADS)

    Jarihani, Ben; Sidle, Roy; Bartley, Rebecca; Roth, Christian

    2017-04-01

    The term "hydrological connectivity" is increasingly used instead of sediment delivery ratio to describe the linkage between the sources of water and sediment within a catchment to the catchment outlet. Sediment delivery ratio is an empirical parameter that is highly site-specific and tends to lump all processes, whilst hydrological connectivity focuses on the spatially-explicit hydrologic drivers of surficial processes. Detailed topographic information plays a fundamental role in geomorphological interpretations as well as quantitative modelling of sediment fluxes and connectivity. Geomorphometric analysis permits a detailed characterization of drainage area and drainage pattern together with the possibility of characterizing surface roughness. High resolution topographic data (i.e., LiDAR) are not available for all areas; however, remotely sensed topographic data from multiple sources with different grid sizes are used to undertake geomorphologic analysis in data-sparse regions. The Index of Connectivity (IC), a geomorphometric model based only on DEM data, is applied in two small savanna catchments in Queensland, Australia. The influence of the scale of the topographic data is explored by using DEMs from LiDAR ( 1 m), WorldDEM ( 10 m), raw SRTM and hydrologically corrected SRTM derived data ( 30 m) to calculate the index of connectivity. The effect of the grid size is also investigated by resampling the high resolution LiDAR DEM to multiple grid sizes (e.g. 5, 10, 20 m) and comparing the extracted IC.

  1. Creation of parallel algorithms for the solution of problems of gas dynamics on multi-core computers and GPU

    NASA Astrophysics Data System (ADS)

    Rybakin, B.; Bogatencov, P.; Secrieru, G.; Iliuha, N.

    2013-10-01

    The paper deals with a parallel algorithm for calculations on multiprocessor computers and GPU accelerators. The calculations of shock waves interaction with low-density bubble results and the problem of the gas flow with the forces of gravity are presented. This algorithm combines a possibility to capture a high resolution of shock waves, the second-order accuracy for TVD schemes, and a possibility to observe a low-level diffusion of the advection scheme. Many complex problems of continuum mechanics are numerically solved on structured or unstructured grids. To improve the accuracy of the calculations is necessary to choose a sufficiently small grid (with a small cell size). This leads to the drawback of a substantial increase of computation time. Therefore, for the calculations of complex problems it is reasonable to use the method of Adaptive Mesh Refinement. That is, the grid refinement is performed only in the areas of interest of the structure, where, e.g., the shock waves are generated, or a complex geometry or other such features exist. Thus, the computing time is greatly reduced. In addition, the execution of the application on the resulting sequence of nested, decreasing nets can be parallelized. Proposed algorithm is based on the AMR method. Utilization of AMR method can significantly improve the resolution of the difference grid in areas of high interest, and from other side to accelerate the processes of the multi-dimensional problems calculating. Parallel algorithms of the analyzed difference models realized for the purpose of calculations on graphic processors using the CUDA technology [1].

  2. Reliability analysis in interdependent smart grid systems

    NASA Astrophysics Data System (ADS)

    Peng, Hao; Kan, Zhe; Zhao, Dandan; Han, Jianmin; Lu, Jianfeng; Hu, Zhaolong

    2018-06-01

    Complex network theory is a useful way to study many real complex systems. In this paper, a reliability analysis model based on complex network theory is introduced in interdependent smart grid systems. In this paper, we focus on understanding the structure of smart grid systems and studying the underlying network model, their interactions, and relationships and how cascading failures occur in the interdependent smart grid systems. We propose a practical model for interdependent smart grid systems using complex theory. Besides, based on percolation theory, we also study the effect of cascading failures effect and reveal detailed mathematical analysis of failure propagation in such systems. We analyze the reliability of our proposed model caused by random attacks or failures by calculating the size of giant functioning components in interdependent smart grid systems. Our simulation results also show that there exists a threshold for the proportion of faulty nodes, beyond which the smart grid systems collapse. Also we determine the critical values for different system parameters. In this way, the reliability analysis model based on complex network theory can be effectively utilized for anti-attack and protection purposes in interdependent smart grid systems.

  3. Sizing and modelling of photovoltaic water pumping system

    NASA Astrophysics Data System (ADS)

    Al-Badi, A.; Yousef, H.; Al Mahmoudi, T.; Al-Shammaki, M.; Al-Abri, A.; Al-Hinai, A.

    2018-05-01

    With the decline in price of the photovoltaics (PVs) their use as a power source for water pumping is the most attractive solution instead of using diesel generators or electric motors driven by a grid system. In this paper, a method to design a PV pumping system is presented and discussed, which is then used to calculate the required size of the PV for an existing farm. Furthermore, the amount of carbon dioxide emissions saved by the use of PV water pumping system instead of using diesel-fuelled generators or electrical motor connected to the grid network is calculated. In addition, an experimental set-up is developed for the PV water pumping system using both DC and AC motors with batteries. The experimental tests are used to validate the developed MATLAB model. This research work demonstrates that using the PV water pumping system is not only improving the living conditions in rural areas but it is also protecting the environment and can be a cost-effective application in remote locations.

  4. Improvements in sub-grid, microphysics averages using quadrature based approaches

    NASA Astrophysics Data System (ADS)

    Chowdhary, K.; Debusschere, B.; Larson, V. E.

    2013-12-01

    Sub-grid variability in microphysical processes plays a critical role in atmospheric climate models. In order to account for this sub-grid variability, Larson and Schanen (2013) propose placing a probability density function on the sub-grid cloud microphysics quantities, e.g. autoconversion rate, essentially interpreting the cloud microphysics quantities as a random variable in each grid box. Random sampling techniques, e.g. Monte Carlo and Latin Hypercube, can be used to calculate statistics, e.g. averages, on the microphysics quantities, which then feed back into the model dynamics on the coarse scale. We propose an alternate approach using numerical quadrature methods based on deterministic sampling points to compute the statistical moments of microphysics quantities in each grid box. We have performed a preliminary test on the Kessler autoconversion formula, and, upon comparison with Latin Hypercube sampling, our approach shows an increased level of accuracy with a reduction in sample size by almost two orders of magnitude. Application to other microphysics processes is the subject of ongoing research.

  5. ELIPGRID-PC: A PC program for calculating hot spot probabilities

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Davidson, J.R.

    1994-10-01

    ELIPGRID-PC, a new personal computer program has been developed to provide easy access to Singer`s 1972 ELIPGRID algorithm for hot-spot detection probabilities. Three features of the program are the ability to determine: (1) the grid size required for specified conditions, (2) the smallest hot spot that can be sampled with a given probability, and (3) the approximate grid size resulting from specified conditions and sampling cost. ELIPGRID-PC also provides probability of hit versus cost data for graphing with spread-sheets or graphics software. The program has been successfully tested using Singer`s published ELIPGRID results. An apparent error in the original ELIPGRIDmore » code has been uncovered and an appropriate modification incorporated into the new program.« less

  6. 3D magnetospheric parallel hybrid multi-grid method applied to planet–plasma interactions

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Leclercq, L., E-mail: ludivine.leclercq@latmos.ipsl.fr; Modolo, R., E-mail: ronan.modolo@latmos.ipsl.fr; Leblanc, F.

    2016-03-15

    We present a new method to exploit multiple refinement levels within a 3D parallel hybrid model, developed to study planet–plasma interactions. This model is based on the hybrid formalism: ions are kinetically treated whereas electrons are considered as a inertia-less fluid. Generally, ions are represented by numerical particles whose size equals the volume of the cells. Particles that leave a coarse grid subsequently entering a refined region are split into particles whose volume corresponds to the volume of the refined cells. The number of refined particles created from a coarse particle depends on the grid refinement rate. In order tomore » conserve velocity distribution functions and to avoid calculations of average velocities, particles are not coalesced. Moreover, to ensure the constancy of particles' shape function sizes, the hybrid method is adapted to allow refined particles to move within a coarse region. Another innovation of this approach is the method developed to compute grid moments at interfaces between two refinement levels. Indeed, the hybrid method is adapted to accurately account for the special grid structure at the interfaces, avoiding any overlapping grid considerations. Some fundamental test runs were performed to validate our approach (e.g. quiet plasma flow, Alfven wave propagation). Lastly, we also show a planetary application of the model, simulating the interaction between Jupiter's moon Ganymede and the Jovian plasma.« less

  7. Impact of dose size in single fraction spatially fractionated (grid) radiotherapy for melanoma

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhang, Hualin, E-mail: hualin.zhang@northwestern.edu, E-mail: hualinzhang@yahoo.com; Zhong, Hualiang; Barth, Rolf F.

    2014-02-15

    Purpose: To evaluate the impact of dose size in single fraction, spatially fractionated (grid) radiotherapy for selectively killing infiltrated melanoma cancer cells of different tumor sizes, using different radiobiological models. Methods: A Monte Carlo technique was employed to calculate the 3D dose distribution of a commercially available megavoltage grid collimator in a 6 MV beam. The linear-quadratic (LQ) and modified linear quadratic (MLQ) models were used separately to evaluate the therapeutic outcome of a series of single fraction regimens that employed grid therapy to treat both acute and late responding melanomas of varying sizes. The dose prescription point was atmore » the center of the tumor volume. Dose sizes ranging from 1 to 30 Gy at 100% dose line were modeled. Tumors were either touching the skin surface or having their centers at a depth of 3 cm. The equivalent uniform dose (EUD) to the melanoma cells and the therapeutic ratio (TR) were defined by comparing grid therapy with the traditional open debulking field. The clinical outcomes from recent reports were used to verify the authors’ model. Results: Dose profiles at different depths and 3D dose distributions in a series of 3D melanomas treated with grid therapy were obtained. The EUDs and TRs for all sizes of 3D tumors involved at different doses were derived through the LQ and MLQ models, and a practical equation was derived. The EUD was only one fifth of the prescribed dose. The TR was dependent on the prescribed dose and on the LQ parameters of both the interspersed cancer and normal tissue cells. The results from the LQ model were consistent with those of the MLQ model. At 20 Gy, the EUD and TR by the LQ model were 2.8% higher and 1% lower than by the MLQ, while at 10 Gy, the EUD and TR as defined by the LQ model were only 1.4% higher and 0.8% lower, respectively. The dose volume histograms of grid therapy for a 10 cm tumor showed different dosimetric characteristics from those of conventional radiotherapy. A significant portion of the tumor volume received a very large dose in grid therapy, which ensures significant tumor cell killing in these regions. Conversely, some areas received a relatively small dose, thereby sparing interspersed normal cells and increasing radiation tolerance. The radiobiology modeling results indicated that grid therapy could be useful for treating acutely responding melanomas infiltrating radiosensitive normal tissues. The theoretical model predictions were supported by the clinical outcomes. Conclusions: Grid therapy functions by selectively killing infiltrating tumor cells and concomitantly sparing interspersed normal cells. The TR depends on the radiosensitivity of the cell population, dose, tumor size, and location. Because the volumes of very high dose regions are small, the LQ model can be used safely to predict the clinical outcomes of grid therapy. When treating melanomas with a dose of 15 Gy or higher, single fraction grid therapy is clearly advantageous for sparing interspersed normal cells. The existence of a threshold fraction dose, which was found in the authors’ theoretical simulations, was confirmed by clinical observations.« less

  8. Optimal Grid Size for Inter-Comparability of MODIS And VIIRS Vegetation Indices at Level 2G or Higher

    NASA Astrophysics Data System (ADS)

    Campagnolo, M.; Schaaf, C.

    2016-12-01

    Due to the necessity of time compositing and other user requirements, vegetation indices, as well as many other EOS derived products, are distributed in a gridded format (level L2G or higher) using an equal area sinusoidal grid, at grid sizes of 232 m, 463 m or 926 m. In this process, the actual surface signal suffers somewhat of a degradation, caused by both the sensor's point spread function and this resampling from swath to the regular grid. The magnitude of that degradation depends on a number of factors, such as surface heterogeneity, band nominal resolution, observation geometry and grid size. In this research, the effect of grid size is quantified for MODIS and VIIRS (at five EOS validation sites with distinct land covers), for the full range of view zenith angles, and at grid sizes of 232 m, 253 m, 309 m, 371 m, 397 m and 463 m. This allows us to compare MODIS and VIIRS gridded products for the same scenes, and to determine the grid size at which these products are most similar. Towards that end, simulated MODIS and VIIRS bands are generated from Landsat 8 surface reflectance images at each site and gridded products are then derived by using maximum obscov resampling. Then, for every grid size, the original Landsat 8 NDVI and the derived MODIS and VIIRS NDVI products are compared. This methodology can be applied to other bands and products, to determine which spatial aggregation overall is best suited for EOS to S-NPP product continuity. Results for MODIS (250 m bands) and VIIRS (375 m bands) NDVI products show that finer grid sizes tend to be better at preserving the original signal. Significant degradation for gridded NDVI occurs when grid size is larger then 253 m (MODIS) and 371 m (VIIRS). Our results suggest that current MODIS "500 m" (actually 463 m) grid size is best for product continuity. Note however, that up to that grid size value, MODIS gridded products are somewhat better at preserving the surface signal than VIIRS, except for at very high VZA.

  9. Target intersection probabilities for parallel-line and continuous-grid types of search

    USGS Publications Warehouse

    McCammon, R.B.

    1977-01-01

    The expressions for calculating the probability of intersection of hidden targets of different sizes and shapes for parallel-line and continuous-grid types of search can be formulated by vsing the concept of conditional probability. When the prior probability of the orientation of a widden target is represented by a uniform distribution, the calculated posterior probabilities are identical with the results obtained by the classic methods of probability. For hidden targets of different sizes and shapes, the following generalizations about the probability of intersection can be made: (1) to a first approximation, the probability of intersection of a hidden target is proportional to the ratio of the greatest dimension of the target (viewed in plane projection) to the minimum line spacing of the search pattern; (2) the shape of the hidden target does not greatly affect the probability of the intersection when the largest dimension of the target is small relative to the minimum spacing of the search pattern, (3) the probability of intersecting a target twice for a particular type of search can be used as a lower bound if there is an element of uncertainty of detection for a particular type of tool; (4) the geometry of the search pattern becomes more critical when the largest dimension of the target equals or exceeds the minimum spacing of the search pattern; (5) for elongate targets, the probability of intersection is greater for parallel-line search than for an equivalent continuous square-grid search when the largest dimension of the target is less than the minimum spacing of the search pattern, whereas the opposite is true when the largest dimension exceeds the minimum spacing; (6) the probability of intersection for nonorthogonal continuous-grid search patterns is not greatly different from the probability of intersection for the equivalent orthogonal continuous-grid pattern when the orientation of the target is unknown. The probability of intersection for an elliptically shaped target can be approximated by treating the ellipse as intermediate between a circle and a line. A search conducted along a continuous rectangular grid can be represented as intermediate between a search along parallel lines and along a continuous square grid. On this basis, an upper and lower bound for the probability of intersection of an elliptically shaped target for a continuous rectangular grid can be calculated. Charts have been constructed that permit the values for these probabilities to be obtained graphically. The use of conditional probability allows the explorationist greater flexibility in considering alternate search strategies for locating hidden targets. ?? 1977 Plenum Publishing Corp.

  10. Clinical study using novel endoscopic system for measuring size of gastrointestinal lesion

    PubMed Central

    Oka, Kiyoshi; Seki, Takeshi; Akatsu, Tomohiro; Wakabayashi, Takao; Inui, Kazuo; Yoshino, Junji

    2014-01-01

    AIM: To verify the performance of a lesion size measurement system through a clinical study. METHODS: Our proposed system, which consists of a conventional endoscope, an optical device, an optical probe, and a personal computer, generates a grid scale to measure the lesion size from an endoscopic image. The width of the grid scale is constantly adjusted according to the distance between the tip of the endoscope and lesion because the lesion size on an endoscopic image changes according to the distance. The shape of the grid scale was corrected to match the distortion of the endoscopic image. The distance was calculated using the amount of laser light reflected from the lesion through an optical probe inserted into the instrument channel of the endoscope. The endoscopist can thus measure the lesion size without contact by comparing the lesion with the size of the grid scale on the endoscopic image. (1) A basic test was performed to verify the relationship between the measurement error eM and the tilt angle of the endoscope; and (2) The sizes of three colon polyps were measured using our system during endoscopy. These sizes were immediately measured by scale after their removal. RESULTS: There was no error at α = 0°. In addition, the values of eM (mean ± SD) were 0.24 ± 0.11 mm (α = 10°), 0.90 ± 0.58 mm (α = 20°) and 2.31 ± 1.41 mm (α = 30°). According to these results, our system has been confirmed to measure accurately when the tilt angle is less than 20°. The measurement error was approximately 1 mm in the clinical study. Therefore, it was concluded that our proposed measurement system was also effective in clinical examinations. CONCLUSION: By combining simple optical equipment with a conventional endoscope, a quick and accurate system for measuring lesion size was established. PMID:24744595

  11. A Numerical Study of Three Moving-Grid Methods for One-Dimensional Partial Differential Equations Which Are Based on the Method of Lines

    NASA Astrophysics Data System (ADS)

    Furzeland, R. M.; Verwer, J. G.; Zegeling, P. A.

    1990-08-01

    In recent years, several sophisticated packages based on the method of lines (MOL) have been developed for the automatic numerical integration of time-dependent problems in partial differential equations (PDEs), notably for problems in one space dimension. These packages greatly benefit from the very successful developments of automatic stiff ordinary differential equation solvers. However, from the PDE point of view, they integrate only in a semiautomatic way in the sense that they automatically adjust the time step sizes, but use just a fixed space grid, chosen a priori, for the entire calculation. For solutions possessing sharp spatial transitions that move, e.g., travelling wave fronts or emerging boundary and interior layers, a grid held fixed for the entire calculation is computationally inefficient, since for a good solution this grid often must contain a very large number of nodes. In such cases methods which attempt automatically to adjust the sizes of both the space and the time steps are likely to be more successful in efficiently resolving critical regions of high spatial and temporal activity. Methods and codes that operate this way belong to the realm of adaptive or moving-grid methods. Following the MOL approach, this paper is devoted to an evaluation and comparison, mainly based on extensive numerical tests, of three moving-grid methods for 1D problems, viz., the finite-element method of Miller and co-workers, the method published by Petzold, and a method based on ideas adopted from Dorfi and Drury. Our examination of these three methods is aimed at assessing which is the most suitable from the point of view of retaining the acknowledged features of reliability, robustness, and efficiency of the conventional MOL approach. Therefore, considerable attention is paid to the temporal performance of the methods.

  12. Application of multi-grid method on the simulation of incremental forging processes

    NASA Astrophysics Data System (ADS)

    Ramadan, Mohamad; Khaled, Mahmoud; Fourment, Lionel

    2016-10-01

    Numerical simulation becomes essential in manufacturing large part by incremental forging processes. It is a splendid tool allowing to show physical phenomena however behind the scenes, an expensive bill should be paid, that is the computational time. That is why many techniques are developed to decrease the computational time of numerical simulation. Multi-Grid method is a numerical procedure that permits to reduce computational time of numerical calculation by performing the resolution of the system of equations on several mesh of decreasing size which allows to smooth faster the low frequency of the solution as well as its high frequency. In this paper a Multi-Grid method is applied to cogging process in the software Forge 3. The study is carried out using increasing number of degrees of freedom. The results shows that calculation time is divide by two for a mesh of 39,000 nodes. The method is promising especially if coupled with Multi-Mesh method.

  13. Spatial Variability of CCN Sized Aerosol Particles

    NASA Astrophysics Data System (ADS)

    Asmi, A.; Väänänen, R.

    2014-12-01

    The computational limitations restrict the grid size used in GCM models, and for many cloud types they are too large when compared to the scale of the cloud formation processes. Several parameterizations for e.g. convective cloud formation exist, but information on spatial subgrid variation of the cloud condensation nuclei (CCNs) sized aerosol concentration is not known. We quantify this variation as a function of the spatial scale by using datasets from airborne aerosol measurement campaigns around the world including EUCAARI LONGREX, ATAR, INCA, INDOEX, CLAIRE, PEGASOS and several regional airborne campaigns in Finland. The typical shapes of the distributions are analyzed. When possible, we use information obtained by CCN counters. In some other cases, we use particle size distribution measured by for example SMPS to get approximated CCN concentration. Other instruments used include optical particle counters or condensational particle counters. When using the GCM models, the CCN concentration used for each the grid-box is often considered to be either flat, or as an arithmetic mean of the concentration inside the grid-box. However, the aircraft data shows that the concentration values are often lognormal distributed. This, combined with the subgrid variations in the land use and atmospheric properties, might cause that the aerosol-cloud interactions calculated by using mean values to vary significantly from the true effects both temporary and spatially. This, in turn, can cause non-linear bias into the GCMs. We calculate the CCN aerosol concentration distribution as a function of different spatial scales. The measurements allow us to study the variation of these distributions within from hundreds of meters up to hundreds of kilometers. This is used to quantify the potential error when mean values are used in GCMs.

  14. Local Fitting of the Kohn-Sham Density in a Gaussian and Plane Waves Scheme for Large-Scale Density Functional Theory Simulations.

    PubMed

    Golze, Dorothea; Iannuzzi, Marcella; Hutter, Jürg

    2017-05-09

    A local resolution-of-the-identity (LRI) approach is introduced in combination with the Gaussian and plane waves (GPW) scheme to enable large-scale Kohn-Sham density functional theory calculations. In GPW, the computational bottleneck is typically the description of the total charge density on real-space grids. Introducing the LRI approximation, the linear scaling of the GPW approach with respect to system size is retained, while the prefactor for the grid operations is reduced. The density fitting is an O(N) scaling process implemented by approximating the atomic pair densities by an expansion in one-center fit functions. The computational cost for the grid-based operations becomes negligible in LRIGPW. The self-consistent field iteration is up to 30 times faster for periodic systems dependent on the symmetry of the simulation cell and on the density of grid points. However, due to the overhead introduced by the local density fitting, single point calculations and complete molecular dynamics steps, including the calculation of the forces, are effectively accelerated by up to a factor of ∼10. The accuracy of LRIGPW is assessed for different systems and properties, showing that total energies, reaction energies, intramolecular and intermolecular structure parameters are well reproduced. LRIGPW yields also high quality results for extended condensed phase systems such as liquid water, ice XV, and molecular crystals.

  15. Coarse Grid Modeling of Turbine Film Cooling Flows Using Volumetric Source Terms

    NASA Technical Reports Server (NTRS)

    Heidmann, James D.; Hunter, Scott D.

    2001-01-01

    The recent trend in numerical modeling of turbine film cooling flows has been toward higher fidelity grids and more complex geometries. This trend has been enabled by the rapid increase in computing power available to researchers. However, the turbine design community requires fast turnaround time in its design computations, rendering these comprehensive simulations ineffective in the design cycle. The present study describes a methodology for implementing a volumetric source term distribution in a coarse grid calculation that can model the small-scale and three-dimensional effects present in turbine film cooling flows. This model could be implemented in turbine design codes or in multistage turbomachinery codes such as APNASA, where the computational grid size may be larger than the film hole size. Detailed computations of a single row of 35 deg round holes on a flat plate have been obtained for blowing ratios of 0.5, 0.8, and 1.0, and density ratios of 1.0 and 2.0 using a multiblock grid system to resolve the flows on both sides of the plate as well as inside the hole itself. These detailed flow fields were spatially averaged to generate a field of volumetric source terms for each conservative flow variable. Solutions were also obtained using three coarse grids having streamwise and spanwise grid spacings of 3d, 1d, and d/3. These coarse grid solutions used the integrated hole exit mass, momentum, energy, and turbulence quantities from the detailed solutions as volumetric source terms. It is shown that a uniform source term addition over a distance from the wall on the order of the hole diameter is able to predict adiabatic film effectiveness better than a near-wall source term model, while strictly enforcing correct values of integrated boundary layer quantities.

  16. Synchronization of cyclic power grids: Equilibria and stability of the synchronous state

    NASA Astrophysics Data System (ADS)

    Xi, Kaihua; Dubbeldam, Johan L. A.; Lin, Hai Xiang

    2017-01-01

    Synchronization is essential for the proper functioning of power grids; we investigate the synchronous states and their stability for cyclic power grids. We calculate the number of stable equilibria and investigate both the linear and nonlinear stabilities of the synchronous state. The linear stability analysis shows that the stability of the state, determined by the smallest nonzero eigenvalue, is inversely proportional to the size of the network. We use the energy barrier to measure the nonlinear stability and calculate it by comparing the potential energy of the type-1 saddles with that of the stable synchronous state. We find that the energy barrier depends on the network size (N) in a more complicated fashion compared to the linear stability. In particular, when the generators and consumers are evenly distributed in an alternating way, the energy barrier decreases to a constant when N approaches infinity. For a heterogeneous distribution of generators and consumers, the energy barrier decreases with N. The more heterogeneous the distribution is, the stronger the energy barrier depends on N. Finally, we find that by comparing situations with equal line loads in cyclic and tree networks, tree networks exhibit reduced stability. This difference disappears in the limit of N →∞ . This finding corroborates previous results reported in the literature and suggests that cyclic (sub)networks may be applied to enhance power transfer while maintaining stable synchronous operation.

  17. Revision of the documentation for a model for calculating effects of liquid waste disposal in deep saline aquifers

    USGS Publications Warehouse

    INTERA Environmental Consultants, Inc.

    1979-01-01

    The major limitation of the model arises using second-order correct (central-difference) finite-difference approximation in space. To avoid numerical oscillations in the solution, the user must restrict grid block and time step sizes depending upon the magnitude of the dispersivity.

  18. Assessment of time-dependent density functional theory with the restricted excitation space approximation for excited state calculations of large systems

    NASA Astrophysics Data System (ADS)

    Hanson-Heine, Magnus W. D.; George, Michael W.; Besley, Nicholas A.

    2018-06-01

    The restricted excitation subspace approximation is explored as a basis to reduce the memory storage required in linear response time-dependent density functional theory (TDDFT) calculations within the Tamm-Dancoff approximation. It is shown that excluding the core orbitals and up to 70% of the virtual orbitals in the construction of the excitation subspace does not result in significant changes in computed UV/vis spectra for large molecules. The reduced size of the excitation subspace greatly reduces the size of the subspace vectors that need to be stored when using the Davidson procedure to determine the eigenvalues of the TDDFT equations. Furthermore, additional screening of the two-electron integrals in combination with a reduction in the size of the numerical integration grid used in the TDDFT calculation leads to significant computational savings. The use of these approximations represents a simple approach to extend TDDFT to the study of large systems and make the calculations increasingly tractable using modest computing resources.

  19. Spatial Distribution of Bed Particles in Natural Boulder-Bed Streams

    NASA Astrophysics Data System (ADS)

    Clancy, K. F.; Prestegaard, K. L.

    2001-12-01

    The Wolman pebble count is used to obtain the size distribution of bed particles in natural streams. Statistics such as median particle size (D50) are used in resistance calculations. Additional information such as bed particle heterogeneity may also be obtained from the particle distribution, which is used to predict sediment transport rates (Hey, 1979), (Ferguson, Prestegaard, Ashworth, 1989). Boulder-bed streams have an extreme range of particles in the particle size distribution ranging from sand size particles to particles larger than 0.5-m. A study of a natural boulder-bed reach demonstrated that the spatial distribution of the particles is a significant factor in predicting sediment transport and stream bed and bank stability. Further experiments were performed to test the limits of the spatial distribution's effect on sediment transport. Three stream reaches 40-m in length were selected with similar hydrologic characteristics and spatial distributions but varying average size particles. We used a grid 0.5 by 0.5-m and measured four particles within each grid cell. Digital photographs of the streambed were taken in each grid cell. The photographs were examined using image analysis software to obtain particle size and position of the largest particles (D84) within the reach's particle distribution. Cross section, topography and stream depth were surveyed. Velocity and velocity profiles were measured and recorded. With these data and additional surveys of bankfull floods, we tested the significance of the spatial distributions as average particle size decreases. The spatial distribution of streambed particles may provide information about stream valley formation, bank stability, sediment transport, and the growth rate of riparian vegetation.

  20. Modeling Lidar Multiple Scattering

    NASA Astrophysics Data System (ADS)

    Sato, Kaori; Okamoto, Hajime; Ishimoto, Hiroshi

    2016-06-01

    A practical model to simulate multiply scattered lidar returns from inhomogeneous cloud layers are developed based on Backward Monte Carlo (BMC) simulations. The estimated time delay of the backscattered intensities returning from different vertical grids by the developed model agreed well with that directly obtained from BMC calculations. The method was applied to the Cloud-Aerosol Lidar and Infrared Pathfinder Satellite Observations (CALIPSO) satellite data to improve the synergetic retrieval of cloud microphysics with CloudSat radar data at optically thick cloud grids. Preliminary results for retrieving mass fraction of co-existing cloud particles and drizzle size particles within lowlevel clouds are demonstrated.

  1. The R package 'icosa' for coarse resolution global triangular and penta-hexagonal gridding

    NASA Astrophysics Data System (ADS)

    Kocsis, Adam T.

    2017-04-01

    With the development of the internet and the computational power of personal computers, open source programming environments have become indispensable for science in the past decade. This includes the increase of the GIS capacity of the free R environment, which was originally developed for statistical analyses. The flexibility of R made it a preferred programming tool in a multitude of disciplines from the area of the biological and geological sciences. Many of these subdisciplines operate with incidence (occurrence) data that are in a large number of cases to be grained before further analyses can be conducted. This graining is executed mostly by gridding data to cells of a Gaussian grid of various resolutions to increase the density of data in a single unit of the analyses. This method has obvious shortcomings despite the ease of its application: well-known systematic biases are induced to cell sizes and shapes that can interfere with the results of statistical procedures, especially if the number of incidence points influences the metrics in question. The 'icosa' package employs a common method to overcome this obstacle by implementing grids with roughly equal cell sizes and shapes that are based on tessellated icosahedra. These grid objects are essentially polyhedra with xyz Cartesian vertex data that are linked to tables of faces and edges. At its current developmental stage, the package uses a single method of tessellation which balances grid cell size and shape distortions, but its structure allows the implementation of various other types of tessellation algorithms. The resolution of the grids can be set by the number of breakpoints inserted into a segment forming an edge of the original icosahedron. Both the triangular and their inverted penta-hexagonal grids are available for creation with the package. The package also incorporates functions to look up coordinates in the grid very effectively and data containers to link data to the grid structure. The classes defined in the package are communicating with classes of the 'sp' and 'raster' packages and functions are supplied that allow resolution change and type conversions. Three-dimensional rendering is made available with the 'rgl' package and two-dimensional projections can be calculated using 'sp' and 'rgdal'. The package was developed as part of a project funded by the Deutsche Forschungsgemeinschaft (KO - 5382/1-1).

  2. Recent Enhancements To The FUN3D Flow Solver For Moving-Mesh Applications

    NASA Technical Reports Server (NTRS)

    Biedron, Robert T,; Thomas, James L.

    2009-01-01

    An unsteady Reynolds-averaged Navier-Stokes solver for unstructured grids has been extended to handle general mesh movement involving rigid, deforming, and overset meshes. Mesh deformation is achieved through analogy to elastic media by solving the linear elasticity equations. A general method for specifying the motion of moving bodies within the mesh has been implemented that allows for inherited motion through parent-child relationships, enabling simulations involving multiple moving bodies. Several example calculations are shown to illustrate the range of potential applications. For problems in which an isolated body is rotating with a fixed rate, a noninertial reference-frame formulation is available. An example calculation for a tilt-wing rotor is used to demonstrate that the time-dependent moving grid and noninertial formulations produce the same results in the limit of zero time-step size.

  3. Grid cells on steeply sloping terrain: evidence for planar rather than volumetric encoding

    PubMed Central

    Hayman, Robin M. A.; Casali, Giulio; Wilson, Jonathan J.; Jeffery, Kate J.

    2015-01-01

    Neural encoding of navigable space involves a network of structures centered on the hippocampus, whose neurons –place cells – encode current location. Input to the place cells includes afferents from the entorhinal cortex, which contains grid cells. These are neurons expressing spatially localized activity patches, or firing fields, that are evenly spaced across the floor in a hexagonal close-packed array called a grid. It is thought that grids function to enable the calculation of distances. The question arises as to whether this odometry process operates in three dimensions, and so we queried whether grids permeate three-dimensional (3D) space – that is, form a lattice – or whether they simply follow the environment surface. If grids form a 3D lattice then this lattice would ordinarily be aligned horizontally (to explain the usual hexagonal pattern observed). A tilted floor would transect several layers of this putative lattice, resulting in interruption of the hexagonal pattern. We model this prediction with simulated grid lattices, and show that the firing of a grid cell on a 40°-tilted surface should cover proportionally less of the surface, with smaller field size, fewer fields, and reduced hexagonal symmetry. However, recording of real grid cells as animals foraged on a 40°-tilted surface found that firing of grid cells was almost indistinguishable, in pattern or rate, from that on the horizontal surface, with if anything increased coverage and field number, and preserved field size. It thus appears unlikely that the sloping surface transected a lattice. However, grid cells on the slope displayed slightly degraded firing patterns, with reduced coherence and slightly reduced symmetry. These findings collectively suggest that the grid cell component of the metric representation of space is not fixed in absolute 3D space but is influenced both by the surface the animal is on and by the relationship of this surface to the horizontal, supporting the hypothesis that the neural map of space is “multi-planar” rather than fully volumetric. PMID:26236245

  4. A Simulation Algorithm to Approximate the Area of Mapped Forest Inventory Plots

    Treesearch

    William A. Bechtold; Naser E. Heravi; Matthew E. Kinkenon

    2003-01-01

    Calculating the area of polygons associated with mapped forest inventory plots can be mathematically cumbersome, especially when computing change between inventories. We developed a simulation technique that utilizes a computer-generated dot grid and geometry to estimate the area of mapped polygons within any size circle. The technique also yields a matrix of change in...

  5. A New Algorithm to Optimize Maximal Information Coefficient

    PubMed Central

    Luo, Feng; Yuan, Zheming

    2016-01-01

    The maximal information coefficient (MIC) captures dependences between paired variables, including both functional and non-functional relationships. In this paper, we develop a new method, ChiMIC, to calculate the MIC values. The ChiMIC algorithm uses the chi-square test to terminate grid optimization and then removes the restriction of maximal grid size limitation of original ApproxMaxMI algorithm. Computational experiments show that ChiMIC algorithm can maintain same MIC values for noiseless functional relationships, but gives much smaller MIC values for independent variables. For noise functional relationship, the ChiMIC algorithm can reach the optimal partition much faster. Furthermore, the MCN values based on MIC calculated by ChiMIC can capture the complexity of functional relationships in a better way, and the statistical powers of MIC calculated by ChiMIC are higher than those calculated by ApproxMaxMI. Moreover, the computational costs of ChiMIC are much less than those of ApproxMaxMI. We apply the MIC values tofeature selection and obtain better classification accuracy using features selected by the MIC values from ChiMIC. PMID:27333001

  6. Commissioning and comprehensive quality assurance of commercial 3D treatment planning system using IAEA Technical Report Series-430.

    PubMed

    Jamema, S V; Upreti, R R; Sharma, S; Deshpande, D D

    2008-09-01

    The purpose of this work is to report the results of commissioning and to establish a quality assurance (QA) program for commercial 3D treatment planning system (TPS) based on IAEA Technical Report Series 430. Eclipse v 7.3.10, (Varian Medical Systems, Palo Alto, CA, U.S.A.) TPS was commissioned for a Clinac 6EX (Varian Medical Systems, Palo Alto, CA, USA) linear accelerator. CT images of a phantom with various known in-homogeneities were acquired. The images were transferred to TPS and tested for various parameters related to patient data acquisition, anatomical modeling, plan evaluation and dose calculation. Dosimetric parameters including open, asymmetric and wedged shaped fields, oblique incidence, buildup region behavior and SSD dependence were evaluated. Representative clinical cases were tested for MU calculation and point doses. The maximum variation between the measured and the known CT numbers was 20 +/- 11.7 HU (1 SD). The results of all non-dosimetric tests were found within tolerance, however expansion at the sharp corners was found distorted. The accuracy of the DVH calculations depends on the grid size. TPS calculations of all the dosimetric parameters were in good agreement with the measured values, however for asymmetric open and wedged fields, few points were found out of tolerance. Smaller grid size calculation showed better agreement of dose calculation in the build-up region. Independent tests for MU calculation showed a variation within +/-2% (relative to planning system), meanwhile variation of 3.0% was observed when the central axis was blocked. The test results were in agreement with the tolerance specified by IAEA TRS 430. A subset of the commissioning tests has been identified as a baseline data for an ongoing QA program.

  7. A computational study of coherent structures in the wakes of two-dimensional bluff bodies

    NASA Astrophysics Data System (ADS)

    Pearce, Jeffrey Alan

    1988-08-01

    The periodic shedding of vortices from bluff bodies was first recognized in the late 1800's. Currently, there is great interest concerning the effect of vortex shedding on structures and on vehicle stability. In the design of bluff structures which will be exposed to a flow, knowledge of the shedding frequency and the amplitude of the aerodynamic forces is critical. The ability to computationally predict parameters associated with periodic vortex shedding is thus a valuable tool. In this study, the periodic shedding of vortices from several bluff body geometries is predicted. The study is conducted with a two-dimensional finite-difference code employed on various grid sizes. The effects of the grid size and time step on the accuracy of the solution are addressed. Strouhal numbers and aerodynamic force coefficients are computed for all of the bodies considered and compared with previous experimental results. Results indicate that the finite-difference code is capable of predicting periodic vortex shedding for all of the geometries tested. Refinement of the finite-difference grid was found to give little improvement in the prediction; however, the choice of time step size was shown to be critical. Predictions of Strouhal numbers were generally accurate, and the calculated aerodynamic forces generally exhibited behavior consistent with previous studies.

  8. Three-dimensional hydrodynamic Bondi-Hoyle accretion. 2: Homogeneous medium at Mach 3 with gamma = 5/3

    NASA Technical Reports Server (NTRS)

    Ruffert, Maximilian; Arnett, David

    1994-01-01

    We investigate the hydrodynamics of three-dimensional classical Bondi-Hoyle accretion. Totally absorbing spheres of varying sizes (from 10 down to 0.01 accretion radii) move at Mach 3 relative to a homogeneous and slightly perturbed medium, which is taken to be an ideal gas (gamma = 5/3). To accommodate the long-range gravitational forces, the extent of the computational volume is 32(exp 3) accretion radii. We examine the influence of numerical procedure on physical behavior. The hydrodynamics is modeled by the 'piecewise parabolic method.' No energy sources (nuclear burning) or sinks (radiation, conduction) are included. The resolution in the vicinity of the accretor is increased by multiply nesting several (5-10) grids around the sphere, each finer grid being a factor of 2 smaller in zone dimension that the next coarser grid. The largest dynamic range (ratio of size of the largest grid to size of the finest zone) is 16,384. This allows us to include a coarse model for the surface of the accretor (vacuum sphere) on the finest grid, while at the same time evolving the gas on the coarser grids. Initially (at time t = 0-10), a shock front is set up, a Mach cone develops, and the accretion column is observable. Eventually the flow becomes unstable, destroying axisymmetry. This happens approximately when the mass accretion rate reaches the values (+/- 10%) predicted by the Bondi-Hoyle accretion formula (factor of 2 included). However, our three-dimensional models do not show the highly dynamic flip-flop flow so prominent in two-dimensional calculations performed by other authors. The flow, and thus the accretion rate of all quantities, shows quasi-periodic (P approximately equals 5) cycles between quiescent and active states. The interpolation formula proposed in an accompanying paper is found to follow the collected numerical data to within approximately 30%. The specific angular momentum accreted is of the same order of magnitude as the values previously found for two-dimensional flows.

  9. Electromagnetic plasma simulation in realistic geometries

    NASA Astrophysics Data System (ADS)

    Brandon, S.; Ambrosiano, J. J.; Nielsen, D.

    1991-08-01

    Particle-in-Cell (PIC) calculations have become an indispensable tool to model the nonlinear collective behavior of charged particle species in electromagnetic fields. Traditional finite difference codes, such as CONDOR (2-D) and ARGUS (3-D), are used extensively to design experiments and develop new concepts. A wide variety of physical processes can be modeled simply and efficiently by these codes. However, experiments have become more complex. Geometrical shapes and length scales are becoming increasingly more difficult to model. Spatial resolution requirements for the electromagnetic calculation force large grids and small time steps. Many hours of CRAY YMP time may be required to complete 2-D calculation -- many more for 3-D calculations. In principle, the number of mesh points and particles need only to be increased until all relevant physical processes are resolved. In practice, the size of a calculation is limited by the computer budget. As a result, experimental design is being limited by the ability to calculate, not by the experimenters ingenuity or understanding of the physical processes involved. Several approaches to meet these computational demands are being pursued. Traditional PIC codes continue to be the major design tools. These codes are being actively maintained, optimized, and extended to handle large and more complex problems. Two new formulations are being explored to relax the geometrical constraints of the finite difference codes. A modified finite volume test code, TALUS, uses a data structure compatible with that of standard finite difference meshes. This allows a basic conformal boundary/variable grid capability to be retrofitted to CONDOR. We are also pursuing an unstructured grid finite element code, MadMax. The unstructured mesh approach provides maximum flexibility in the geometrical model while also allowing local mesh refinement.

  10. Aspects on HTS applications in confined power grids

    NASA Astrophysics Data System (ADS)

    Arndt, T.; Grundmann, J.; Kuhnert, A.; Kummeth, P.; Nick, W.; Oomen, M.; Schacherer, C.; Schmidt, W.

    2014-12-01

    In an increasing number of electric power grids the share of distributed energy generation is also increasing. The grids have to cope with a considerable change of power flow, which has an impact on the optimum topology of the grids and sub-grids (high-voltage, medium-voltage and low-voltage sub-grids) and the size of quasi-autonomous grid sections. Furthermore the stability of grids is influenced by its size. Thus special benefits of HTS applications in the power grid might become most visible in confined power grids.

  11. SU-F-T-506: Development and Commissioning of the Effective and Efficient Grid Therapy Using High Dose Rate Flattening Filter Free Beam and Multileaf Collimator

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Liu, M; Wen, N; Beyer, C

    Purpose: Treating bulky tumors with grid therapy (GT) has demonstrated high response rates. Long delivery time (∼15min), with consequent increased risk of intrafraction motion, is a major disadvantage of conventional MLC-based GT (MLC-GT). The goal of this study was to develop and commission a MLC-GT technique with similar dosimetric characteristics, but more efficient delivery. Methods: Grid plan was designed with 10X-FFF (2400MU/min) beam and MLC in a commercial treatment planning system (TPS). Grid size was 1cm by 1cm and grid-to-grid distance was 2cm. Field-in-field technique was used to flatten the dose profile at depth of 10cm. Prescription was 15Gy atmore » 1.5cm depth. Doses were verified at depths of 1.5cm, 5cm and 10cm. Point dose was measured with a plastic scintillator detector (PSD) while the planar dose was measured with calibrated Gafchromic EBT3 films in a 20cm think, 30cmx30cm solid water phantom. The measured doses were compared to the doses calculated in the treatment planning system. Percent depth dose (PDD) within the grid was also measured using EBT3 film. Five clinical cases were planned to compare beam-on time. Results: The valley-to-peak dose ratio at the 3 depths was approximately 10–15%, which is very similar to published result. The average point dose difference between the PSD measurements and TPS calculation is 2.1±0.6%. Film dosimetry revealed good agreement between the delivered and calculated dose. The average gamma passing rates at the 3 depths were 95% (3%, 1mm). The average percent difference between the measured PDD and calculated PDD was 2.1% within the depth of 20cm. The phantom plan delivery time was 3.6 min. Average beam-on time was reduced by 66.1±5.6% for the 5 clinical cases. Conclusion: An effective and efficient GT technique was developed and commissioned for the treatment of bulky tumors using FFF beam combined with MLC and automation. The Department of Radiation Oncology at Henry Ford Health System receives research support from Varian Medical Systems and Philips Health Care.« less

  12. Three-dimensional radiochromic film dosimetry for volumetric modulated arc therapy using a spiral water phantom.

    PubMed

    Tanooka, Masao; Doi, Hiroshi; Miura, Hideharu; Inoue, Hiroyuki; Niwa, Yasue; Takada, Yasuhiro; Fujiwara, Masayuki; Sakai, Toshiyuki; Sakamoto, Kiyoshi; Kamikonya, Norihiko; Hirota, Shozo

    2013-11-01

    We validated 3D radiochromic film dosimetry for volumetric modulated arc therapy (VMAT) using a newly developed spiral water phantom. The phantom consists of a main body and an insert box, each of which has an acrylic wall thickness of 3 mm and is filled with water. The insert box includes a spiral film box used for dose-distribution measurement, and a film holder for positioning a radiochromic film. The film holder has two parallel walls whose facing inner surfaces are equipped with spiral grooves in a mirrored configuration. The film is inserted into the spiral grooves by its side edges and runs along them to be positioned on a spiral plane. Dose calculation was performed by applying clinical VMAT plans to the spiral water phantom using a commercial Monte Carlo-based treatment-planning system, Monaco, whereas dose was measured by delivering the VMAT beams to the phantom. The calculated dose distributions were resampled on the spiral plane, and the dose distributions recorded on the film were scanned. Comparisons between the calculated and measured dose distributions yielded an average gamma-index pass rate of 87.0% (range, 91.2-84.6%) in nine prostate VMAT plans under 3 mm/3% criteria with a dose-calculation grid size of 2 mm. The pass rates were increased beyond 90% (average, 91.1%; range, 90.1-92.0%) when the dose-calculation grid size was decreased to 1 mm. We have confirmed that 3D radiochromic film dosimetry using the spiral water phantom is a simple and cost-effective approach to VMAT dose verification.

  13. Outstanding performance of configuration interaction singles and doubles using exact exchange Kohn-Sham orbitals in real-space numerical grid method

    NASA Astrophysics Data System (ADS)

    Lim, Jaechang; Choi, Sunghwan; Kim, Jaewook; Kim, Woo Youn

    2016-12-01

    To assess the performance of multi-configuration methods using exact exchange Kohn-Sham (KS) orbitals, we implemented configuration interaction singles and doubles (CISD) in a real-space numerical grid code. We obtained KS orbitals with the exchange-only optimized effective potential under the Krieger-Li-Iafrate (KLI) approximation. Thanks to the distinctive features of KLI orbitals against Hartree-Fock (HF), such as bound virtual orbitals with compact shapes and orbital energy gaps similar to excitation energies; KLI-CISD for small molecules shows much faster convergence as a function of simulation box size and active space (i.e., the number of virtual orbitals) than HF-CISD. The former also gives more accurate excitation energies with a few dominant configurations than the latter, even with many more configurations. The systematic control of basis set errors is straightforward in grid bases. Therefore, grid-based multi-configuration methods using exact exchange KS orbitals provide a promising new way to make accurate electronic structure calculations.

  14. SU-F-T-436: A Method to Evaluate Dosimetric Properties of SFGRT in Eclipse TPS

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Xu, M; Tobias, R; Pankuch, M

    Purpose: The objective was to develop a method for dose distribution calculation of spatially-fractionated-GRID-radiotherapy (SFGRT) in Eclipse treatment-planning-system (TPS). Methods: Patient treatment-plans with SFGRT for bulky tumors were generated in Varian Eclipse version11. A virtual structure based on the GRID pattern was created and registered to a patient CT image dataset. The virtual GRID structure was positioned on the iso-center level together with matching beam geometries to simulate a commercially available GRID block made of brass. This method overcame the difficulty in treatment-planning and dose-calculation due to the lack o-the option to insert a GRID block add-on in Eclipse TPS.more » The patient treatment-planning displayed GRID effects on the target, critical structures, and dose distribution. The dose calculations were compared to the measurement results in phantom. Results: The GRID block structure was created to follow the beam divergence to the patient CT images. The inserted virtual GRID block made it possible to calculate the dose distributions and profiles at various depths in Eclipse. The virtual GRID block was added as an option to TPS. The 3D representation of the isodose distribution of the spatially-fractionated beam was generated in axial, coronal, and sagittal planes. Physics of GRID can be different from that for fields shaped by regular blocks because the charge-particle-equilibrium cannot be guaranteed for small field openings. Output factor (OF) measurement was required to calculate the MU to deliver the prescribed dose. The calculated OF based on the virtual GRID agreed well with the measured OF in phantom. Conclusion: The method to create the virtual GRID block has been proposed for the first time in Eclipse TPS. The dosedistributions, in-plane and cross-plane profiles in PTV can be displayed in 3D-space. The calculated OF’s based on the virtual GRID model compare well to the measured OF’s for SFGRT clinical use.« less

  15. DenInv3D: a geophysical software for three-dimensional density inversion of gravity field data

    NASA Astrophysics Data System (ADS)

    Tian, Yu; Ke, Xiaoping; Wang, Yong

    2018-04-01

    This paper presents a three-dimensional density inversion software called DenInv3D that operates on gravity and gravity gradient data. The software performs inversion modelling, kernel function calculation, and inversion calculations using the improved preconditioned conjugate gradient (PCG) algorithm. In the PCG algorithm, due to the uncertainty of empirical parameters, such as the Lagrange multiplier, we use the inflection point of the L-curve as the regularisation parameter. The software can construct unequally spaced grids and perform inversions using such grids, which enables changing the resolution of the inversion results at different depths. Through inversion of airborne gradiometry data on the Australian Kauring test site, we discovered that anomalous blocks of different sizes are present within the study area in addition to the central anomalies. The software of DenInv3D can be downloaded from http://159.226.162.30.

  16. Importance of Grid Center Arrangement

    NASA Astrophysics Data System (ADS)

    Pasaogullari, O.; Usul, N.

    2012-12-01

    In Digital Elevation Modeling, grid size is accepted to be the most important parameter. Despite the point density and/or scale of the source data, it is freely decided by the user. Most of the time, arrangement of the grid centers are ignored, even most GIS packages omit the choice of grid center coordinate selection. In our study; importance of the arrangement of grid centers is investigated. Using the analogy between "Raster Grid DEM" and "Bitmap Image", importance of placement of grid centers in DEMs are measured. The study has been conducted on four different grid DEMs obtained from a half ellipsoid. These grid DEMs are obtained in such a way that they are half grid size apart from each other. Resulting grid DEMs are investigated through similarity measures. Image processing scientists use different measures to investigate the dis/similarity between the images and the amount of different information they carry. Grid DEMs are projected to a finer grid in order to co-center. Similarity measures are then applied to each grid DEM pairs. These similarity measures are adapted to DEM with band reduction and real number operation. One of the measures gives function graph and the others give measure matrices. Application of similarity measures to six grid DEM pairs shows interesting results. These four different grid DEMs are created with the same method for the same area, surprisingly; thirteen out of 14 measures state that, the half grid size apart grid DEMs are different from each other. The results indicated that although grid DEMs carry mutual information, they have also additional individual information. In other words, half grid size apart constructed grid DEMs have non-redundant information.; Joint Probability Distributions Function Graphs

  17. An adaptive grid to improve the efficiency and accuracy of modelling underwater noise from shipping

    NASA Astrophysics Data System (ADS)

    Trigg, Leah; Chen, Feng; Shapiro, Georgy; Ingram, Simon; Embling, Clare

    2017-04-01

    Underwater noise from shipping is becoming a significant concern and has been listed as a pollutant under Descriptor 11 of the Marine Strategy Framework Directive. Underwater noise models are an essential tool to assess and predict noise levels for regulatory procedures such as environmental impact assessments and ship noise monitoring. There are generally two approaches to noise modelling. The first is based on simplified energy flux models, assuming either spherical or cylindrical propagation of sound energy. These models are very quick but they ignore important water column and seabed properties, and produce significant errors in the areas subject to temperature stratification (Shapiro et al., 2014). The second type of model (e.g. ray-tracing and parabolic equation) is based on an advanced physical representation of sound propagation. However, these acoustic propagation models are computationally expensive to execute. Shipping noise modelling requires spatial discretization in order to group noise sources together using a grid. A uniform grid size is often selected to achieve either the greatest efficiency (i.e. speed of computations) or the greatest accuracy. In contrast, this work aims to produce efficient and accurate noise level predictions by presenting an adaptive grid where cell size varies with distance from the receiver. The spatial range over which a certain cell size is suitable was determined by calculating the distance from the receiver at which propagation loss becomes uniform across a grid cell. The computational efficiency and accuracy of the resulting adaptive grid was tested by comparing it to uniform 1 km and 5 km grids. These represent an accurate and computationally efficient grid respectively. For a case study of the Celtic Sea, an application of the adaptive grid over an area of 160×160 km reduced the number of model executions required from 25600 for a 1 km grid to 5356 in December and to between 5056 and 13132 in August, which represents a 2 to 5-fold increase in efficiency. The 5 km grid reduces the number of model executions further to 1024. However, over the first 25 km the 5 km grid produces errors of up to 13.8 dB when compared to the highly accurate but inefficient 1 km grid. The newly developed adaptive grid generates much smaller errors of less than 0.5 dB while demonstrating high computational efficiency. Our results show that the adaptive grid provides the ability to retain the accuracy of noise level predictions and improve the efficiency of the modelling process. This can help safeguard sensitive marine ecosystems from noise pollution by improving the underwater noise predictions that inform management activities. References Shapiro, G., Chen, F., Thain, R., 2014. The Effect of Ocean Fronts on Acoustic Wave Propagation in a Shallow Sea, Journal of Marine System, 139: 217 - 226. http://dx.doi.org/10.1016/j.jmarsys.2014.06.007.

  18. Small-mammal density estimation: A field comparison of grid-based vs. web-based density estimators

    USGS Publications Warehouse

    Parmenter, R.R.; Yates, Terry L.; Anderson, D.R.; Burnham, K.P.; Dunnum, J.L.; Franklin, A.B.; Friggens, M.T.; Lubow, B.C.; Miller, M.; Olson, G.S.; Parmenter, Cheryl A.; Pollard, J.; Rexstad, E.; Shenk, T.M.; Stanley, T.R.; White, Gary C.

    2003-01-01

    Statistical models for estimating absolute densities of field populations of animals have been widely used over the last century in both scientific studies and wildlife management programs. To date, two general classes of density estimation models have been developed: models that use data sets from capture–recapture or removal sampling techniques (often derived from trapping grids) from which separate estimates of population size (NÌ‚) and effective sampling area (AÌ‚) are used to calculate density (DÌ‚ = NÌ‚/AÌ‚); and models applicable to sampling regimes using distance-sampling theory (typically transect lines or trapping webs) to estimate detection functions and densities directly from the distance data. However, few studies have evaluated these respective models for accuracy, precision, and bias on known field populations, and no studies have been conducted that compare the two approaches under controlled field conditions. In this study, we evaluated both classes of density estimators on known densities of enclosed rodent populations. Test data sets (n = 11) were developed using nine rodent species from capture–recapture live-trapping on both trapping grids and trapping webs in four replicate 4.2-ha enclosures on the Sevilleta National Wildlife Refuge in central New Mexico, USA. Additional “saturation” trapping efforts resulted in an enumeration of the rodent populations in each enclosure, allowing the computation of true densities. Density estimates (DÌ‚) were calculated using program CAPTURE for the grid data sets and program DISTANCE for the web data sets, and these results were compared to the known true densities (D) to evaluate each model's relative mean square error, accuracy, precision, and bias. In addition, we evaluated a variety of approaches to each data set's analysis by having a group of independent expert analysts calculate their best density estimates without a priori knowledge of the true densities; this “blind” test allowed us to evaluate the influence of expertise and experience in calculating density estimates in comparison to simply using default values in programs CAPTURE and DISTANCE. While the rodent sample sizes were considerably smaller than the recommended minimum for good model results, we found that several models performed well empirically, including the web-based uniform and half-normal models in program DISTANCE, and the grid-based models Mb and Mbh in program CAPTURE (with AÌ‚ adjusted by species-specific full mean maximum distance moved (MMDM) values). These models produced accurate DÌ‚ values (with 95% confidence intervals that included the true D values) and exhibited acceptable bias but poor precision. However, in linear regression analyses comparing each model's DÌ‚ values to the true D values over the range of observed test densities, only the web-based uniform model exhibited a regression slope near 1.0; all other models showed substantial slope deviations, indicating biased estimates at higher or lower density values. In addition, the grid-based DÌ‚ analyses using full MMDM values for WÌ‚ area adjustments required a number of theoretical assumptions of uncertain validity, and we therefore viewed their empirical successes with caution. Finally, density estimates from the independent analysts were highly variable, but estimates from web-based approaches had smaller mean square errors and better achieved confidence-interval coverage of D than did grid-based approaches. Our results support the contention that web-based approaches for density estimation of small-mammal populations are both theoretically and empirically superior to grid-based approaches, even when sample size is far less than often recommended. In view of the increasing need for standardized environmental measures for comparisons among ecosystems and through time, analytical models based on distance sampling appear to offer accurate density estimation approaches for research studies involving small-mammal abundances.

  19. A Fast Full Tensor Gravity computation algorithm for High Resolution 3D Geologic Interpretations

    NASA Astrophysics Data System (ADS)

    Jayaram, V.; Crain, K.; Keller, G. R.

    2011-12-01

    We present an algorithm to rapidly calculate the vertical gravity and full tensor gravity (FTG) values due to a 3-D geologic model. This algorithm can be implemented on single, multi-core CPU and graphical processing units (GPU) architectures. Our technique is based on the line element approximation with a constant density within each grid cell. This type of parameterization is well suited for high-resolution elevation datasets with grid size typically in the range of 1m to 30m. The large high-resolution data grids in our studies employ a pre-filtered mipmap pyramid type representation for the grid data known as the Geometry clipmap. The clipmap was first introduced by Microsoft Research in 2004 to do fly-through terrain visualization. This method caches nested rectangular extents of down-sampled data layers in the pyramid to create view-dependent calculation scheme. Together with the simple grid structure, this allows the gravity to be computed conveniently on-the-fly, or stored in a highly compressed format. Neither of these capabilities has previously been available. Our approach can perform rapid calculations on large topographies including crustal-scale models derived from complex geologic interpretations. For example, we used a 1KM Sphere model consisting of 105000 cells at 10m resolution with 100000 gravity stations. The line element approach took less than 90 seconds to compute the FTG and vertical gravity on an Intel Core i7 CPU at 3.07 GHz utilizing just its single core. Also, unlike traditional gravity computational algorithms, the line-element approach can calculate gravity effects at locations interior or exterior to the model. The only condition that must be met is the observation point cannot be located directly above the line element. Therefore, we perform a location test and then apply appropriate formulation to those data points. We will present and compare the computational performance of the traditional prism method versus the line element approach on different CPU-GPU system configurations. The algorithm calculates the expected gravity at station locations where the observed gravity and FTG data were acquired. This algorithm can be used for all fast forward model calculations of 3D geologic interpretations for data from airborne, space and submarine gravity, and FTG instrumentation.

  20. Molecular docking based screening of compounds against VP40 from Ebola virus.

    PubMed

    M Alam El-Din, Hanaa; A Loutfy, Samah; Fathy, Nasra; H Elberry, Mostafa; M Mayla, Ahmed; Kassem, Sara; Naqvi, Asif

    2016-01-01

    Ebola virus causes severe and often fatal hemorrhagic fevers in humans. The 2014 Ebola epidemic affected multiple countries. The virus matrix protein (VP40) plays a central role in virus assembly and budding. Since there is no FDA-approved vaccine or medicine against Ebola viral infection, discovering new compounds with different binding patterns against it is required. Therefore, we aim to identify small molecules that target the Arg 134 RNA binding and active site of VP40 protein. 1800 molecules were retrieved from PubChem compound database based on Structure Similarity and Conformers of pyrimidine-2, 4-dione. Molecular docking approach using Lamarckian Genetic Algorithm was carried out to find the potent inhibitors for VP40 based on calculated ligand-protein pairwise interaction energies. The grid maps representing the protein were calculated using auto grid and grid size was set to 60*60*60 points with grid spacing of 0.375 Ǻ. Ten independent docking runs were carried out for each ligand and results were clustered according to the 1.0 Ǻ RMSD criteria. The post-docking analysis showed that binding energies ranged from -8.87 to 0.6 Kcal/mol. We report 7 molecules, which showed promising ADMET results, LD-50, as well as H-bond interaction in the binding pocket. The small molecules discovered could act as potential inhibitors for VP40 and could interfere with virus assembly and budding process.

  1. Molecular docking based screening of compounds against VP40 from Ebola virus

    PubMed Central

    M Alam El-Din, Hanaa; A. Loutfy, Samah; Fathy, Nasra; H Elberry, Mostafa; M Mayla, Ahmed; Kassem, Sara; Naqvi, Asif

    2016-01-01

    Ebola virus causes severe and often fatal hemorrhagic fevers in humans. The 2014 Ebola epidemic affected multiple countries. The virus matrix protein (VP40) plays a central role in virus assembly and budding. Since there is no FDA-approved vaccine or medicine against Ebola viral infection, discovering new compounds with different binding patterns against it is required. Therefore, we aim to identify small molecules that target the Arg 134 RNA binding and active site of VP40 protein. 1800 molecules were retrieved from PubChem compound database based on Structure Similarity and Conformers of pyrimidine-2, 4-dione. Molecular docking approach using Lamarckian Genetic Algorithm was carried out to find the potent inhibitors for VP40 based on calculated ligand-protein pairwise interaction energies. The grid maps representing the protein were calculated using auto grid and grid size was set to 60*60*60 points with grid spacing of 0.375 Ǻ. Ten independent docking runs were carried out for each ligand and results were clustered according to the 1.0 Ǻ RMSD criteria. The post-docking analysis showed that binding energies ranged from -8.87 to 0.6 Kcal/mol. We report 7 molecules, which showed promising ADMET results, LD-50, as well as H-bond interaction in the binding pocket. The small molecules discovered could act as potential inhibitors for VP40 and could interfere with virus assembly and budding process. PMID:28149054

  2. Effect of mosaic representation of vegetation in land surface schemes on simulated energy and carbon balances

    NASA Astrophysics Data System (ADS)

    Li, R.; Arora, V. K.

    2012-01-01

    Energy and carbon balance implications of representing vegetation using a composite or mosaic approach in a land surface scheme are investigated. In the composite approach the attributes of different plant functional types (PFTs) present in a grid cell are aggregated in some fashion for energy and water balance calculations. The resulting physical environmental conditions (including net radiation, soil moisture and soil temperature) are common to all PFTs and affect their ecosystem processes. In the mosaic approach energy and water balance calculations are performed separately for each PFT tile using its own vegetation attributes, so each PFT "sees" different physical environmental conditions and its carbon balance evolves somewhat differently from that in the composite approach. Simulations are performed at selected boreal, temperate and tropical locations to illustrate the differences caused by using the composite versus mosaic approaches of representing vegetation. These idealized simulations use 50% fractional coverage for each of the two dominant PFTs in a grid cell. Differences in simulated grid averaged primary energy fluxes at selected sites are generally less than 5% between the two approaches. Simulated grid-averaged carbon fluxes and pool sizes at these sites can, however, differ by as much as 46%. Simulation results suggest that differences in carbon balance between the two approaches arise primarily through differences in net radiation which directly affects net primary productivity, and thus leaf area index and vegetation biomass.

  3. Optimal Padding for the Two-Dimensional Fast Fourier Transform

    NASA Technical Reports Server (NTRS)

    Dean, Bruce H.; Aronstein, David L.; Smith, Jeffrey S.

    2011-01-01

    One-dimensional Fast Fourier Transform (FFT) operations work fastest on grids whose size is divisible by a power of two. Because of this, padding grids (that are not already sized to a power of two) so that their size is the next highest power of two can speed up operations. While this works well for one-dimensional grids, it does not work well for two-dimensional grids. For a two-dimensional grid, there are certain pad sizes that work better than others. Therefore, the need exists to generalize a strategy for determining optimal pad sizes. There are three steps in the FFT algorithm. The first is to perform a one-dimensional transform on each row in the grid. The second step is to transpose the resulting matrix. The third step is to perform a one-dimensional transform on each row in the resulting grid. Steps one and three both benefit from padding the row to the next highest power of two, but the second step needs a novel approach. An algorithm was developed that struck a balance between optimizing the grid pad size with prime factors that are small (which are optimal for one-dimensional operations), and with prime factors that are large (which are optimal for two-dimensional operations). This algorithm optimizes based on average run times, and is not fine-tuned for any specific application. It increases the amount of times that processor-requested data is found in the set-associative processor cache. Cache retrievals are 4-10 times faster than conventional memory retrievals. The tested implementation of the algorithm resulted in faster execution times on all platforms tested, but with varying sized grids. This is because various computer architectures process commands differently. The test grid was 512 512. Using a 540 540 grid on a Pentium V processor, the code ran 30 percent faster. On a PowerPC, a 256x256 grid worked best. A Core2Duo computer preferred either a 1040x1040 (15 percent faster) or a 1008x1008 (30 percent faster) grid. There are many industries that can benefit from this algorithm, including optics, image-processing, signal-processing, and engineering applications.

  4. Three-dimensional radiochromic film dosimetry for volumetric modulated arc therapy using a spiral water phantom

    PubMed Central

    Tanooka, Masao; Doi, Hiroshi; Miura, Hideharu; Inoue, Hiroyuki; Niwa, Yasue; Takada, Yasuhiro; Fujiwara, Masayuki; Sakai, Toshiyuki; Sakamoto, Kiyoshi; Kamikonya, Norihiko; Hirota, Shozo

    2013-01-01

    We validated 3D radiochromic film dosimetry for volumetric modulated arc therapy (VMAT) using a newly developed spiral water phantom. The phantom consists of a main body and an insert box, each of which has an acrylic wall thickness of 3 mm and is filled with water. The insert box includes a spiral film box used for dose-distribution measurement, and a film holder for positioning a radiochromic film. The film holder has two parallel walls whose facing inner surfaces are equipped with spiral grooves in a mirrored configuration. The film is inserted into the spiral grooves by its side edges and runs along them to be positioned on a spiral plane. Dose calculation was performed by applying clinical VMAT plans to the spiral water phantom using a commercial Monte Carlo-based treatment-planning system, Monaco, whereas dose was measured by delivering the VMAT beams to the phantom. The calculated dose distributions were resampled on the spiral plane, and the dose distributions recorded on the film were scanned. Comparisons between the calculated and measured dose distributions yielded an average gamma-index pass rate of 87.0% (range, 91.2–84.6%) in nine prostate VMAT plans under 3 mm/3% criteria with a dose-calculation grid size of 2 mm. The pass rates were increased beyond 90% (average, 91.1%; range, 90.1–92.0%) when the dose-calculation grid size was decreased to 1 mm. We have confirmed that 3D radiochromic film dosimetry using the spiral water phantom is a simple and cost-effective approach to VMAT dose verification. PMID:23685667

  5. A single-source photon source model of a linear accelerator for Monte Carlo dose calculation

    PubMed Central

    Glatting, Gerhard; Wenz, Frederik; Fleckenstein, Jens

    2017-01-01

    Purpose To introduce a new method of deriving a virtual source model (VSM) of a linear accelerator photon beam from a phase space file (PSF) for Monte Carlo (MC) dose calculation. Materials and methods A PSF of a 6 MV photon beam was generated by simulating the interactions of primary electrons with the relevant geometries of a Synergy linear accelerator (Elekta AB, Stockholm, Sweden) and recording the particles that reach a plane 16 cm downstream the electron source. Probability distribution functions (PDFs) for particle positions and energies were derived from the analysis of the PSF. These PDFs were implemented in the VSM using inverse transform sampling. To model particle directions, the phase space plane was divided into a regular square grid. Each element of the grid corresponds to an area of 1 mm2 in the phase space plane. The average direction cosines, Pearson correlation coefficient (PCC) between photon energies and their direction cosines, as well as the PCC between the direction cosines were calculated for each grid element. Weighted polynomial surfaces were then fitted to these 2D data. The weights are used to correct for heteroscedasticity across the phase space bins. The directions of the particles created by the VSM were calculated from these fitted functions. The VSM was validated against the PSF by comparing the doses calculated by the two methods for different square field sizes. The comparisons were performed with profile and gamma analyses. Results The doses calculated with the PSF and VSM agree to within 3% /1 mm (>95% pixel pass rate) for the evaluated fields. Conclusion A new method of deriving a virtual photon source model of a linear accelerator from a PSF file for MC dose calculation was developed. Validation results show that the doses calculated with the VSM and the PSF agree to within 3% /1 mm. PMID:28886048

  6. A single-source photon source model of a linear accelerator for Monte Carlo dose calculation.

    PubMed

    Nwankwo, Obioma; Glatting, Gerhard; Wenz, Frederik; Fleckenstein, Jens

    2017-01-01

    To introduce a new method of deriving a virtual source model (VSM) of a linear accelerator photon beam from a phase space file (PSF) for Monte Carlo (MC) dose calculation. A PSF of a 6 MV photon beam was generated by simulating the interactions of primary electrons with the relevant geometries of a Synergy linear accelerator (Elekta AB, Stockholm, Sweden) and recording the particles that reach a plane 16 cm downstream the electron source. Probability distribution functions (PDFs) for particle positions and energies were derived from the analysis of the PSF. These PDFs were implemented in the VSM using inverse transform sampling. To model particle directions, the phase space plane was divided into a regular square grid. Each element of the grid corresponds to an area of 1 mm2 in the phase space plane. The average direction cosines, Pearson correlation coefficient (PCC) between photon energies and their direction cosines, as well as the PCC between the direction cosines were calculated for each grid element. Weighted polynomial surfaces were then fitted to these 2D data. The weights are used to correct for heteroscedasticity across the phase space bins. The directions of the particles created by the VSM were calculated from these fitted functions. The VSM was validated against the PSF by comparing the doses calculated by the two methods for different square field sizes. The comparisons were performed with profile and gamma analyses. The doses calculated with the PSF and VSM agree to within 3% /1 mm (>95% pixel pass rate) for the evaluated fields. A new method of deriving a virtual photon source model of a linear accelerator from a PSF file for MC dose calculation was developed. Validation results show that the doses calculated with the VSM and the PSF agree to within 3% /1 mm.

  7. Fast and precise dense grid size measurement method based on coaxial dual optical imaging system

    NASA Astrophysics Data System (ADS)

    Guo, Jiping; Peng, Xiang; Yu, Jiping; Hao, Jian; Diao, Yan; Song, Tao; Li, Ameng; Lu, Xiaowei

    2015-10-01

    Test sieves with dense grid structure are widely used in many fields, accurate gird size calibration is rather critical for success of grading analysis and test sieving. But traditional calibration methods suffer from the disadvantages of low measurement efficiency and shortage of sampling number of grids which could lead to quality judgment risk. Here, a fast and precise test sieve inspection method is presented. Firstly, a coaxial imaging system with low and high optical magnification probe is designed to capture the grid images of the test sieve. Then, a scaling ratio between low and high magnification probes can be obtained by the corresponding grids in captured images. With this, all grid dimensions in low magnification image can be obtained by measuring few corresponding grids in high magnification image with high accuracy. Finally, by scanning the stage of the tri-axis platform of the measuring apparatus, whole surface of the test sieve can be quickly inspected. Experiment results show that the proposed method can measure the test sieves with higher efficiency compare to traditional methods, which can measure 0.15 million grids (gird size 0.1mm) within only 60 seconds, and it can measure grid size range from 20μm to 5mm precisely. In a word, the presented method can calibrate the grid size of test sieve automatically with high efficiency and accuracy. By which, surface evaluation based on statistical method can be effectively implemented, and the quality judgment will be more reasonable.

  8. Dynamic fisheye grids for binary black hole simulations

    NASA Astrophysics Data System (ADS)

    Zilhão, Miguel; Noble, Scott C.

    2014-03-01

    We present a new warped gridding scheme adapted to simulating gas dynamics in binary black hole spacetimes. The grid concentrates grid points in the vicinity of each black hole to resolve the smaller scale structures there, and rarefies grid points away from each black hole to keep the overall problem size at a practical level. In this respect, our system can be thought of as a ‘double’ version of the fisheye coordinate system, used before in numerical relativity codes for evolving binary black holes. The gridding scheme is constructed as a mapping between a uniform coordinate system—in which the equations of motion are solved—to the distorted system representing the spatial locations of our grid points. Since we are motivated to eventually use this system for circumbinary disc calculations, we demonstrate how the distorted system can be constructed to asymptote to the typical spherical polar coordinate system, amenable to efficiently simulating orbiting gas flows about central objects with little numerical diffusion. We discuss its implementation in the Harm3d code, tailored to evolve the magnetohydrodynamics equations in curved spacetimes. We evaluate the performance of the system’s implementation in Harm3d with a series of tests, such as the advected magnetic field loop test, magnetized Bondi accretion, and evolutions of hydrodynamic discs about a single black hole and about a binary black hole. Like we have done with Harm3d, this gridding scheme can be implemented in other unigrid codes as a (possibly) simpler alternative to adaptive mesh refinement.

  9. A fast dynamic grid adaption scheme for meteorological flows

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fiedler, B.H.; Trapp, R.J.

    1993-10-01

    The continuous dynamic grid adaption (CDGA) technique is applied to a compressible, three-dimensional model of a rising thermal. The computational cost, per grid point per time step, of using CDGA instead of a fixed, uniform Cartesian grid is about 53% of the total cost of the model with CDGA. The use of general curvilinear coordinates contributes 11.7% to this total, calculating and moving the grid 6.1%, and continually updating the transformation relations 20.7%. Costs due to calculations that involve the gridpoint velocities (as well as some substantial unexplained costs) contribute the remaining 14.5%. A simple way to limit the costmore » of calculating the grid is presented. The grid is adapted by solving an elliptic equation for gridpoint coordinates on a coarse grid and then interpolating the full finite-difference grid. In this application, the additional costs per grid point of CDGA are shown to be easily offset by the savings resulting from the reduction in the required number of grid points. In simulation of the thermal costs are reduced by a factor of 3, as compared with those of a companion model with a fixed, uniform Cartesian grid. 8 refs., 8 figs.« less

  10. Predicting grid-size-dependent fracture strains of DP980 with a microstructure-based post-necking model

    DOE PAGES

    Cheng, G.; Hu, X. H.; Choi, K. S.; ...

    2017-07-08

    Ductile fracture is a local phenomenon, and it is well established that fracture strain levels depend on both stress triaxiality and the resolution (grid size) of strain measurements. Two-dimensional plane strain post-necking models with different model sizes are used in this paper to predict the grid-size-dependent fracture strain of a commercial dual-phase steel, DP980. The models are generated from the actual microstructures, and the individual phase flow properties and literature-based individual phase damage parameters for the Johnson–Cook model are used for ferrite and martensite. A monotonic relationship is predicted: the smaller the model size, the higher the fracture strain. Thus,more » a general framework is developed to quantify the grid-size-dependent fracture strains for multiphase materials. In addition to the grid-size dependency, the influences of intrinsic microstructure features, i.e., the flow curve and fracture strains of the two constituent phases, on the predicted fracture strains also are examined. Finally, application of the derived fracture strain versus model size relationship is demonstrated with large clearance trimming simulations with different element sizes.« less

  11. Predicting grid-size-dependent fracture strains of DP980 with a microstructure-based post-necking model

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cheng, G.; Hu, X. H.; Choi, K. S.

    Ductile fracture is a local phenomenon, and it is well established that fracture strain levels depend on both stress triaxiality and the resolution (grid size) of strain measurements. Two-dimensional plane strain post-necking models with different model sizes are used in this paper to predict the grid-size-dependent fracture strain of a commercial dual-phase steel, DP980. The models are generated from the actual microstructures, and the individual phase flow properties and literature-based individual phase damage parameters for the Johnson–Cook model are used for ferrite and martensite. A monotonic relationship is predicted: the smaller the model size, the higher the fracture strain. Thus,more » a general framework is developed to quantify the grid-size-dependent fracture strains for multiphase materials. In addition to the grid-size dependency, the influences of intrinsic microstructure features, i.e., the flow curve and fracture strains of the two constituent phases, on the predicted fracture strains also are examined. Finally, application of the derived fracture strain versus model size relationship is demonstrated with large clearance trimming simulations with different element sizes.« less

  12. Characterizing the Spatial Contiguity of Extreme Precipitation over the US in the Recent Past

    NASA Astrophysics Data System (ADS)

    Touma, D. E.; Swain, D. L.; Diffenbaugh, N. S.

    2016-12-01

    The spatial characteristics of extreme precipitation over an area can define the hydrologic response in a basin, subsequently affecting the flood risk in the region. Here, we examine the spatial extent of extreme precipitation in the US by defining its "footprint": a contiguous area of rainfall exceeding a certain threshold (e.g., 90th percentile) on a given day. We first characterize the climatology of extreme rainfall footprint sizes across the US from 1980-2015 using Daymet, a high-resolution observational gridded rainfall dataset. We find that there are distinct regional and seasonal differences in average footprint sizes of extreme daily rainfall. In the winter, the Midwest shows footprints exceeding 500,000 sq. km while the Front Range exhibits footprints of 10,000 sq. km. Alternatively, the summer average footprint size is generally smaller and more uniform across the US, ranging from 10,000 sq. km in the Southwest to 100,000 sq. km in Montana and North Dakota. Moreover, we find that there are some significant increasing trends of average footprint size between 1980-2015, specifically in the Southwest in the winter and the Northeast in the spring. While gridded daily rainfall datasets allow for a practical framework in calculating footprint size, this calculation heavily depends on the interpolation methods that have been used in creating the dataset. Therefore, we assess footprint size using the GHCN-Daily station network and use geostatistical methods to define footprints of extreme rainfall directly from station data. Compared to the findings from Daymet, preliminary results using this method show fewer small daily footprint sizes over the US while large footprints are of similar number and magnitude to Daymet. Overall, defining the spatial characteristics of extreme rainfall as well as observed and expected changes in these characteristics allows us to better understand the hydrologic response to extreme rainfall and how to better characterize flood risks.

  13. Efficient radiative transfer methods for continuum and line transfer in large three-dimensional models

    NASA Astrophysics Data System (ADS)

    Juvela, Mika J.

    The relationship between physical conditions of an interstellar cloud and the observed radiation is defined by the radiative transfer problem. Radiative transfer calculations are needed if, e.g., one wants to disentangle abundance variations from excitation effects or wants to model variations of dust properties inside an interstellar cloud. New observational facilities (e.g., ALMA and Herschel) will bring improved accuracy both in terms of intensity and spatial resolution. This will enable detailed studies of the densest sub-structures of interstellar clouds and star forming regions. Such observations must be interpreted with accurate radiative transfer methods and realistic source models. In many cases this will mean modelling in three dimensions. High optical depths and observed wide range of linear scales are, however, challenging for radiative transfer modelling. A large range of linear scales can be accessed only with hierarchical models. Figure 1 shows an example of the use of a hierarchical grid for radiative transfer calculations when the original model cloud (L=10 pc, =500 cm-3) was based a MHD simulation carried out on a regular grid (Juvela & Padoan, 2005). For computed line intensities an accuracy of 10% was still reached when the number of individual cells (and the run time) was reduced by a factor of ten. This illustrates how, as long as cloud is not extremely optically thick, most of the emission comes from a small sub-volume. It is also worth noting that while errors are ~10% for any given point they are much smaller when compared with intensity variations. In particular, calculations on hierarchical grid recovered the spatial power spectrum of line emission with very good accuracy. Monte Carlo codes are used widely in both continuum and line transfer calculations. Like any lambda iteration schemes these suffer from slow convergence when models are optically thick. In line transfer Accelerated Monte Carlo methods (AMC) present a partial solution to this problem (Juvela & Padoan, 2000; Hogerheijde & van der Tak, 2000). AMC methods can be used similarly in continuum calculations to speed up the computation of dust temperatures (Juvela, 2005). The sampling problems associated with high optical depths can be solved with weighted sampling and the handling of models with τV ~ 1000 is perfectly feasible. Transiently heated small dust grains pose another problem because the calculation of their temperature distribution is very time consuming. However, a 3D model will contain thousands of cells at very similar conditions. If dust temperature distributions are calculated only once for such a set an approximate solution can be found in a much shorter time time. (Juvela & Padoan, 2003; see Figure 2a). MHD simulations with Automatic Mesh Refinement (AMR) techniques present an exciting development for the modelling of interstellar clouds. Cloud models consist of a hierarchy of grids with different grid steps and the ratio between the cloud size and the smallest resolution elements can be 106 or even larger. We are currently working on radiative transfer codes (line and continuum) that could be used efficiently on such grids (see Figure 2b). The radiative transfer problem can be solved relatively independently on each of the sub-grids. This means that the use of convergence acceleration methods can be limited to those sub-grids where they are needed and, on the other hand, parallelization of the code is straightforward.

  14. Constructing a Climatology of Whistler Wave Energy from Lightning in Low Earth Orbit

    DTIC Science & Technology

    2011-12-16

    geocentric coordinates are not equal area) and a great circle distance between the grid centers, an additional normalization is included to account for the...calculated at each corner of a 1°x1° geocentric grid as discussed for the apex latitude calculations; pseudopower was calculated within each grid...1°x1° geocentric grid at the altitude in question. Along with the model at 660 km (Figure 15a and 15b) and the conjugate location at 660 km (Figure

  15. Transonic cascade flow calculations using non-periodic C-type grids

    NASA Technical Reports Server (NTRS)

    Arnone, Andrea; Liou, Meng-Sing; Povinelli, Louis A.

    1991-01-01

    A new kind of C-type grid is proposed for turbomachinery flow calculations. This grid is nonperiodic on the wake and results in minimum skewness for cascades with high turning and large camber. Euler and Reynolds averaged Navier-Stokes equations are discretized on this type of grid using a finite volume approach. The Baldwin-Lomax eddy-viscosity model is used for turbulence closure. Jameson's explicit Runge-Kutta scheme is adopted for the integration in time, and computational efficiency is achieved through accelerating strategies such as multigriding and residual smoothing. A detailed numerical study was performed for a turbine rotor and for a vane. A grid dependence analysis is presented and the effect of artificial dissipation is also investigated. Comparison of calculations with experiments clearly demonstrates the advantage of the proposed grid.

  16. Filter size definition in anisotropic subgrid models for large eddy simulation on irregular grids

    NASA Astrophysics Data System (ADS)

    Abbà, Antonella; Campaniello, Dario; Nini, Michele

    2017-06-01

    The definition of the characteristic filter size to be used for subgrid scales models in large eddy simulation using irregular grids is still an unclosed problem. We investigate some different approaches to the definition of the filter length for anisotropic subgrid scale models and we propose a tensorial formulation based on the inertial ellipsoid of the grid element. The results demonstrate an improvement in the prediction of several key features of the flow when the anisotropicity of the grid is explicitly taken into account with the tensorial filter size.

  17. Evaluation of myocardial defect detection between parallel-hole and fan-beam SPECT using the Hotelling trace

    NASA Astrophysics Data System (ADS)

    Wollenweber, S. D.; Tsui, B. M. W.; Lalush, D. S.; Frey, E. C.; Gullberg, G. T.

    1998-08-01

    The objective of this study was to implement the Hotelling trace (HT) to evaluate the potential increase in defect detection in myocardial SPECT using high-resolution fan-beam (HRF) versus parallel-hole (HRP) collimation and compare results to a previously reported human observer study (G.K. Gregoriou et al., ibid., vol. 42, p. 1267-75, 1995). Projection data from the 3D MCAT torso phantom were simulated including the effects of attenuation, collimator-detector response blurring and scatter. Poisson noise fluctuations were then simulated. The HRP and HRF collimators had the same spatial resolution at 20 cm. The total counts in the projection data sets were proportional to the detection efficiencies of the collimators and on the order of that found in clinical Tc-99m studies. In six left-ventricular defect locations, the HT found for HRF was superior to that for HRP collimation. For HRF collimation, the HT was calculated for reconstructed images using 64/spl times/64, 128/spl times/128 and 192/spl times/192 grid sizes. The results demonstrate substantial improvement in myocardial defect detection when the grid size was increased from 64/spl times/64 to 128/spl times/128 and slight improvement from 128/spl times/128 to 192/spl times/192. Also, the performance of the Hotelling observer in terms of the HT at the different grid sizes correlates at better than 0.95 to that found in human observers in a previously reported observer experiment and ROC study.

  18. Effect of mosaic representation of vegetation in land surface schemes on simulated energy and carbon balances

    NASA Astrophysics Data System (ADS)

    Li, R.; Arora, V. K.

    2011-06-01

    Energy and carbon balance implications of representing vegetation using a composite or mosaic approach in a land surface scheme are investigated. In the composite approach the attributes of different plant functional types (PFTs) present in a grid cell are aggregated in some fashion for energy and water balance calculations. The resulting physical environmental conditions (including net radiation, soil moisture and soil temperature) are common to all PFTs and affect their ecosystem processes. In the mosaic approach energy and water balance calculations are performed separately for each PFT tile using its own vegetation attributes, so each PFT "sees" different physical environmental conditions and its carbon balance evolves somewhat differently from that in the composite approach. Simulations are performed at selected boreal, temperate and tropical locations to illustrate the differences caused by using the composite versus the mosaic approaches of representing vegetation. Differences in grid averaged primary energy fluxes are generally less than 5 % between the two approaches. Grid-averaged carbon fluxes and pool sizes can, however, differ by as much as 46 %. Simulation results suggest that differences in carbon balance between the two approaches arise primarily through differences in net radiation which directly affects net primary productivity, and thus leaf area index and vegetation biomass.

  19. Increasing accuracy of dispersal kernels in grid-based population models

    USGS Publications Warehouse

    Slone, D.H.

    2011-01-01

    Dispersal kernels in grid-based population models specify the proportion, distance and direction of movements within the model landscape. Spatial errors in dispersal kernels can have large compounding effects on model accuracy. Circular Gaussian and Laplacian dispersal kernels at a range of spatial resolutions were investigated, and methods for minimizing errors caused by the discretizing process were explored. Kernels of progressively smaller sizes relative to the landscape grid size were calculated using cell-integration and cell-center methods. These kernels were convolved repeatedly, and the final distribution was compared with a reference analytical solution. For large Gaussian kernels (σ > 10 cells), the total kernel error was <10 &sup-11; compared to analytical results. Using an invasion model that tracked the time a population took to reach a defined goal, the discrete model results were comparable to the analytical reference. With Gaussian kernels that had σ ≤ 0.12 using the cell integration method, or σ ≤ 0.22 using the cell center method, the kernel error was greater than 10%, which resulted in invasion times that were orders of magnitude different than theoretical results. A goal-seeking routine was developed to adjust the kernels to minimize overall error. With this, corrections for small kernels were found that decreased overall kernel error to <10-11 and invasion time error to <5%.

  20. Collection Efficiency and Ice Accretion Characteristics of Two Full Scale and One 1/4 Scale Business Jet Horizontal Tails

    NASA Technical Reports Server (NTRS)

    Bidwell, Colin S.; Papadakis, Michael

    2005-01-01

    Collection efficiency and ice accretion calculations have been made for a series of business jet horizontal tail configurations using a three-dimensional panel code, an adaptive grid code, and the NASA Glenn LEWICE3D grid based ice accretion code. The horizontal tail models included two full scale wing tips and a 25 percent scale model. Flow solutions for the horizontal tails were generated using the PMARC panel code. Grids used in the ice accretion calculations were generated using the adaptive grid code ICEGRID. The LEWICE3D grid based ice accretion program was used to calculate impingement efficiency and ice shapes. Ice shapes typifying rime and mixed icing conditions were generated for a 30 minute hold condition. All calculations were performed on an SGI Octane computer. The results have been compared to experimental flow and impingement data. In general, the calculated flow and collection efficiencies compared well with experiment, and the ice shapes appeared representative of the rime and mixed icing conditions for which they were calculated.

  1. Air-core grid for scattered x-ray rejection

    DOEpatents

    Logan, C.M.; Lane, S.M.

    1995-10-03

    The invention is directed to a grid used in x-ray imaging applications to block scattered radiation while allowing the desired imaging radiation to pass through, and to process for making the grid. The grid is composed of glass containing lead oxide, and eliminates the spacer material used in prior known grids, and is therefore, an air-core grid. The glass is arranged in a pattern so that a large fraction of the area is open allowing the imaging radiation to pass through. A small pore size is used and the grid has a thickness chosen to provide high scatter rejection. For example, the grid may be produced with a 200 {micro}m pore size, 80% open area, and 4 mm thickness. 2 figs.

  2. Air-core grid for scattered x-ray rejection

    DOEpatents

    Logan, Clinton M.; Lane, Stephen M.

    1995-01-01

    The invention is directed to a grid used in x-ray imaging applications to block scattered radiation while allowing the desired imaging radiation to pass through, and to process for making the grid. The grid is composed of glass containing lead oxide, and eliminates the spacer material used in prior known grids, and is therefore, an air-core grid. The glass is arranged in a pattern so that a large fraction of the area is open allowing the imaging radiation to pass through. A small pore size is used and the grid has a thickness chosen to provide high scatter rejection. For example, the grid may be produced with a 200 .mu.m pore size, 80% open area, and 4 mm thickness.

  3. Finite-difference modeling with variable grid-size and adaptive time-step in porous media

    NASA Astrophysics Data System (ADS)

    Liu, Xinxin; Yin, Xingyao; Wu, Guochen

    2014-04-01

    Forward modeling of elastic wave propagation in porous media has great importance for understanding and interpreting the influences of rock properties on characteristics of seismic wavefield. However, the finite-difference forward-modeling method is usually implemented with global spatial grid-size and time-step; it consumes large amounts of computational cost when small-scaled oil/gas-bearing structures or large velocity-contrast exist underground. To overcome this handicap, combined with variable grid-size and time-step, this paper developed a staggered-grid finite-difference scheme for elastic wave modeling in porous media. Variable finite-difference coefficients and wavefield interpolation were used to realize the transition of wave propagation between regions of different grid-size. The accuracy and efficiency of the algorithm were shown by numerical examples. The proposed method is advanced with low computational cost in elastic wave simulation for heterogeneous oil/gas reservoirs.

  4. Directional kriging implementation for gridded data interpolation and comparative study with common methods

    NASA Astrophysics Data System (ADS)

    Mahmoudabadi, H.; Briggs, G.

    2016-12-01

    Gridded data sets, such as geoid models or datum shift grids, are commonly used in coordinate transformation algorithms. Grid files typically contain known or measured values at regular fixed intervals. The process of computing a value at an unknown location from the values in the grid data set is called "interpolation". Generally, interpolation methods predict a value at a given point by computing a weighted average of the known values in the neighborhood of the point. Geostatistical Kriging is a widely used interpolation method for irregular networks. Kriging interpolation first analyzes the spatial structure of the input data, then generates a general model to describe spatial dependencies. This model is used to calculate values at unsampled locations by finding direction, shape, size, and weight of neighborhood points. Because it is based on a linear formulation for the best estimation, Kriging it the optimal interpolation method in statistical terms. The Kriging interpolation algorithm produces an unbiased prediction, as well as the ability to calculate the spatial distribution of uncertainty, allowing you to estimate the errors in an interpolation for any particular point. Kriging is not widely used in geospatial applications today, especially applications that run on low power devices or deal with large data files. This is due to the computational power and memory requirements of standard Kriging techniques. In this paper, improvements are introduced in directional kriging implementation by taking advantage of the structure of the grid files. The regular spacing of points simplifies finding the neighborhood points and computing their pairwise distances, reducing the the complexity and improving the execution time of the Kriging algorithm. Also, the proposed method iteratively loads small portion of interest areas in different directions to reduce the amount of required memory. This makes the technique feasible on almost any computer processor. Comparison between kriging and other standard interpolation methods demonstrated more accurate estimations in less denser data files.

  5. Voidage correction algorithm for unresolved Euler-Lagrange simulations

    NASA Astrophysics Data System (ADS)

    Askarishahi, Maryam; Salehi, Mohammad-Sadegh; Radl, Stefan

    2018-04-01

    The effect of grid coarsening on the predicted total drag force and heat exchange rate in dense gas-particle flows is investigated using Euler-Lagrange (EL) approach. We demonstrate that grid coarsening may reduce the predicted total drag force and exchange rate. Surprisingly, exchange coefficients predicted by the EL approach deviate more significantly from the exact value compared to results of Euler-Euler (EE)-based calculations. The voidage gradient is identified as the root cause of this peculiar behavior. Consequently, we propose a correction algorithm based on a sigmoidal function to predict the voidage experienced by individual particles. Our correction algorithm can significantly improve the prediction of exchange coefficients in EL models, which is tested for simulations involving Euler grid cell sizes between 2d_p and 12d_p . It is most relevant in simulations of dense polydisperse particle suspensions featuring steep voidage profiles. For these suspensions, classical approaches may result in an error of the total exchange rate of up to 30%.

  6. Experimental investigation of a 2.5 centimeter diameter Kaufman microthruster

    NASA Technical Reports Server (NTRS)

    Cohen, A. J.

    1973-01-01

    A 2.5-centimeter-diameter Kaufman electron bombardment microthruster was fabricated and tested. The microthruster design was based on the 15-centimeter-diameter SERT 2 and 5-centimeter-diameter Lewis experimental thruster designs. The microthruster with a two-grid system, operating at a net accelerating potential of 600 volts and an accelerator potential of 500 volts, produced a calculated 445 micronewton thrust when it was run with a 9-milliampere beam current. A glass grid was initially used in testing. Later a two-grid system was successfully incorporated. Both the propellant utilization efficiency and the total power efficiency were lower than for large-size advanced thrusters, as expected; but they were sufficiently high that 2.5-centimeter thrusters show promise for future space applications. Total power of the microthruster with an assumed 7-watt hollow-cathode neutralizer was less than 30 watts at a thrust level of 445 micronewton (100 Nu LBf). The hollow cathode was operated at zero tip heater power for power requirement tests.

  7. Forward modeling magnetic fields of induced and remanent magnetization in the lithosphere using tesseroids

    NASA Astrophysics Data System (ADS)

    Baykiev, Eldar; Ebbing, Jörg; Brönner, Marco; Fabian, Karl

    2016-11-01

    A newly developed software package to calculate the magnetic field in a spherical coordinate system near the Earth's surface and on satellite height is shown to produce reliable modeling results for global and regional applications. The discretization cells of the model are uniformly magnetized spherical prisms, so called tesseroids. The presented algorithm extends an existing code for gravity calculations by applying Poisson's relation to identify the magnetic potential with the sum over pseudogravity fields of tesseroids. By testing different lithosphere discretization grids it is possible to determine the optimal size of tesseroids for field calculations on satellite altitude within realistic measurement error bounds. Also the influence of the Earth's ellipticity upon the modeling result is estimated and global examples are studied. The new software calculates induced and remanent magnetic fields for models at global and regional scale. For regional models far-field effects are evaluated and discussed. This provides bounds for the minimal size of a regional model that is necessary to predict meaningful satellite total field anomalies over the corresponding area.

  8. Smart electric vehicle (EV) charging and grid integration apparatus and methods

    DOEpatents

    Gadh, Rajit; Mal, Siddhartha; Prabhu, Shivanand; Chu, Chi-Cheng; Sheikh, Omar; Chung, Ching-Yen; He, Lei; Xiao, Bingjun; Shi, Yiyu

    2015-05-05

    An expert system manages a power grid wherein charging stations are connected to the power grid, with electric vehicles connected to the charging stations, whereby the expert system selectively backfills power from connected electric vehicles to the power grid through a grid tie inverter (if present) within the charging stations. In more traditional usage, the expert system allows for electric vehicle charging, coupled with user preferences as to charge time, charge cost, and charging station capabilities, without exceeding the power grid capacity at any point. A robust yet accurate state of charge (SOC) calculation method is also presented, whereby initially an open circuit voltage (OCV) based on sampled battery voltages and currents is calculated, and then the SOC is obtained based on a mapping between a previously measured reference OCV (ROCV) and SOC. The OCV-SOC calculation method accommodates likely any battery type with any current profile.

  9. A Composite Source Model With Fractal Subevent Size Distribution

    NASA Astrophysics Data System (ADS)

    Burjanek, J.; Zahradnik, J.

    A composite source model, incorporating different sized subevents, provides a pos- sible description of complex rupture processes during earthquakes. The number of subevents with characteristic dimension greater than R is proportional to R-2. The subevents do not overlap with each other, and the sum of their areas equals to the area of the target event (e.g. mainshock) . The subevents are distributed randomly over the fault. Each subevent is modeled as a finite source, using kinematic approach (radial rupture propagation, constant rupture velocity, boxcar slip-velocity function, with constant rise time on the subevent). The final slip at each subevent is related to its characteristic dimension, using constant stress-drop scaling. Variation of rise time with subevent size is a free parameter of modeling. The nucleation point of each subevent is taken as the point closest to mainshock hypocentre. The synthetic Green's functions are calculated by the discrete-wavenumber method in a 1D horizontally lay- ered crustal model in a relatively coarse grid of points covering the fault plane. The Green's functions needed for the kinematic model in a fine grid are obtained by cu- bic spline interpolation. As different frequencies may be efficiently calculated with different sampling, the interpolation simplifies and speeds-up the procedure signifi- cantly. The composite source model described above allows interpretation in terms of a kinematic model with non-uniform final slip and rupture velocity spatial distribu- tions. The 1994 Northridge earthquake (Mw = 6.7) is used as a validation event. The strong-ground motion modeling of the 1999 Athens earthquake (Mw = 5.9) is also performed.

  10. Modal element method for potential flow in non-uniform ducts: Combining closed form analysis with CFD

    NASA Technical Reports Server (NTRS)

    Baumeister, Kenneth J.; Baumeister, Joseph F.

    1994-01-01

    An analytical procedure is presented, called the modal element method, that combines numerical grid based algorithms with eigenfunction expansions developed by separation of variables. A modal element method is presented for solving potential flow in a channel with two-dimensional cylindrical like obstacles. The infinite computational region is divided into three subdomains; the bounded finite element domain, which is characterized by the cylindrical obstacle and the surrounding unbounded uniform channel entrance and exit domains. The velocity potential is represented approximately in the grid based domain by a finite element solution and is represented analytically by an eigenfunction expansion in the uniform semi-infinite entrance and exit domains. The calculated flow fields are in excellent agreement with exact analytical solutions. By eliminating the grid surrounding the obstacle, the modal element method reduces the numerical grid size, employs a more precise far field boundary condition, as well as giving theoretical insight to the interaction of the obstacle with the mean flow. Although the analysis focuses on a specific geometry, the formulation is general and can be applied to a variety of problems as seen by a comparison to companion theories in aeroacoustics and electromagnetics.

  11. Design and operation of grid-interactive thin-film silicon PV systems

    NASA Astrophysics Data System (ADS)

    Marion, Bill; Atmaram, Gobind; Lashway, Clin; Strachan, John W.

    Results are described from the operation of 11 thin-film amorphous silicon photovoltaic systems at three test facilities: the Florida Solar Energy Center, the New Mexico Solar Energy Institute, and Sandia National Laboratories. Commercially available modules from four US manufacturers are used in these systems, with array sizes from 133 to 750 W peak. Measured array efficiencies are from 3.1 to 4.8 percent. Except for one manufacturer, array peak power is in agreement with the calculated design ratings. For certain grid-connected systems, nonoptimal operation exists because the array peak power voltage is below the lower voltage limit of the power conditioning system. Reliability problems are found in two manufacturers' modules when shorts to ground and terminal corrosion occur. Array leakage current data are presented.

  12. Optimizing solar-cell grid geometry

    NASA Technical Reports Server (NTRS)

    Crossley, A. P.

    1969-01-01

    Trade-off analysis and mathematical expressions calculate optimum grid geometry in terms of various cell parameters. Determination of the grid geometry provides proper balance between grid resistance and cell output to optimize the energy conversion process.

  13. SU-F-T-74: Experimental Validation of Monaco Electron Monte Carlo Dose Calculation for Small Fields

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Varadhan; Way, S; Arentsen, L

    2016-06-15

    Purpose: To verify experimentally the accuracy of Monaco (Elekta) electron Monte Carlo (eMC) algorithm to calculate small field size depth doses, monitor units and isodose distributions. Methods: Beam modeling of eMC algorithm was performed for electron energies of 6, 9, 12 15 and 18 Mev for a Elekta Infinity Linac and all available ( 6, 10, 14 20 and 25 cone) applicator sizes. Electron cutouts of incrementally smaller field sizes (20, 40, 60 and 80% blocked from open cone) were fabricated. Dose calculation was performed using a grid size smaller than one-tenth of the R{sub 80–20} electron distal falloff distancemore » and number of particle histories was set at 500,000 per cm{sup 2}. Percent depth dose scans and beam profiles at dmax, d{sub 90} and d{sub 80} depths were measured for each cutout and energy with Wellhoffer (IBA) Blue Phantom{sup 2} scanning system and compared against eMC calculated doses. Results: The measured dose and output factors of incrementally reduced cutout sizes (to 3cm diameter) agreed with eMC calculated doses within ± 2.5%. The profile comparisons at dmax, d{sub 90} and d{sub 80} depths and percent depth doses at reduced field sizes agreed within 2.5% or 2mm. Conclusion: Our results indicate that the Monaco eMC algorithm can accurately predict depth doses, isodose distributions, and monitor units in homogeneous water phantom for field sizes as small as 3.0 cm diameter for energies in the 6 to 18 MeV range at 100 cm SSD. Consequently, the old rule of thumb to approximate limiting cutout size for an electron field determined by the lateral scatter equilibrium (E (MeV)/2.5 in centimeters of water) does not apply to Monaco eMC algorithm.« less

  14. Business Pattern of Distributed Energy in Electric Power System Reformation

    NASA Astrophysics Data System (ADS)

    Liang, YUE; Zhuochu, LIU; Jun, LI; Siwei, LI

    2017-05-01

    Under the trend of the electric power system revolution, the operation mode of micro power grid that including distributed power will be more diversified. User’s demand response and different strategies on electricity all have great influence on the operation of distributed power grid. This paper will not only research sensitive factors of micro power grid operation, but also analyze and calculate the cost and benefit of micro power grid operation upon different types. Then it will build a tech-economic calculation model, which applies to different types of micro power grid under the reformation of electric power system.

  15. Patched-grid calculations with the Euler and Navier-Stokes equations: Theory and applications

    NASA Technical Reports Server (NTRS)

    Rai, M. M.

    1986-01-01

    A patched-grid approach is one in which the flow region of interest is divided into subregions which are then discretized independently using existing grid generator. The equations of motion are integrated in each subregion in conjunction with patch-boundary schemes which allow proper information transfer across interfaces that separate subregions. The patched-grid approach greatly simplifies the treatment of complex geometries and also the addition of grid points to selected regions of the flow. A conservative patch-boundary condition that can be used with explicit, implicit factored and implicit relaxation schemes is described. Several example calculations that demonstrate the capabilities of the patched-grid scheme are also included.

  16. Probabilistic seismic hazard analysis (PSHA) for Ethiopia and the neighboring region

    NASA Astrophysics Data System (ADS)

    Ayele, Atalay

    2017-10-01

    Seismic hazard calculation is carried out for the Horn of Africa region (0°-20° N and 30°-50°E) based on the probabilistic seismic hazard analysis (PSHA) method. The earthquakes catalogue data obtained from different sources were compiled, homogenized to Mw magnitude scale and declustered to remove the dependent events as required by Poisson earthquake source model. The seismotectonic map of the study area that avails from recent studies is used for area sources zonation. For assessing the seismic hazard, the study area was divided into small grids of size 0.5° × 0.5°, and the hazard parameters were calculated at the center of each of these grid cells by considering contributions from all seismic sources. Peak Ground Acceleration (PGA) corresponding to 10% and 2% probability of exceedance in 50 years were calculated for all the grid points using generic rock site with Vs = 760 m/s. Obtained values vary from 0.0 to 0.18 g and 0.0-0.35 g for 475 and 2475 return periods, respectively. The corresponding contour maps showing the spatial variation of PGA values for the two return periods are presented here. Uniform hazard response spectrum (UHRS) for 10% and 2% probability of exceedance in 50 years and hazard curves for PGA and 0.2 s spectral acceleration (Sa) all at rock site are developed for the city of Addis Ababa. The hazard map of this study corresponding to the 475 return periods has already been used to update and produce the 3rd generation building code of Ethiopia.

  17. Predictions of Transient Flame Lift-Off Length With Comparison to Single-Cylinder Optical Engine Experiments

    DOE PAGES

    Senecal, P. K.; Pomraning, E.; Anders, J. W.; ...

    2014-05-28

    A state-of-the-art, grid-convergent simulation methodology was applied to three-dimensional calculations of a single-cylinder optical engine. A mesh resolution study on a sector-based version of the engine geometry further verified the RANS-based cell size recommendations previously presented by Senecal et al. (“Grid Convergent Spray Models for Internal Combustion Engine CFD Simulations,” ASME Paper No. ICEF2012-92043). Convergence of cylinder pressure, flame lift-off length, and emissions was achieved for an adaptive mesh refinement cell size of 0.35 mm. Furthermore, full geometry simulations, using mesh settings derived from the grid convergence study, resulted in excellent agreement with measurements of cylinder pressure, heat release rate,more » and NOx emissions. On the other hand, the full geometry simulations indicated that the flame lift-off length is not converged at 0.35 mm for jets not aligned with the computational mesh. Further simulations suggested that the flame lift-off lengths for both the nonaligned and aligned jets appear to be converged at 0.175 mm. With this increased mesh resolution, both the trends and magnitudes in flame lift-off length were well predicted with the current simulation methodology. Good agreement between the overall predicted flame behavior and the available chemiluminescence measurements was also achieved. Our present study indicates that cell size requirements for accurate prediction of full geometry flame lift-off lengths may be stricter than those for global combustion behavior. This may be important when accurate soot predictions are required.« less

  18. Predictions of Transient Flame Lift-Off Length With Comparison to Single-Cylinder Optical Engine Experiments

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Senecal, P. K.; Pomraning, E.; Anders, J. W.

    A state-of-the-art, grid-convergent simulation methodology was applied to three-dimensional calculations of a single-cylinder optical engine. A mesh resolution study on a sector-based version of the engine geometry further verified the RANS-based cell size recommendations previously presented by Senecal et al. (“Grid Convergent Spray Models for Internal Combustion Engine CFD Simulations,” ASME Paper No. ICEF2012-92043). Convergence of cylinder pressure, flame lift-off length, and emissions was achieved for an adaptive mesh refinement cell size of 0.35 mm. Furthermore, full geometry simulations, using mesh settings derived from the grid convergence study, resulted in excellent agreement with measurements of cylinder pressure, heat release rate,more » and NOx emissions. On the other hand, the full geometry simulations indicated that the flame lift-off length is not converged at 0.35 mm for jets not aligned with the computational mesh. Further simulations suggested that the flame lift-off lengths for both the nonaligned and aligned jets appear to be converged at 0.175 mm. With this increased mesh resolution, both the trends and magnitudes in flame lift-off length were well predicted with the current simulation methodology. Good agreement between the overall predicted flame behavior and the available chemiluminescence measurements was also achieved. Our present study indicates that cell size requirements for accurate prediction of full geometry flame lift-off lengths may be stricter than those for global combustion behavior. This may be important when accurate soot predictions are required.« less

  19. Siting and sizing of distributed generators based on improved simulated annealing particle swarm optimization.

    PubMed

    Su, Hongsheng

    2017-12-18

    Distributed power grids generally contain multiple diverse types of distributed generators (DGs). Traditional particle swarm optimization (PSO) and simulated annealing PSO (SA-PSO) algorithms have some deficiencies in site selection and capacity determination of DGs, such as slow convergence speed and easily falling into local trap. In this paper, an improved SA-PSO (ISA-PSO) algorithm is proposed by introducing crossover and mutation operators of genetic algorithm (GA) into SA-PSO, so that the capabilities of the algorithm are well embodied in global searching and local exploration. In addition, diverse types of DGs are made equivalent to four types of nodes in flow calculation by the backward or forward sweep method, and reactive power sharing principles and allocation theory are applied to determine initial reactive power value and execute subsequent correction, thus providing the algorithm a better start to speed up the convergence. Finally, a mathematical model of the minimum economic cost is established for the siting and sizing of DGs under the location and capacity uncertainties of each single DG. Its objective function considers investment and operation cost of DGs, grid loss cost, annual purchase electricity cost, and environmental pollution cost, and the constraints include power flow, bus voltage, conductor current, and DG capacity. Through applications in an IEEE33-node distributed system, it is found that the proposed method can achieve desirable economic efficiency and safer voltage level relative to traditional PSO and SA-PSO algorithms, and is a more effective planning method for the siting and sizing of DGs in distributed power grids.

  20. Spatial analyses for nonoverlapping objects with size variations and their application to coral communities.

    PubMed

    Muko, Soyoka; Shimatani, Ichiro K; Nozawa, Yoko

    2014-07-01

    Spatial distributions of individuals are conventionally analysed by representing objects as dimensionless points, in which spatial statistics are based on centre-to-centre distances. However, if organisms expand without overlapping and show size variations, such as is the case for encrusting corals, interobject spacing is crucial for spatial associations where interactions occur. We introduced new pairwise statistics using minimum distances between objects and demonstrated their utility when examining encrusting coral community data. We also calculated the conventional point process statistics and the grid-based statistics to clarify the advantages and limitations of each spatial statistical method. For simplicity, coral colonies were approximated by disks in these demonstrations. Focusing on short-distance effects, the use of minimum distances revealed that almost all coral genera were aggregated at a scale of 1-25 cm. However, when fragmented colonies (ramets) were treated as a genet, a genet-level analysis indicated weak or no aggregation, suggesting that most corals were randomly distributed and that fragmentation was the primary cause of colony aggregations. In contrast, point process statistics showed larger aggregation scales, presumably because centre-to-centre distances included both intercolony spacing and colony sizes (radius). The grid-based statistics were able to quantify the patch (aggregation) scale of colonies, but the scale was strongly affected by the colony size. Our approach quantitatively showed repulsive effects between an aggressive genus and a competitively weak genus, while the grid-based statistics (covariance function) also showed repulsion although the spatial scale indicated from the statistics was not directly interpretable in terms of ecological meaning. The use of minimum distances together with previously proposed spatial statistics helped us to extend our understanding of the spatial patterns of nonoverlapping objects that vary in size and the associated specific scales. © 2013 The Authors. Journal of Animal Ecology © 2013 British Ecological Society.

  1. Optimal Sizing Tool for Battery Storage in Grid Applications

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    2015-09-24

    The battery storage sizing tool developed at Pacific Northwest National Laboratory can be used to evaluate economic performance and determine the optimal size of battery storage in different use cases considering multiple power system applications. The considered use cases include i) utility owned battery storage, and ii) battery storage behind customer meter. The power system applications from energy storage include energy arbitrage, balancing services, T&D deferral, outage mitigation, demand charge reduction etc. Most of existing solutions consider only one or two grid services simultaneously, such as balancing service and energy arbitrage. ES-select developed by Sandia and KEMA is able tomore » consider multiple grid services but it stacks the grid services based on priorities instead of co-optimization. This tool is the first one that provides a co-optimization for systematic and local grid services.« less

  2. Uncertainty in coal property valuation in West Virginia: A case study

    USGS Publications Warehouse

    Hohn, M.E.; McDowell, R.R.

    2001-01-01

    Interpolated grids of coal bed thickness are being considered for use in a proposed method for taxation of coal in the state of West Virginia (United States). To assess the origin and magnitude of possible inaccuracies in calculated coal tonnage, we used conditional simulation to generate equiprobable realizations of net coal thickness for two coals on a 7 1/2 min topographic quadrangle, and a third coal in a second quadrangle. Coals differed in average thickness and proportion of original coal that had been removed by erosion; all three coals crop out in the study area. Coal tonnage was calculated for each realization and for each interpolated grid for actual and artificial property parcels, and differences were summarized as graphs of percent difference between tonnage calculated from the grid and average tonnage from simulations. Coal in individual parcels was considered minable for valuation purposes if average thickness in each parcel exceeded 30 inches. Results of this study show that over 75% of the parcels are classified correctly as minable or unminable based on interpolation grids of coal bed thickness. Although between 80 and 90% of the tonnages differ by less than 20% between interpolated values and simulated values, a nonlinear conditional bias might exist in estimation of coal tonnage from interpolated thickness, such that tonnage is underestimated where coal is thin, and overestimated where coal is thick. The largest percent differences occur for parcels that are small in area, although because of the small quantities of coal in question, bias is small on an absolute scale for these parcels. For a given parcel size, maximum apparent overestimation of coal tonnage occurs in parcels with an average coal bed thickness near the minable cutoff of 30 in. Conditional bias in tonnage for parcels having a coal thickness exceeding the cutoff by 10 in. or more is constant for two of the three coals studied, and increases slightly with average thickness for the third coal. ?? 2001 International Association for Mathematical Geology.

  3. Pathloss Calculation Using the Transmission Line Matrix and Finite Difference Time Domain Methods With Coarse Grids

    DOE PAGES

    Nutaro, James; Kuruganti, Teja

    2017-02-24

    Numerical simulations of the wave equation that are intended to provide accurate time domain solutions require a computational mesh with grid points separated by a distance less than the wavelength of the source term and initial data. However, calculations of radio signal pathloss generally do not require accurate time domain solutions. This paper describes an approach for calculating pathloss by using the finite difference time domain and transmission line matrix models of wave propagation on a grid with points separated by distances much greater than the signal wavelength. The calculated pathloss can be kept close to the true value formore » freespace propagation with an appropriate selection of initial conditions. This method can also simulate diffraction with an error governed by the ratio of the signal wavelength to the grid spacing.« less

  4. 10 CFR 950.25 - Calculation of covered costs.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... by the Intercontinental Exchange (ICE) or an alternate electronic marketplace deemed reliable by the... to the grid for calculating payments to the Nuclear Waste Fund to the nameplate capacity for each... generation to the grid for calculating payments to the Nuclear Waste Fund to the nameplate capacity for each...

  5. 10 CFR 950.25 - Calculation of covered costs.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... by the Intercontinental Exchange (ICE) or an alternate electronic marketplace deemed reliable by the... to the grid for calculating payments to the Nuclear Waste Fund to the nameplate capacity for each... generation to the grid for calculating payments to the Nuclear Waste Fund to the nameplate capacity for each...

  6. 10 CFR 950.25 - Calculation of covered costs.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... by the Intercontinental Exchange (ICE) or an alternate electronic marketplace deemed reliable by the... to the grid for calculating payments to the Nuclear Waste Fund to the nameplate capacity for each... generation to the grid for calculating payments to the Nuclear Waste Fund to the nameplate capacity for each...

  7. 10 CFR 950.25 - Calculation of covered costs.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... by the Intercontinental Exchange (ICE) or an alternate electronic marketplace deemed reliable by the... to the grid for calculating payments to the Nuclear Waste Fund to the nameplate capacity for each... generation to the grid for calculating payments to the Nuclear Waste Fund to the nameplate capacity for each...

  8. 10 CFR 950.25 - Calculation of covered costs.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... by the Intercontinental Exchange (ICE) or an alternate electronic marketplace deemed reliable by the... to the grid for calculating payments to the Nuclear Waste Fund to the nameplate capacity for each... generation to the grid for calculating payments to the Nuclear Waste Fund to the nameplate capacity for each...

  9. Stability Test for Transient-Temperature Calculations

    NASA Technical Reports Server (NTRS)

    Campbell, W.

    1984-01-01

    Graphical test helps assure numerical stability of calculations of transient temperature or diffusion in composite medium. Rectangular grid forms basis of two-dimensional finite-difference model for heat conduction or other diffusion like phenomena. Model enables calculation of transient heat transfer among up to four different materials that meet at grid point.

  10. Meteorological modeling of arrival and deposition of fallout at intermediate distances downwind of the Nevada Test Site.

    PubMed

    Cederwall, R T; Peterson, K R

    1990-11-01

    A three-dimensional atmospheric transport and diffusion model is used to calculate the arrival and deposition of fallout from 13 selected nuclear tests at the Nevada Test Site (NTS) in the 1950s. Results are used to extend NTS fallout patterns to intermediate downwind distances (300 to 1200 km). The radioactive cloud is represented in the model by a population of Lagrangian marker particles, with concentrations calculated on an Eulerian grid. Use of marker particles, with fall velocities dependent on particle size, provides a realistic simulation of fallout as the debris cloud travels downwind. The three-dimensional wind field is derived from observed data, adjusted for mass consistency. Terrain is represented in the grid, which extends up to 1200 km downwind of NTS and has 32-km horizontal resolution and 1-km vertical resolution. Ground deposition is calculated by a deposition-velocity approach. Source terms and relationships between deposition and exposure rate are based on work by Hicks. Uncertainty in particle size and vertical distributions within the debris cloud (and stem) allow for some model "tuning" to better match measured ground-deposition values. Particle trajectories representing different sizes and starting heights above ground zero are used to guide source specification. An hourly time history of the modeled fallout pattern as the debris cloud moves downwind provides estimates of fallout arrival times. Results for event HARRY illustrate the methodology. The composite deposition pattern for all 13 tests is characterized by two lobes extending out to the north-northeast and east-northeast, respectively, at intermediate distances from NTS. Arrival estimates, along with modeled deposition values, augment measured deposition data in the development of data bases at the county level; these data bases are used for estimating radiation exposure at intermediate distances downwind of NTS. Results from a study of event TRINITY are also presented.

  11. Multidimensional radiative transfer with multilevel atoms. II. The non-linear multigrid method.

    NASA Astrophysics Data System (ADS)

    Fabiani Bendicho, P.; Trujillo Bueno, J.; Auer, L.

    1997-08-01

    A new iterative method for solving non-LTE multilevel radiative transfer (RT) problems in 1D, 2D or 3D geometries is presented. The scheme obtains the self-consistent solution of the kinetic and RT equations at the cost of only a few (<10) formal solutions of the RT equation. It combines, for the first time, non-linear multigrid iteration (Brandt, 1977, Math. Comp. 31, 333; Hackbush, 1985, Multi-Grid Methods and Applications, springer-Verlag, Berlin), an efficient multilevel RT scheme based on Gauss-Seidel iterations (cf. Trujillo Bueno & Fabiani Bendicho, 1995ApJ...455..646T), and accurate short-characteristics formal solution techniques. By combining a valid stopping criterion with a nested-grid strategy a converged solution with the desired true error is automatically guaranteed. Contrary to the current operator splitting methods the very high convergence speed of the new RT method does not deteriorate when the grid spatial resolution is increased. With this non-linear multigrid method non-LTE problems discretized on N grid points are solved in O(N) operations. The nested multigrid RT method presented here is, thus, particularly attractive in complicated multilevel transfer problems where small grid-sizes are required. The properties of the method are analyzed both analytically and with illustrative multilevel calculations for Ca II in 1D and 2D schematic model atmospheres.

  12. Navier-Stokes simulation of rotor-body flowfield in hover using overset grids

    NASA Technical Reports Server (NTRS)

    Srinivasan, G. R.; Ahmad, J. U.

    1993-01-01

    A free-wake Navier-Stokes numerical scheme and multiple Chimera overset grids have been utilized for calculating the quasi-steady hovering flowfield of a Boeing-360 rotor mounted on an axisymmetric whirl-tower. The entire geometry of this rotor-body configuration is gridded-up with eleven different overset grids. The composite grid has 1.3 million grid points for the entire flow domain. The numerical results, obtained using coarse grids and a rigid rotor assumption, show a thrust value that is within 5% of the experimental value at a flow condition of M(sub tip) = 0.63, Theta(sub c) = 8 deg, and Re = 2.5 x 10(exp 6). The numerical method thus demonstrates the feasibility of using a multi-block scheme for calculating the flowfields of complex configurations consisting of rotating and non-rotating components.

  13. Control of nanoparticle size and amount by using the mesh grid and applying DC-bias to the substrate in silane ICP-CVD process

    NASA Astrophysics Data System (ADS)

    Yoo, Seung-Wan; Hwang, Nong-Moon; You, Shin-Jae; Kim, Jung-Hyung; Seong, Dae-Jin

    2017-11-01

    The effect of applying a bias to the substrate on the size and amount of charged crystalline silicon nanoparticles deposited on the substrate was investigated in the inductively coupled plasma chemical vapor deposition process. By inserting the grounded grid with meshes above the substrate, the region just above the substrate was separated from the plasma. Thereby, crystalline Si nanoparticles formed by the gas-phase reaction in the plasma could be deposited directly on the substrate, successfully avoiding the formation of a film. Moreover, the size and the amount of deposited nanoparticles could be changed by applying direct current bias to the substrate. When the grid of 1 × 1-mm-sized mesh was used, the nanoparticle flux was increased as the negative substrate bias increased from 0 to - 50 V. On the other hand, when a positive bias was applied to the substrate, Si nanoparticles were not deposited at all. Regardless of substrate bias voltages, the most frequently observed nanoparticles synthesized with the grid of 1 × 1-mm-sized mesh had the size range of 10-12 nm in common. When the square mesh grid of 2-mm size was used, as the substrate bias was increased from - 50 to 50 V, the size of the nanoparticles observed most frequently increased from the range of 8-10 to 40-45 nm but the amount that was deposited on the substrate decreased.

  14. Deep Part Load Flow Analysis in a Francis Model turbine by means of two-phase unsteady flow simulations

    NASA Astrophysics Data System (ADS)

    Conrad, Philipp; Weber, Wilhelm; Jung, Alexander

    2017-04-01

    Hydropower plants are indispensable to stabilize the grid by reacting quickly to changes of the energy demand. However, an extension of the operating range towards high and deep part load conditions without fatigue of the hydraulic components is desirable to increase their flexibility. In this paper a model sized Francis turbine at low discharge operating conditions (Q/QBEP = 0.27) is analyzed by means of computational fluid dynamics (CFD). Unsteady two-phase simulations for two Thoma-number conditions are conducted. Stochastic pressure oscillations, observed on the test rig at low discharge, require sophisticated numerical models together with small time steps, large grid sizes and long simulation times to cope with these fluctuations. In this paper the BSL-EARSM model (Explicit Algebraic Reynolds Stress) was applied as a compromise between scale resolving and two-equation turbulence models with respect to computational effort and accuracy. Simulation results are compared to pressure measurements showing reasonable agreement in resolving the frequency spectra and amplitude. Inner blade vortices were predicted successfully in shape and size. Surface streamlines in blade-to-blade view are presented, giving insights to the formation of the inner blade vortices. The acquired time dependent pressure fields can be used for quasi-static structural analysis (FEA) for fatigue calculations in the future.

  15. Boundary element method for normal non-adhesive and adhesive contacts of power-law graded elastic materials

    NASA Astrophysics Data System (ADS)

    Li, Qiang; Popov, Valentin L.

    2018-03-01

    Recently proposed formulation of the boundary element method for adhesive contacts has been generalized for contacts of power-law graded materials with and without adhesion. Proceeding from the fundamental solution for single force acting on the surface of an elastic half space, first the influence matrix is obtained for a rectangular grid. The inverse problem for the calculation of required stress in the contact area from a known surface displacement is solved using the conjugate-gradient technique. For the transformation between the stresses and displacements, the Fast Fourier Transformation is used. For the adhesive contact of graded material, the detachment criterion based on the energy balance is proposed. The method is validated by comparison with known exact analytical solutions as well as by proving the independence of the mesh size and the grid orientation.

  16. Fast calculation method of computer-generated hologram using a depth camera with point cloud gridding

    NASA Astrophysics Data System (ADS)

    Zhao, Yu; Shi, Chen-Xiao; Kwon, Ki-Chul; Piao, Yan-Ling; Piao, Mei-Lan; Kim, Nam

    2018-03-01

    We propose a fast calculation method for a computer-generated hologram (CGH) of real objects that uses a point cloud gridding method. The depth information of the scene is acquired using a depth camera and the point cloud model is reconstructed virtually. Because each point of the point cloud is distributed precisely to the exact coordinates of each layer, each point of the point cloud can be classified into grids according to its depth. A diffraction calculation is performed on the grids using a fast Fourier transform (FFT) to obtain a CGH. The computational complexity is reduced dramatically in comparison with conventional methods. The feasibility of the proposed method was confirmed by numerical and optical experiments.

  17. X-ray photon correlation spectroscopy using a fast pixel array detector with a grid mask resolution enhancer.

    PubMed

    Hoshino, Taiki; Kikuchi, Moriya; Murakami, Daiki; Harada, Yoshiko; Mitamura, Koji; Ito, Kiminori; Tanaka, Yoshihito; Sasaki, Sono; Takata, Masaki; Jinnai, Hiroshi; Takahara, Atsushi

    2012-11-01

    The performance of a fast pixel array detector with a grid mask resolution enhancer has been demonstrated for X-ray photon correlation spectroscopy (XPCS) measurements to investigate fast dynamics on a microscopic scale. A detecting system, in which each pixel of a single-photon-counting pixel array detector, PILATUS, is covered by grid mask apertures, was constructed for XPCS measurements of silica nanoparticles in polymer melts. The experimental results are confirmed to be consistent by comparison with other independent experiments. By applying this method, XPCS measurements can be carried out by customizing the hole size of the grid mask to suit the experimental conditions, such as beam size, detector size and sample-to-detector distance.

  18. Propagation of Disturbances in AC Electricity Grids.

    PubMed

    Tamrakar, Samyak; Conrath, Michael; Kettemann, Stefan

    2018-04-24

    The energy transition towards high shares of renewable energy will affect the stability of electricity grids in many ways. Here, we aim to study its impact on propagation of disturbances by solving nonlinear swing equations describing coupled rotating masses of synchronous generators and motors on different grid topologies. We consider a tree, a square grid and as a real grid topology, the german transmission grid. We identify ranges of parameters with different transient dynamics: the disturbance decays exponentially in time, superimposed by oscillations with the fast decay rate of a single node, or with a smaller decay rate without oscillations. Most remarkably, as the grid inertia is lowered, nodes may become correlated, slowing down the propagation from ballistic to diffusive motion, decaying with a power law in time. Applying linear response theory we show that tree grids have a spectral gap leading to exponential relaxation as protected by topology and independent on grid size. Meshed grids are found to have a spectral gap which decreases with increasing grid size, leading to slow power law relaxation and collective diffusive propagation of disturbances. We conclude by discussing consequences if no measures are undertaken to preserve the grid inertia in the energy transition.

  19. CFL3D, FUN3d, and NSU3D Contributions to the Fifth Drag Prediction Workshop

    NASA Technical Reports Server (NTRS)

    Park, Michael A.; Laflin, Kelly R.; Chaffin, Mark S.; Powell, Nicholas; Levy, David W.

    2013-01-01

    Results presented at the Fifth Drag Prediction Workshop using CFL3D, FUN3D, and NSU3D are described. These are calculations on the workshop provided grids and drag adapted grids. The NSU3D results have been updated to reflect an improvement to skin friction calculation on skewed grids. FUN3D results generated after the workshop are included for custom participant generated grids and a grid from a previous workshop. Uniform grid refinement at the design condition shows a tight grouping in calculated drag, where the variation in the pressure component of drag is larger than the skin friction component. At this design condition, A fine-grid drag value was predicted with a smaller drag adjoint adapted grid via tetrahedral adaption to a metric and mixed-element subdivision. The buffet study produced larger variation than the design case, which is attributed to large differences in the predicted side-of-body separation extent. Various modeling and discretization approaches had a strong impact on predicted side-of-body separation. This large wing root separation bubble was not observed in wind tunnel tests indicating that more work is necessary in modeling wing root juncture flows to predict experiments.

  20. Convergance experiments with a hydrodynamic model of Port Royal Sound, South Carolina

    USGS Publications Warehouse

    Lee, J.K.; Schaffranek, R.W.; Baltzer, R.A.

    1989-01-01

    A two-demensional, depth-averaged, finite-difference, flow/transport model, SIM2D, is being used to simulate tidal circulation and transport in the Port Royal Sound, South Carolina, estuarine system. Models of a subregion of the Port Royal Sound system have been derived from an earlier-developed model of the entire system having a grid size of 600 ft. The submodels were implemented with grid sizes of 600, 300, and 150 ft in order to determine the effects of changes in grid size on computed flows in the subregion, which is characterized by narrow channels and extensive tidal flats that flood and dewater with each rise and fall of the tide. Tidal amplitudes changes less than 5 percent as the grid size was decreased. Simulations were performed with the 300-foot submodel for time steps of 60, 30, and 15 s. Study results are discussed.

  1. Error Estimates of the Ares I Computed Turbulent Ascent Longitudinal Aerodynamic Analysis

    NASA Technical Reports Server (NTRS)

    Abdol-Hamid, Khaled S.; Ghaffari, Farhad

    2012-01-01

    Numerical predictions of the longitudinal aerodynamic characteristics for the Ares I class of vehicles, along with the associated error estimate derived from an iterative convergence grid refinement, are presented. Computational results are based on an unstructured grid, Reynolds-averaged Navier-Stokes analysis. The validity of the approach to compute the associated error estimates, derived from a base grid to an extrapolated infinite-size grid, was first demonstrated on a sub-scaled wind tunnel model at representative ascent flow conditions for which the experimental data existed. Such analysis at the transonic flow conditions revealed a maximum deviation of about 23% between the computed longitudinal aerodynamic coefficients with the base grid and the measured data across the entire roll angles. This maximum deviation from the wind tunnel data was associated with the computed normal force coefficient at the transonic flow condition and was reduced to approximately 16% based on the infinite-size grid. However, all the computed aerodynamic coefficients with the base grid at the supersonic flow conditions showed a maximum deviation of only about 8% with that level being improved to approximately 5% for the infinite-size grid. The results and the error estimates based on the established procedure are also presented for the flight flow conditions.

  2. Laser-induced superhydrophobic grid patterns on PDMS for droplet arrays formation

    NASA Astrophysics Data System (ADS)

    Farshchian, Bahador; Gatabi, Javad R.; Bernick, Steven M.; Park, Sooyeon; Lee, Gwan-Hyoung; Droopad, Ravindranath; Kim, Namwon

    2017-02-01

    We demonstrate a facile single step laser treatment process to render a polydimethylsiloxane (PDMS) surface superhydrophobic. By synchronizing a pulsed nanosecond laser source with a motorized stage, superhydrophobic grid patterns were written on the surface of PDMS. Hierarchical micro and nanostructures were formed in the irradiated areas while non-irradiated areas were covered by nanostructures due to deposition of ablated particles. Arrays of droplets form spontaneously on the laser-patterned PDMS with superhydrophobic grid pattern when the PDMS sample is simply immersed in and withdrawn from water due to different wetting properties of the irradiated and non-irradiated areas. The effects of withdrawal speed and pitch size of superhydrophobic grid on the size of formed droplets were investigated experimentally. The droplet size increases initially with increasing the withdrawal speed and then does not change significantly beyond certain points. Moreover, larger droplets are formed by increasing the pitch size of the superhydrophobic grid. The droplet arrays formed on the laser-patterned PDMS with wettability contrast can be used potentially for patterning of particles, chemicals, and bio-molecules and also for cell screening applications.

  3. Numerical generation of two-dimensional grids by the use of Poisson equations with grid control at boundaries

    NASA Technical Reports Server (NTRS)

    Sorenson, R. L.; Steger, J. L.

    1980-01-01

    A method for generating boundary-fitted, curvilinear, two dimensional grids by the use of the Poisson equations is presented. Grids of C-type and O-type were made about airfoils and other shapes, with circular, rectangular, cascade-type, and other outer boundary shapes. Both viscous and inviscid spacings were used. In all cases, two important types of grid control can be exercised at both inner and outer boundaries. First is arbitrary control of the distances between the boundaries and the adjacent lines of the same coordinate family, i.e., stand-off distances. Second is arbitrary control of the angles with which lines of the opposite coordinate family intersect the boundaries. Thus, both grid cell size (or aspect ratio) and grid cell skewness are controlled at boundaries. Reasonable cell size and shape are ensured even in cases wherein extreme boundary shapes would tend to cause skewness or poorly controlled grid spacing. An inherent feature of the Poisson equations is that lines in the interior of the grid smoothly connect the boundary points (the grid mapping functions are second order differentiable).

  4. Multiscale image processing and antiscatter grids in digital radiography.

    PubMed

    Lo, Winnie Y; Hornof, William J; Zwingenberger, Allison L; Robertson, Ian D

    2009-01-01

    Scatter radiation is a source of noise and results in decreased signal-to-noise ratio and thus decreased image quality in digital radiography. We determined subjectively whether a digitally processed image made without a grid would be of similar quality to an image made with a grid but without image processing. Additionally the effects of exposure dose and of a using a grid with digital radiography on overall image quality were studied. Thoracic and abdominal radiographs of five dogs of various sizes were made. Four acquisition techniques were included (1) with a grid, standard exposure dose, digital image processing; (2) without a grid, standard exposure dose, digital image processing; (3) without a grid, half the exposure dose, digital image processing; and (4) with a grid, standard exposure dose, no digital image processing (to mimic a film-screen radiograph). Full-size radiographs as well as magnified images of specific anatomic regions were generated. Nine reviewers rated the overall image quality subjectively using a five-point scale. All digitally processed radiographs had higher overall scores than nondigitally processed radiographs regardless of patient size, exposure dose, or use of a grid. The images made at half the exposure dose had a slightly lower quality than those made at full dose, but this was only statistically significant in magnified images. Using a grid with digital image processing led to a slight but statistically significant increase in overall quality when compared with digitally processed images made without a grid but whether this increase in quality is clinically significant is unknown.

  5. Deterministic seismic hazard macrozonation of India

    NASA Astrophysics Data System (ADS)

    Kolathayar, Sreevalsa; Sitharam, T. G.; Vipin, K. S.

    2012-10-01

    Earthquakes are known to have occurred in Indian subcontinent from ancient times. This paper presents the results of seismic hazard analysis of India (6°-38°N and 68°-98°E) based on the deterministic approach using latest seismicity data (up to 2010). The hazard analysis was done using two different source models (linear sources and point sources) and 12 well recognized attenuation relations considering varied tectonic provinces in the region. The earthquake data obtained from different sources were homogenized and declustered and a total of 27,146 earthquakes of moment magnitude 4 and above were listed in the study area. The sesismotectonic map of the study area was prepared by considering the faults, lineaments and the shear zones which are associated with earthquakes of magnitude 4 and above. A new program was developed in MATLAB for smoothing of the point sources. For assessing the seismic hazard, the study area was divided into small grids of size 0.1° × 0.1° (approximately 10 × 10 km), and the hazard parameters were calculated at the center of each of these grid cells by considering all the seismic sources within a radius of 300 to 400 km. Rock level peak horizontal acceleration (PHA) and spectral accelerations for periods 0.1 and 1 s have been calculated for all the grid points with a deterministic approach using a code written in MATLAB. Epistemic uncertainty in hazard definition has been tackled within a logic-tree framework considering two types of sources and three attenuation models for each grid point. The hazard evaluation without logic tree approach also has been done for comparison of the results. The contour maps showing the spatial variation of hazard values are presented in the paper.

  6. ADAPTIVE TETRAHEDRAL GRID REFINEMENT AND COARSENING IN MESSAGE-PASSING ENVIRONMENTS

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hallberg, J.; Stagg, A.

    2000-10-01

    A grid refinement and coarsening scheme has been developed for tetrahedral and triangular grid-based calculations in message-passing environments. The element adaption scheme is based on an edge bisection of elements marked for refinement by an appropriate error indicator. Hash-table/linked-list data structures are used to store nodal and element formation. The grid along inter-processor boundaries is refined and coarsened consistently with the update of these data structures via MPI calls. The parallel adaption scheme has been applied to the solution of a transient, three-dimensional, nonlinear, groundwater flow problem. Timings indicate efficiency of the grid refinement process relative to the flow solvermore » calculations.« less

  7. Search for β2 Adrenergic Receptor Ligands by Virtual Screening via Grid Computing and Investigation of Binding Modes by Docking and Molecular Dynamics Simulations

    PubMed Central

    Bai, Qifeng; Shao, Yonghua; Pan, Dabo; Zhang, Yang; Liu, Huanxiang; Yao, Xiaojun

    2014-01-01

    We designed a program called MolGridCal that can be used to screen small molecule database in grid computing on basis of JPPF grid environment. Based on MolGridCal program, we proposed an integrated strategy for virtual screening and binding mode investigation by combining molecular docking, molecular dynamics (MD) simulations and free energy calculations. To test the effectiveness of MolGridCal, we screened potential ligands for β2 adrenergic receptor (β2AR) from a database containing 50,000 small molecules. MolGridCal can not only send tasks to the grid server automatically, but also can distribute tasks using the screensaver function. As for the results of virtual screening, the known agonist BI-167107 of β2AR is ranked among the top 2% of the screened candidates, indicating MolGridCal program can give reasonable results. To further study the binding mode and refine the results of MolGridCal, more accurate docking and scoring methods are used to estimate the binding affinity for the top three molecules (agonist BI-167107, neutral antagonist alprenolol and inverse agonist ICI 118,551). The results indicate agonist BI-167107 has the best binding affinity. MD simulation and free energy calculation are employed to investigate the dynamic interaction mechanism between the ligands and β2AR. The results show that the agonist BI-167107 also has the lowest binding free energy. This study can provide a new way to perform virtual screening effectively through integrating molecular docking based on grid computing, MD simulations and free energy calculations. The source codes of MolGridCal are freely available at http://molgridcal.codeplex.com. PMID:25229694

  8. Grid generation in three dimensions by Poisson equations with control of cell size and skewness at boundary surfaces

    NASA Technical Reports Server (NTRS)

    Sorenson, R. L.; Steger, J. L.

    1983-01-01

    An algorithm for generating computational grids about arbitrary three-dimensional bodies is developed. The elliptic partial differential equation (PDE) approach developed by Steger and Sorenson and used in the NASA computer program GRAPE is extended from two to three dimensions. Forcing functions which are found automatically by the algorithm give the user the ability to control mesh cell size and skewness at boundary surfaces. This algorithm, as is typical of PDE grid generators, gives smooth grid lines and spacing in the interior of the grid. The method is applied to a rectilinear wind-tunnel case and to two body shapes in spherical coordinates.

  9. Decay of grid turbulence in superfluid helium-4: Mesh dependence

    NASA Astrophysics Data System (ADS)

    Yang, J.; Ihas, G. G.

    2018-03-01

    Temporal decay of grid turbulence is experimentally studied in superfluid 4He in a large square channel. The second sound attenuation method is used to measure the turbulent vortex line density (L) with a phase locked tracking technique to minimize frequency shift effects induced by temperature fluctuations. Two different grids (0.8 mm and 3.0 mm mesh) are pulled to generate turbulence. Different power laws for decaying behavior are predicted by a theory. According to this theory, L should decay as t‑11/10 when the length scale of energy containing eddies grows from the grid mesh size to the size of the channel. At later time, after the energy containing eddy size becomes comparable to the channel, L should follow t‑3/2. Our recent experimental data exhibit evidence for t‑11/10 during the early time and t‑2 instead of t‑3/2 for later time. Moreover, a consistent bump/plateau feature is prominent between the two decay regimes for smaller (0.8 mm) grid mesh holes but absent with a grid mesh hole of 3.0 mm. This implies that in the large channel different types of turbulence are generated, depending on mesh hole size (mesh Reynolds number) compared to channel Reynolds number.

  10. Method of assembly of molecular-sized nets and scaffolding

    DOEpatents

    Michl, Josef; Magnera, Thomas F.; David, Donald E.; Harrison, Robin M.

    1999-01-01

    The present invention relates to methods and starting materials for forming molecular-sized grids or nets, or other structures based on such grids and nets, by creating molecular links between elementary molecular modules constrained to move in only two directions on an interface or surface by adhesion or bonding to that interface or surface. In the methods of this invention, monomers are employed as the building blocks of grids and more complex structures. Monomers are introduced onto and allowed to adhere or bond to an interface. The connector groups of adjacent adhered monomers are then polymerized with each other to form a regular grid in two dimensions above the interface. Modules that are not bound or adhered to the interface are removed prior to reaction of the connector groups to avoid undesired three-dimensional cross-linking and the formation of non-grid structures. Grids formed by the methods of this invention are useful in a variety of applications, including among others, for separations technology, as masks for forming regular surface structures (i.e., metal deposition) and as templates for three-dimensional molecular-sized structures.

  11. Method of assembly of molecular-sized nets and scaffolding

    DOEpatents

    Michl, J.; Magnera, T.F.; David, D.E.; Harrison, R.M.

    1999-03-02

    The present invention relates to methods and starting materials for forming molecular-sized grids or nets, or other structures based on such grids and nets, by creating molecular links between elementary molecular modules constrained to move in only two directions on an interface or surface by adhesion or bonding to that interface or surface. In the methods of this invention, monomers are employed as the building blocks of grids and more complex structures. Monomers are introduced onto and allowed to adhere or bond to an interface. The connector groups of adjacent adhered monomers are then polymerized with each other to form a regular grid in two dimensions above the interface. Modules that are not bound or adhered to the interface are removed prior to reaction of the connector groups to avoid undesired three-dimensional cross-linking and the formation of non-grid structures. Grids formed by the methods of this invention are useful in a variety of applications, including among others, for separations technology, as masks for forming regular surface structures (i.e., metal deposition) and as templates for three-dimensional molecular-sized structures. 9 figs.

  12. Icing Branch Current Research Activities in Icing Physics

    NASA Technical Reports Server (NTRS)

    Vargas, Mario

    2009-01-01

    Current development: A grid block transformation scheme which allows the input of grids in arbitrary reference frames, the use of mirror planes, and grids with relative velocities has been developed. A simple ice crystal and sand particle bouncing scheme has been included. Added an SLD splashing model based on that developed by William Wright for the LEWICE 3.2.2 software. A new area based collection efficiency algorithm will be incorporated which calculates trajectories from inflow block boundaries to outflow block boundaries. This method will be used for calculating and passing collection efficiency data between blade rows for turbo-machinery calculations.

  13. Decomposition of the Seismic Source Using Numerical Simulations and Observations of Nuclear Explosions

    DTIC Science & Technology

    2017-05-31

    SUBJECT TERMS nonlinear finite element calculations, nuclear explosion monitoring, topography 16. SECURITY CLASSIFICATION OF: 17. LIMITATION OF ABSTRACT 18...3D North Korea calculations........ Figure 6. The CRAM 3D finite element outer grid (left) is rectangular......................... Figure 7. Stress...Figure 6. The CRAM 3D finite element outer grid (left) is rectangular. The inner grid (center) is shaped to match the shape of the explosion shock wave

  14. Large-scale ground motion simulation using GPGPU

    NASA Astrophysics Data System (ADS)

    Aoi, S.; Maeda, T.; Nishizawa, N.; Aoki, T.

    2012-12-01

    Huge computation resources are required to perform large-scale ground motion simulations using 3-D finite difference method (FDM) for realistic and complex models with high accuracy. Furthermore, thousands of various simulations are necessary to evaluate the variability of the assessment caused by uncertainty of the assumptions of the source models for future earthquakes. To conquer the problem of restricted computational resources, we introduced the use of GPGPU (General purpose computing on graphics processing units) which is the technique of using a GPU as an accelerator of the computation which has been traditionally conducted by the CPU. We employed the CPU version of GMS (Ground motion Simulator; Aoi et al., 2004) as the original code and implemented the function for GPU calculation using CUDA (Compute Unified Device Architecture). GMS is a total system for seismic wave propagation simulation based on 3-D FDM scheme using discontinuous grids (Aoi&Fujiwara, 1999), which includes the solver as well as the preprocessor tools (parameter generation tool) and postprocessor tools (filter tool, visualization tool, and so on). The computational model is decomposed in two horizontal directions and each decomposed model is allocated to a different GPU. We evaluated the performance of our newly developed GPU version of GMS on the TSUBAME2.0 which is one of the Japanese fastest supercomputer operated by the Tokyo Institute of Technology. First we have performed a strong scaling test using the model with about 22 million grids and achieved 3.2 and 7.3 times of the speed-up by using 4 and 16 GPUs. Next, we have examined a weak scaling test where the model sizes (number of grids) are increased in proportion to the degree of parallelism (number of GPUs). The result showed almost perfect linearity up to the simulation with 22 billion grids using 1024 GPUs where the calculation speed reached to 79.7 TFlops and about 34 times faster than the CPU calculation using the same number of cores. Finally, we applied GPU calculation to the simulation of the 2011 Tohoku-oki earthquake. The model was constructed using a slip model from inversion of strong motion data (Suzuki et al., 2012), and a geological- and geophysical-based velocity structure model comprising all the Tohoku and Kanto regions as well as the large source area, which consists of about 1.9 billion grids. The overall characteristics of observed velocity seismograms for a longer period than range of 8 s were successfully reproduced (Maeda et al., 2012 AGU meeting). The turn around time for 50 thousand-step calculation (which correspond to 416 s in seismograph) using 100 GPUs was 52 minutes which is fairly short, especially considering this is the performance for the realistic and complex model.

  15. Computational Modeling of the Ames 11-Ft Transonic Wind Tunnel in Conjunction with IofNEWT

    NASA Technical Reports Server (NTRS)

    Djomehri, M. Jahed; Buning, Pieter G.; Erickson, Larry L.; George, Michael W. (Technical Monitor)

    1995-01-01

    Technical advances in Computational Fluid Dynamics have now made it possible to simulate complex three-dimensional internal flows about models of various size placed in a Transonic Wind Tunnel. TWT wall interference effects have been a source of error in predicting flight data from actual wind tunnel measured data. An advantage of such internal CFD calculations is to directly compare numerical results with the actual tunnel data for code assessment and tunnel flow analysis. A CFD capability has recently been devised for flow analysis of the NASA/Ames 11-Ft TWT facility. The primary objectives of this work are to provide a CFD tool to study the NASA/Ames 11-Ft TWT flow characteristics, to understand the slotted wall interference effects, and to validate CFD codes. A secondary objective is to integrate the internal flowfield calculations with the Pressure Sensitive Paint data, a surface pressure distribution capability in Ames' production wind tunnels. The effort has been part of the Ames IofNEWT, Integration of Numerical and Experimental Wind Tunnels project, which is aimed at providing further analytical tools for industrial application. We used the NASA/Ames OVERFLOW code to solve the thin-layer Navier-Stokes equations. Viscosity effects near the model are captured by Baldwin-Lomax or Baldwin-Barth turbulence models. The solver was modified to model the flow behavior in the vicinity of the tunnel longitudinal slotted walls. A suitable porous type wall boundary condition was coded to account for the cross-flow through the test section. Viscous flow equations were solved in generalized coordinates with a three-factor implicit central difference scheme in conjunction with the Chimera grid procedure. The internal flow field about the model and the tunnel walls were descretized by the Chimera overset grid system. This approach allows the application of efficient grid generation codes about individual components of the configuration; separate minor grids were developed to resolve the model and overset onto a main grid which discretizes the interior of the tunnel test section. Individual grid components axe not required to have mesh boundaries joined in any special way to each other or to the main tunnel grid. Programs have been developed to rotate the model about the tunnel pivot point and rotation axis, similar to that of the tunnel turntable mechanism for adjusting the pitch of the physical model in the test section.

  16. Multipinhole SPECT helical scan parameters and imaging volume

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Yao, Rutao, E-mail: rutaoyao@buffalo.edu; Deng, Xiao; Wei, Qingyang

    Purpose: The authors developed SPECT imaging capability on an animal PET scanner using a multiple-pinhole collimator and step-and-shoot helical data acquisition protocols. The objective of this work was to determine the preferred helical scan parameters, i.e., the angular and axial step sizes, and the imaging volume, that provide optimal imaging performance. Methods: The authors studied nine helical scan protocols formed by permuting three rotational and three axial step sizes. These step sizes were chosen around the reference values analytically calculated from the estimated spatial resolution of the SPECT system and the Nyquist sampling theorem. The nine helical protocols were evaluatedmore » by two figures-of-merit: the sampling completeness percentage (SCP) and the root-mean-square (RMS) resolution. SCP was an analytically calculated numerical index based on projection sampling. RMS resolution was derived from the reconstructed images of a sphere-grid phantom. Results: The RMS resolution results show that (1) the start and end pinhole planes of the helical scheme determine the axial extent of the effective field of view (EFOV), and (2) the diameter of the transverse EFOV is adequately calculated from the geometry of the pinhole opening, since the peripheral region beyond EFOV would introduce projection multiplexing and consequent effects. The RMS resolution results of the nine helical scan schemes show optimal resolution is achieved when the axial step size is the half, and the angular step size is about twice the corresponding values derived from the Nyquist theorem. The SCP results agree in general with that of RMS resolution but are less critical in assessing the effects of helical parameters and EFOV. Conclusions: The authors quantitatively validated the effective FOV of multiple pinhole helical scan protocols and proposed a simple method to calculate optimal helical scan parameters.« less

  17. Efficient O(N) integration for all-electron electronic structure calculation using numeric basis functions

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Havu, V.; Fritz Haber Institute of the Max Planck Society, Berlin; Blum, V.

    2009-12-01

    We consider the problem of developing O(N) scaling grid-based operations needed in many central operations when performing electronic structure calculations with numeric atom-centered orbitals as basis functions. We outline the overall formulation of localized algorithms, and specifically the creation of localized grid batches. The choice of the grid partitioning scheme plays an important role in the performance and memory consumption of the grid-based operations. Three different top-down partitioning methods are investigated, and compared with formally more rigorous yet much more expensive bottom-up algorithms. We show that a conceptually simple top-down grid partitioning scheme achieves essentially the same efficiency as themore » more rigorous bottom-up approaches.« less

  18. Effect of particle size distribution on the hydrodynamics of dense CFB risers

    NASA Astrophysics Data System (ADS)

    Bakshi, Akhilesh; Khanna, Samir; Venuturumilli, Raj; Altantzis, Christos; Ghoniem, Ahmed

    2015-11-01

    Circulating Fluidized Beds (CFB) are favorable in the energy and chemical industries, due to their high efficiency. While accurate hydrodynamic modeling is essential for optimizing performance, most CFB riser simulations are performed assuming equally-sized solid particles, owing to limited computational resources. Even though this approach yields reasonable predictions, it neglects commonly observed experimental findings suggesting the strong effect of particle size distribution (psd) on the hydrodynamics and chemical conversion. Thus, this study is focused on the inclusion of discrete particle sizes to represent the psd and its effect on fluidization via 2D numerical simulations. The particle sizes and corresponding mass fluxes are obtained using experimental data in dense CFB riser while the modeling framework is described in Bakshi et al 2015. Simulations are conducted at two scales: (a) fine grid to resolve heterogeneous structures and (b) coarse grid using EMMS sub-grid modifications. Using suitable metrics which capture bed dynamics, this study provides insights into segregation and mixing of particles as well as highlights need for improved sub-grid models.

  19. Dose calculation algorithm of fast fine-heterogeneity correction for heavy charged particle radiotherapy.

    PubMed

    Kanematsu, Nobuyuki

    2011-04-01

    This work addresses computing techniques for dose calculations in treatment planning with proton and ion beams, based on an efficient kernel-convolution method referred to as grid-dose spreading (GDS) and accurate heterogeneity-correction method referred to as Gaussian beam splitting. The original GDS algorithm suffered from distortion of dose distribution for beams tilted with respect to the dose-grid axes. Use of intermediate grids normal to the beam field has solved the beam-tilting distortion. Interplay of arrangement between beams and grids was found as another intrinsic source of artifact. Inclusion of rectangular-kernel convolution in beam transport, to share the beam contribution among the nearest grids in a regulatory manner, has solved the interplay problem. This algorithmic framework was applied to a tilted proton pencil beam and a broad carbon-ion beam. In these cases, while the elementary pencil beams individually split into several tens, the calculation time increased only by several times with the GDS algorithm. The GDS and beam-splitting methods will complementarily enable accurate and efficient dose calculations for radiotherapy with protons and ions. Copyright © 2010 Associazione Italiana di Fisica Medica. Published by Elsevier Ltd. All rights reserved.

  20. FAS multigrid calculations of three dimensional flow using non-staggered grids

    NASA Technical Reports Server (NTRS)

    Matovic, D.; Pollard, A.; Becker, H. A.; Grandmaison, E. W.

    1993-01-01

    Grid staggering is a well known remedy for the problem of velocity/pressure coupling in incompressible flow calculations. Numerous inconveniences occur, however, when staggered grids are implemented, particularly when a general-purpose code, capable of handling irregular three-dimensional domains, is sought. In several non-staggered grid numerical procedures proposed in the literature, the velocity/pressure coupling is achieved by either pressure or velocity (momentum) averaging. This approach is not convenient for simultaneous (block) solvers that are preferred when using multigrid methods. A new method is introduced in this paper that is based upon non-staggered grid formulation with a set of virtual cell face velocities used for pressure/velocity coupling. Instead of pressure or velocity averaging, a momentum balance at the cell face is used as a link between the momentum and mass balance constraints. The numerical stencil is limited to 9 nodes (in 2D) or 27 nodes (in 3D), both during the smoothing and inter-grid transfer, which is a convenient feature when a block point solver is applied. The results for a lid-driven cavity and a cube in a lid-driven cavity are presented and compared to staggered grid calculations using the same multigrid algorithm. The method is shown to be stable and produce a smooth (wiggle-free) pressure field.

  1. On-the-fly Numerical Surface Integration for Finite-Difference Poisson-Boltzmann Methods.

    PubMed

    Cai, Qin; Ye, Xiang; Wang, Jun; Luo, Ray

    2011-11-01

    Most implicit solvation models require the definition of a molecular surface as the interface that separates the solute in atomic detail from the solvent approximated as a continuous medium. Commonly used surface definitions include the solvent accessible surface (SAS), the solvent excluded surface (SES), and the van der Waals surface. In this study, we present an efficient numerical algorithm to compute the SES and SAS areas to facilitate the applications of finite-difference Poisson-Boltzmann methods in biomolecular simulations. Different from previous numerical approaches, our algorithm is physics-inspired and intimately coupled to the finite-difference Poisson-Boltzmann methods to fully take advantage of its existing data structures. Our analysis shows that the algorithm can achieve very good agreement with the analytical method in the calculation of the SES and SAS areas. Specifically, in our comprehensive test of 1,555 molecules, the average unsigned relative error is 0.27% in the SES area calculations and 1.05% in the SAS area calculations at the grid spacing of 1/2Å. In addition, a systematic correction analysis can be used to improve the accuracy for the coarse-grid SES area calculations, with the average unsigned relative error in the SES areas reduced to 0.13%. These validation studies indicate that the proposed algorithm can be applied to biomolecules over a broad range of sizes and structures. Finally, the numerical algorithm can also be adapted to evaluate the surface integral of either a vector field or a scalar field defined on the molecular surface for additional solvation energetics and force calculations.

  2. Numerical Simulations of Two-Phase Reacting Flow in a Single-Element Lean Direct Injection (LDI) Combustor Using NCC

    NASA Technical Reports Server (NTRS)

    Liu, Nan-Suey; Shih, Tsan-Hsing; Wey, C. Thomas

    2011-01-01

    A series of numerical simulations of Jet-A spray reacting flow in a single-element lean direct injection (LDI) combustor have been conducted by using the National Combustion Code (NCC). The simulations have been carried out using the time filtered Navier-Stokes (TFNS) approach ranging from the steady Reynolds-averaged Navier-Stokes (RANS), unsteady RANS (URANS), to the dynamic flow structure simulation (DFS). The sub-grid model employed for turbulent mixing and combustion includes the well-mixed model, the linear eddy mixing (LEM) model, and the filtered mass density function (FDF/PDF) model. The starting condition of the injected liquid spray is specified via empirical droplet size correlation, and a five-species single-step global reduced mechanism is employed for fuel chemistry. All the calculations use the same grid whose resolution is of the RANS type. Comparisons of results from various models are presented.

  3. Computation of the acoustic radiation force using the finite-difference time-domain method.

    PubMed

    Cai, Feiyan; Meng, Long; Jiang, Chunxiang; Pan, Yu; Zheng, Hairong

    2010-10-01

    The computational details related to calculating the acoustic radiation force on an object using a 2-D grid finite-difference time-domain method (FDTD) are presented. The method is based on propagating the stress and velocity fields through the grid and determining the energy flow with and without the object. The axial and radial acoustic radiation forces predicted by FDTD method are in excellent agreement with the results obtained by analytical evaluation of the scattering method. In particular, the results indicate that it is possible to trap the steel cylinder in the radial direction by optimizing the width of Gaussian source and the operation frequency. As the sizes of the relating objects are smaller than or comparable to wavelength, the algorithm presented here can be easily extended to 3-D and include torque computation algorithms, thus providing a highly flexible and universally usable computation engine.

  4. Evaluation of the Lattice-Boltzmann Equation Solver PowerFLOW for Aerodynamic Applications

    NASA Technical Reports Server (NTRS)

    Lockard, David P.; Luo, Li-Shi; Singer, Bart A.; Bushnell, Dennis M. (Technical Monitor)

    2000-01-01

    A careful comparison of the performance of a commercially available Lattice-Boltzmann Equation solver (Power-FLOW) was made with a conventional, block-structured computational fluid-dynamics code (CFL3D) for the flow over a two-dimensional NACA-0012 airfoil. The results suggest that the version of PowerFLOW used in the investigation produced solutions with large errors in the computed flow field; these errors are attributed to inadequate resolution of the boundary layer for reasons related to grid resolution and primitive turbulence modeling. The requirement of square grid cells in the PowerFLOW calculations limited the number of points that could be used to span the boundary layer on the wing and still keep the computation size small enough to fit on the available computers. Although not discussed in detail, disappointing results were also obtained with PowerFLOW for a cavity flow and for the flow around a generic helicopter configuration.

  5. Research on Streamlines and Aerodynamic Heating for Unstructured Grids on High-Speed Vehicles

    NASA Technical Reports Server (NTRS)

    DeJarnette, Fred R.; Hamilton, H. Harris (Technical Monitor)

    2001-01-01

    Engineering codes are needed which can calculate convective heating rates accurately and expeditiously on the surfaces of high-speed vehicles. One code which has proven to meet these needs is the Langley Approximate Three-Dimensional Convective Heating (LATCH) code. It uses the axisymmetric analogue in an integral boundary-layer method to calculate laminar and turbulent heating rates along inviscid surface streamlines. It requires the solution of the inviscid flow field to provide the surface properties needed to calculate the streamlines and streamline metrics. The LATCH code has been used with inviscid codes which calculated the flow field on structured grids, Several more recent inviscid codes calculate flow field properties on unstructured grids. The present research develops a method to calculate inviscid surface streamlines, the streamline metrics, and heating rates using the properties calculated from inviscid flow fields on unstructured grids. Mr. Chris Riley, prior to his departure from NASA LaRC, developed a preliminary code in the C language, called "UNLATCH", to accomplish these goals. No publication was made on his research. The present research extends and improves on the code developed by Riley. Particular attention is devoted to the stagnation region, and the method is intended for programming in the FORTRAN 90 language.

  6. Investigation of advanced counterrotation blade configuration concepts for high speed turboprop systems, task 1: Ducted propfan analysis

    NASA Technical Reports Server (NTRS)

    Hall, Edward J.; Delaney, Robert A.; Bettner, James L.

    1990-01-01

    The time-dependent three-dimensional Euler equations of gas dynamics were solved numerically to study the steady compressible transonic flow about ducted propfan propulsion systems. Aerodynamic calculations were based on a four-stage Runge-Kutta time-marching finite volume solution technique with added numerical dissipation. An implicit residual smoothing operator was used to aid convergence. Two calculation grids were employed in this study. The first grid utilized an H-type mesh network with a branch cut opening to represent the axisymmetric cowl. The second grid utilized a multiple-block mesh system with a C-type grid about the cowl. The individual blocks were numerically coupled in the Euler solver. Grid systems were generated by a combined algebraic/elliptic algortihm developed specifically for ducted propfans. Numerical calculations were initially performed for unducted propfans to verify the accuracy of the three-dimensional Euler formulation. The Euler analyses were then applied for the calculation of ducted propfan flows, and predicted results were compared with experimental data for two cases. The three-dimensional Euler analyses displayed exceptional accuracy, although certain parameters were observed to be very sensitive to geometric deflections. Both solution schemes were found to be very robust and demonstrated nearly equal efficiency and accuracy, although it was observed that the multi-block C-grid formulation provided somewhat better resolution of the cowl leading edge region.

  7. Influence of grid resolution, parcel size and drag models on bubbling fluidized bed simulation

    DOE PAGES

    Lu, Liqiang; Konan, Arthur; Benyahia, Sofiane

    2017-06-02

    Here in this paper, a bubbling fluidized bed is simulated with different numerical parameters, such as grid resolution and parcel size. We examined also the effect of using two homogeneous drag correlations and a heterogeneous drag based on the energy minimization method. A fast and reliable bubble detection algorithm was developed based on the connected component labeling. The radial and axial solids volume fraction profiles are compared with experiment data and previous simulation results. These results show a significant influence of drag models on bubble size and voidage distributions and a much less dependence on numerical parameters. With a heterogeneousmore » drag model that accounts for sub-scale structures, the void fraction in the bubbling fluidized bed can be well captured with coarse grid and large computation parcels. Refining the CFD grid and reducing the parcel size can improve the simulation results but with a large increase in computation cost.« less

  8. Heat budget observations for the FIRE/SRB Wisconsin experiment region from October 9 through November 2, 1986

    NASA Technical Reports Server (NTRS)

    Whitlock, Charles H.; Lecroy, Stuart R.

    1987-01-01

    A map and concise tables are presented which show locations, pixel size, and heat budget products from the NOAA-9 satellite for the FIRE/SRB Wisconsin experiment region during the period 9 October through 2 November 1986. In addition to the operational standard products, a narrowband albedo parameter is calculated and presented based on values from AVHRR band 1. This parameter is useful in identifying and/or quantifying clouds on a global basis using a polar-stereographic grid system.

  9. Magnetic resonance imaging metallic artifact of commonly encountered surgical implants and foreign material.

    PubMed

    Sutherland-Smith, James; Tilley, Brenda

    2012-01-01

    Magnetic resonance imaging (MRI) artifacts secondary to metallic implants and foreign bodies are well described. Herein, we provide quantitative data from veterinary implants including total hip arthroplasty implants, cranial cruciate repair implants, surgical screws, a skin staple, ligation clips, an identification microchip, ameroid constrictor, and potential foreign bodies including air gun and BB projectiles and a sewing needle. The objects were scanned in a gelatin phantom with plastic grid using standardized T2-weighted turbo-spin echo (TSE), T1-weighted spin echo, and T2*-weighted gradient recalled echo (GRE) image acquisitions at 1.5 T. Maximum linear dimensions and areas of signal voiding and grid distortion were calculated using a DICOM workstation for each sequence and object. Artifact severity was similar between the T2-weighted TSE and T1-weighted images, while the T2*-weighted images were most susceptible to artifact. Metal type influenced artifact size with the largest artifacts arising from steel objects followed by surgical stainless steel, titanium, and lead. For animals with metallic surgical implants or foreign bodies, the quantification of the artifact size will help guide clinicians on the viability of MRI. © 2012 Veterinary Radiology & Ultrasound.

  10. pyDockWEB: a web server for rigid-body protein-protein docking using electrostatics and desolvation scoring.

    PubMed

    Jiménez-García, Brian; Pons, Carles; Fernández-Recio, Juan

    2013-07-01

    pyDockWEB is a web server for the rigid-body docking prediction of protein-protein complex structures using a new version of the pyDock scoring algorithm. We use here a new custom parallel FTDock implementation, with adjusted grid size for optimal FFT calculations, and a new version of pyDock, which dramatically speeds up calculations while keeping the same predictive accuracy. Given the 3D coordinates of two interacting proteins, pyDockWEB returns the best docking orientations as scored mainly by electrostatics and desolvation energy. The server does not require registration by the user and is freely accessible for academics at http://life.bsc.es/servlet/pydock. Supplementary data are available at Bioinformatics online.

  11. Predicting mosaics and wildlife diversity resulting from fire disturbance to a forest ecosystem

    NASA Astrophysics Data System (ADS)

    Potter, Meredith W.; Kessell, Stephen R.

    1980-05-01

    A model for predicting community mosaics and wildlife diversity resulting from fire disturbance to a forest ecosystem is presented. It applies an algorithm that delineates the size and shape of each patch from grid-based input data and calculates standard diversity measures for the entire mosaic of community patches and their included animal species. The user can print these diversity calculations, maps of the current community-type-age-class mosaic, and maps of habitat utilization by each animal species. Furthermore, the user can print estimates of changes in each resulting from natural disturbance. Although data and resolution level independent, the model is demonstrated and tested with data from the Lewis and Clark National Forest in Montana.

  12. 3D RISM theory with fast reciprocal-space electrostatics.

    PubMed

    Heil, Jochen; Kast, Stefan M

    2015-03-21

    The calculation of electrostatic solute-solvent interactions in 3D RISM ("three-dimensional reference interaction site model") integral equation theory is recast in a form that allows for a computational treatment analogous to the "particle-mesh Ewald" formalism as used for molecular simulations. In addition, relations that connect 3D RISM correlation functions and interaction potentials with thermodynamic quantities such as the chemical potential and average solute-solvent interaction energy are reformulated in a way that calculations of expensive real-space electrostatic terms on the 3D grid are completely avoided. These methodical enhancements allow for both, a significant speedup particularly for large solute systems and a smoother convergence of predicted thermodynamic quantities with respect to box size, as illustrated for several benchmark systems.

  13. Application of the FUN3D Unstructured-Grid Navier-Stokes Solver to the 4th AIAA Drag Prediction Workshop Cases

    NASA Technical Reports Server (NTRS)

    Lee-Rausch, Elizabeth M.; Hammond, Dana P.; Nielsen, Eric J.; Pirzadeh, S. Z.; Rumsey, Christopher L.

    2010-01-01

    FUN3D Navier-Stokes solutions were computed for the 4th AIAA Drag Prediction Workshop grid convergence study, downwash study, and Reynolds number study on a set of node-based mixed-element grids. All of the baseline tetrahedral grids were generated with the VGRID (developmental) advancing-layer and advancing-front grid generation software package following the gridding guidelines developed for the workshop. With maximum grid sizes exceeding 100 million nodes, the grid convergence study was particularly challenging for the node-based unstructured grid generators and flow solvers. At the time of the workshop, the super-fine grid with 105 million nodes and 600 million elements was the largest grid known to have been generated using VGRID. FUN3D Version 11.0 has a completely new pre- and post-processing paradigm that has been incorporated directly into the solver and functions entirely in a parallel, distributed memory environment. This feature allowed for practical pre-processing and solution times on the largest unstructured-grid size requested for the workshop. For the constant-lift grid convergence case, the convergence of total drag is approximately second-order on the finest three grids. The variation in total drag between the finest two grids is only 2 counts. At the finest grid levels, only small variations in wing and tail pressure distributions are seen with grid refinement. Similarly, a small wing side-of-body separation also shows little variation at the finest grid levels. Overall, the FUN3D results compare well with the structured-grid code CFL3D. The FUN3D downwash study and Reynolds number study results compare well with the range of results shown in the workshop presentations.

  14. Getting the current out

    NASA Astrophysics Data System (ADS)

    Burger, D. R.

    1983-11-01

    Progress of a photovoltaic (PV) device from a research concept to a competitive power-generation source requires an increasing concern with current collection. The initial metallization focus is usually on contact resistance, since a good ohmic contact is desirable for accurate device characterization measurements. As the device grows in size, sheet resistance losses become important and a metal grid is usually added to reduce the effective sheet resistance. Later, as size and conversion efficiency continue to increase, grid-line resistance and cell shadowing must be considered simultaneously, because grid-line resistance is inversely related to total grid-line area and cell shadowing is directly related. A PV cell grid design must consider the five power-loss phenomena mentioned above: sheet resistance, contact resistance, grid resistance, bus-bar resistance and cell shadowing. Although cost, reliability and usage are important factors in deciding upon the best metallization system, this paper will focus only upon grid-line design and substrate material problems for flat-plate solar arrays.

  15. GRID BLACKOUT IN VACUUM TUBES

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hardin, K.D.

    1961-06-01

    A method which gives quantitative data is presented which allows for characterization of the grid blackout effect and is applicable to calculation of circuit degradation. Data are presented for several tube types which show developed bias and discharge time constants as a function of pulse input conditions. Blackout can seriously change the performance of any vacuum tube circuit which utilizes the tube in positive grid operation. The effects on CW oscillators and UHF mixers are discussed. An equivalent circuit which simulates some portions of the blackout phenomenon is presented and used to calculate effective capacitance and resistance associated with themore » grid surface. (auth)« less

  16. Locally refined block-centred finite-difference groundwater models: Evaluation of parameter sensitivity and the consequences for inverse modelling

    USGS Publications Warehouse

    Mehl, S.; Hill, M.C.

    2002-01-01

    Models with local grid refinement, as often required in groundwater models, pose special problems for model calibration. This work investigates the calculation of sensitivities and the performance of regression methods using two existing and one new method of grid refinement. The existing local grid refinement methods considered are: (a) a variably spaced grid in which the grid spacing becomes smaller near the area of interest and larger where such detail is not needed, and (b) telescopic mesh refinement (TMR), which uses the hydraulic heads or fluxes of a regional model to provide the boundary conditions for a locally refined model. The new method has a feedback between the regional and local grids using shared nodes, and thereby, unlike the TMR methods, balances heads and fluxes at the interfacing boundary. Results for sensitivities are compared for the three methods and the effect of the accuracy of sensitivity calculations are evaluated by comparing inverse modelling results. For the cases tested, results indicate that the inaccuracies of the sensitivities calculated using the TMR approach can cause the inverse model to converge to an incorrect solution.

  17. Locally refined block-centered finite-difference groundwater models: Evaluation of parameter sensitivity and the consequences for inverse modelling and predictions

    USGS Publications Warehouse

    Mehl, S.; Hill, M.C.

    2002-01-01

    Models with local grid refinement, as often required in groundwater models, pose special problems for model calibration. This work investigates the calculation of sensitivities and performance of regression methods using two existing and one new method of grid refinement. The existing local grid refinement methods considered are (1) a variably spaced grid in which the grid spacing becomes smaller near the area of interest and larger where such detail is not needed and (2) telescopic mesh refinement (TMR), which uses the hydraulic heads or fluxes of a regional model to provide the boundary conditions for a locally refined model. The new method has a feedback between the regional and local grids using shared nodes, and thereby, unlike the TMR methods, balances heads and fluxes at the interfacing boundary. Results for sensitivities are compared for the three methods and the effect of the accuracy of sensitivity calculations are evaluated by comparing inverse modelling results. For the cases tested, results indicate that the inaccuracies of the sensitivities calculated using the TMR approach can cause the inverse model to converge to an incorrect solution.

  18. Limited Area Coverage/High Resolution Picture Transmission (LAC/HRPT) data vegetative index calculation processor user's manual

    NASA Technical Reports Server (NTRS)

    Obrien, S. O. (Principal Investigator)

    1980-01-01

    The program, LACVIN, calculates vegetative indexes numbers on limited area coverage/high resolution picture transmission data for selected IJ grid sections. The IJ grid sections were previously extracted from the full resolution data tapes and stored on disk files.

  19. Sparse grid techniques for particle-in-cell schemes

    NASA Astrophysics Data System (ADS)

    Ricketson, L. F.; Cerfon, A. J.

    2017-02-01

    We propose the use of sparse grids to accelerate particle-in-cell (PIC) schemes. By using the so-called ‘combination technique’ from the sparse grids literature, we are able to dramatically increase the size of the spatial cells in multi-dimensional PIC schemes while paying only a slight penalty in grid-based error. The resulting increase in cell size allows us to reduce the statistical noise in the simulation without increasing total particle number. We present initial proof-of-principle results from test cases in two and three dimensions that demonstrate the new scheme’s efficiency, both in terms of computation time and memory usage.

  20. Dose evaluation of Grid Therapy using a 6 MV flattening filter-free (FFF) photon beam: A Monte Carlo study.

    PubMed

    Martínez-Rovira, Immaculada; Puxeu-Vaqué, Josep; Prezado, Yolanda

    2017-10-01

    Spatially fractionated radiotherapy is a strategy to overcome the main limitation of radiotherapy, i.e., the restrained normal tissue tolerances. A well-known example is Grid Therapy, which is currently performed at some hospitals using megavoltage photon beams delivered by Linacs. Grid Therapy has been successfully used in the management of bulky abdominal tumors with low toxicity. The aim of this work was to evaluate whether an improvement in therapeutic index in Grid Therapy can be obtained by implementing it in a flattening filter-free (FFF) Linac. The rationale behind is that the removal of the flattening filter shifts the beam energy spectrum towards lower energies and increase the photon fluence. Lower energies result in a reduction of lateral scattering and thus, to higher peak-to-valley dose ratios (PVDR) in normal tissues. In addition, the gain in fluence might allow using smaller beams leading a more efficient exploitation of dose-volume effects, and consequently, a better normal tissue sparing. Monte Carlo simulations were used to evaluate realistic dose distributions considering a 6 MV FFF photon beam from a standard medical Linac and a cerrobend mechanical collimator in different configurations: grid sizes of 0.3 × 0.3 cm 2 , 0.5 × 0.5 cm 2 , and 1 × 1 cm 2 and a corresponding center-to-center (ctc) distance of 0.6, 1, and 2 cm, respectively (total field size of 10 × 10 cm 2 ). As figure of merit, peak doses in depth, PVDR, output factors (OF), and penumbra values were assessed. Dose at the entrance is slightly higher than in conventional Grid Therapy. However, it is compensated by the large PVDR obtained at the entrance, reaching a maximum of 35 for a grid size of 1 × 1 cm 2 . Indeed, this grid size leads to very high PVDR values at all depths (≥ 10), which are much higher than in standard Grid Therapy. This may be beneficial for normal tissues but detrimental for tumor control, where a lower PVDR might be requested. In that case, higher valley doses in the tumor could be achieved by using an interlaced approach and/or adapting the ctc distance. The smallest grid size (0.3 × 0.3 cm 2 ) leads to low PVDR at all depths, comparable to standard Grid Therapy. However, the use of very thin beams might increase the normal tissue tolerances with respect to the grid size commonly used (1 × 1 cm 2 ). The gain in fluence provided by FFF implies that the important OF reduction (0.6) will not increase treatment time. Finally, the intermediate configuration (0.5 × 0.5 cm 2 ) provides high PVDR in the first 5 cm, and comparable PVDR to previous Grid Therapy works at depth. Therefore, this configuration might allow increasing the normal tissue tolerances with respect to Grid Therapy thanks to the higher PVDR and thinner beams, while a similar tumor control could be expected. The implementation of Grid Therapy in an FFF photon beam from medical Linac might lead to an improvement of the therapeutic index. Among the cases evaluated, a grid size of 0.5 × 0.5 cm 2 (1-cm-ctc) is the most advantageous configuration from the physics point of view. Radiobiological experiments are needed to fully explore this new avenue and to confirm our results. © 2017 American Association of Physicists in Medicine.

  1. Efficient grid-based techniques for density functional theory

    NASA Astrophysics Data System (ADS)

    Rodriguez-Hernandez, Juan Ignacio

    Understanding the chemical and physical properties of molecules and materials at a fundamental level often requires quantum-mechanical models for these substance's electronic structure. This type of many body quantum mechanics calculation is computationally demanding, hindering its application to substances with more than a few hundreds atoms. The supreme goal of many researches in quantum chemistry---and the topic of this dissertation---is to develop more efficient computational algorithms for electronic structure calculations. In particular, this dissertation develops two new numerical integration techniques for computing molecular and atomic properties within conventional Kohn-Sham-Density Functional Theory (KS-DFT) of molecular electronic structure. The first of these grid-based techniques is based on the transformed sparse grid construction. In this construction, a sparse grid is generated in the unit cube and then mapped to real space according to the pro-molecular density using the conditional distribution transformation. The transformed sparse grid was implemented in program deMon2k, where it is used as the numerical integrator for the exchange-correlation energy and potential in the KS-DFT procedure. We tested our grid by computing ground state energies, equilibrium geometries, and atomization energies. The accuracy on these test calculations shows that our grid is more efficient than some previous integration methods: our grids use fewer points to obtain the same accuracy. The transformed sparse grids were also tested for integrating, interpolating and differentiating in different dimensions (n = 1,2,3,6). The second technique is a grid-based method for computing atomic properties within QTAIM. It was also implemented in deMon2k. The performance of the method was tested by computing QTAIM atomic energies, charges, dipole moments, and quadrupole moments. For medium accuracy, our method is the fastest one we know of.

  2. Seismic Wave Propagation on the Tablet Computer

    NASA Astrophysics Data System (ADS)

    Emoto, K.

    2015-12-01

    Tablet computers widely used in recent years. The performance of the tablet computer is improving year by year. Some of them have performance comparable to the personal computer of a few years ago with respect to the calculation speed and the memory size. The convenience and the intuitive operation are the advantage of the tablet computer compared to the desktop PC. I developed the iPad application of the numerical simulation of the seismic wave propagation. The numerical simulation is based on the 2D finite difference method with the staggered-grid scheme. The number of the grid points is 512 x 384 = 196,608. The grid space is 200m in both horizontal and vertical directions. That is the calculation area is 102km x 77km. The time step is 0.01s. In order to reduce the user waiting time, the image of the wave field is drawn simultaneously with the calculation rather than playing the movie after the whole calculation. P and S wave energies are plotted on the screen every 20 steps (0.2s). There is the trade-off between the smooth simulation and the resolution of the wave field image. In the current setting, it takes about 30s to calculate the 10s wave propagation (50 times image updates). The seismogram at the receiver is displayed below of the wave field updated in real time. The default medium structure consists of 3 layers. The layer boundary is defined by 10 movable points with linear interpolation. Users can intuitively change to the arbitrary boundary shape by moving the point. Also users can easily change the source and the receiver positions. The favorite structure can be saved and loaded. For the advance simulation, users can introduce the random velocity fluctuation whose spectrum can be changed to the arbitrary shape. By using this application, everyone can simulate the seismic wave propagation without the special knowledge of the elastic wave equation. So far, the Japanese version of the application is released on the App Store. Now I am preparing the English version.

  3. Calculating Soil Wetness, Evapotranspiration and Carbon Cycle Processes Over Large Grid Areas Using a New Scaling Technique

    NASA Technical Reports Server (NTRS)

    Sellers, Piers

    2012-01-01

    Soil wetness typically shows great spatial variability over the length scales of general circulation model (GCM) grid areas (approx 100 km ), and the functions relating evapotranspiration and photosynthetic rate to local-scale (approx 1 m) soil wetness are highly non-linear. Soil respiration is also highly dependent on very small-scale variations in soil wetness. We therefore expect significant inaccuracies whenever we insert a single grid area-average soil wetness value into a function to calculate any of these rates for the grid area. For the particular case of evapotranspiration., this method - use of a grid-averaged soil wetness value - can also provoke severe oscillations in the evapotranspiration rate and soil wetness under some conditions. A method is presented whereby the probability distribution timction(pdf) for soil wetness within a grid area is represented by binning. and numerical integration of the binned pdf is performed to provide a spatially-integrated wetness stress term for the whole grid area, which then permits calculation of grid area fluxes in a single operation. The method is very accurate when 10 or more bins are used, can deal realistically with spatially variable precipitation, conserves moisture exactly and allows for precise modification of the soil wetness pdf after every time step. The method could also be applied to other ecological problems where small-scale processes must be area-integrated, or upscaled, to estimate fluxes over large areas, for example in treatments of the terrestrial carbon budget or trace gas generation.

  4. Navier-Stokes calculations on multi-element airfoils using a chimera-based solver

    NASA Technical Reports Server (NTRS)

    Jasper, Donald W.; Agrawal, Shreekant; Robinson, Brian A.

    1993-01-01

    A study of Navier-Stokes calculations of flows about multielement airfoils using a chimera grid approach is presented. The chimera approach utilizes structured, overlapped grids which allow great flexibility of grid arrangement and simplifies grid generation. Calculations are made for two-, three-, and four-element airfoils, and modeling of the effect of gap distance between elements is demonstrated for a two element case. Solutions are obtained using the thin-layer form of the Reynolds averaged Navier-Stokes equations with turbulence closure provided by the Baldwin-Lomax algebraic model or the Baldwin-Barth one equation model. The Baldwin-Barth turbulence model is shown to provide better agreement with experimental data and to dramatically improve convergence rates for some cases. Recently developed, improved farfield boundary conditions are incorporated into the solver for greater efficiency. Computed results show good comparison with experimental data which include aerodynamic forces, surface pressures, and boundary layer velocity profiles.

  5. Application of Approximate Pattern Matching in Two Dimensional Spaces to Grid Layout for Biochemical Network Maps

    PubMed Central

    Inoue, Kentaro; Shimozono, Shinichi; Yoshida, Hideaki; Kurata, Hiroyuki

    2012-01-01

    Background For visualizing large-scale biochemical network maps, it is important to calculate the coordinates of molecular nodes quickly and to enhance the understanding or traceability of them. The grid layout is effective in drawing compact, orderly, balanced network maps with node label spaces, but existing grid layout algorithms often require a high computational cost because they have to consider complicated positional constraints through the entire optimization process. Results We propose a hybrid grid layout algorithm that consists of a non-grid, fast layout (preprocessor) algorithm and an approximate pattern matching algorithm that distributes the resultant preprocessed nodes on square grid points. To demonstrate the feasibility of the hybrid layout algorithm, it is characterized in terms of the calculation time, numbers of edge-edge and node-edge crossings, relative edge lengths, and F-measures. The proposed algorithm achieves outstanding performances compared with other existing grid layouts. Conclusions Use of an approximate pattern matching algorithm quickly redistributes the laid-out nodes by fast, non-grid algorithms on the square grid points, while preserving the topological relationships among the nodes. The proposed algorithm is a novel use of the pattern matching, thereby providing a breakthrough for grid layout. This application program can be freely downloaded from http://www.cadlive.jp/hybridlayout/hybridlayout.html. PMID:22679486

  6. Application of approximate pattern matching in two dimensional spaces to grid layout for biochemical network maps.

    PubMed

    Inoue, Kentaro; Shimozono, Shinichi; Yoshida, Hideaki; Kurata, Hiroyuki

    2012-01-01

    For visualizing large-scale biochemical network maps, it is important to calculate the coordinates of molecular nodes quickly and to enhance the understanding or traceability of them. The grid layout is effective in drawing compact, orderly, balanced network maps with node label spaces, but existing grid layout algorithms often require a high computational cost because they have to consider complicated positional constraints through the entire optimization process. We propose a hybrid grid layout algorithm that consists of a non-grid, fast layout (preprocessor) algorithm and an approximate pattern matching algorithm that distributes the resultant preprocessed nodes on square grid points. To demonstrate the feasibility of the hybrid layout algorithm, it is characterized in terms of the calculation time, numbers of edge-edge and node-edge crossings, relative edge lengths, and F-measures. The proposed algorithm achieves outstanding performances compared with other existing grid layouts. Use of an approximate pattern matching algorithm quickly redistributes the laid-out nodes by fast, non-grid algorithms on the square grid points, while preserving the topological relationships among the nodes. The proposed algorithm is a novel use of the pattern matching, thereby providing a breakthrough for grid layout. This application program can be freely downloaded from http://www.cadlive.jp/hybridlayout/hybridlayout.html.

  7. Evaluation of truncation error and adaptive grid generation for the transonic full potential flow calculations

    NASA Technical Reports Server (NTRS)

    Nakamura, S.

    1983-01-01

    The effects of truncation error on the numerical solution of transonic flows using the full potential equation are studied. The effects of adapting grid point distributions to various solution aspects including shock waves is also discussed. A conclusion is that a rapid change of grid spacing is damaging to the accuracy of the flow solution. Therefore, in a solution adaptive grid application an optimal grid is obtained as a tradeoff between the amount of grid refinement and the rate of grid stretching.

  8. Filter and Grid Resolution in DG-LES

    NASA Astrophysics Data System (ADS)

    Miao, Ling; Sammak, Shervin; Madnia, Cyrus K.; Givi, Peyman

    2017-11-01

    The discontinuous Galerkin (DG) methodology has proven very effective for large eddy simulation (LES) of turbulent flows. Two important parameters in DG-LES are the grid resolution (h) and the filter size (Δ). In most previous work, the filter size is usually set to be proportional to the grid spacing. In this work, the DG method is combined with a subgrid scale (SGS) closure which is equivalent to that of the filtered density function (FDF). The resulting hybrid scheme is particularly attractive because a larger portion of the resolved energy is captured as the order of spectral approximation increases. Different cases for LES of a three-dimensional temporally developing mixing layer are appraised and a systematic parametric study is conducted to investigate the effects of grid resolution, the filter width size, and the order of spectral discretization. Comparative assessments are also made via the use of high resolution direct numerical simulation (DNS) data.

  9. The benefits of grid-scale storage on Oahu

    DOE PAGES

    Ellison, James F.; Rashkin, Lee J.; Serio, Joseph; ...

    2017-12-23

    The Hawaiian Electric Company intends to procure grid-scale Battery Energy Storage System (“BESS”) capacity. The purpose of this study is to determine whether providing contingency reserve or time-of-day shifting is of more benefit to the Oahu grid, and to better understand the relationship between BESS size and level of benefit. This is an independent study by Sandia, and is not being used to support the regulatory case for BESS capacity by Hawaiian Electric. The study team created a production cost model of the Oahu grid using data primarily from the Hawaiian Electric Company. The proposed BESS supplied contingency reserve inmore » one set of runs and time-of-day shifting in another. Supplying contingency reserve led to larger savings than time-of-day energy shifting. Assuming a renewable reserve and a quick-start reserve, and $15/MMBtu for Low-Sulphur Fuel Oil, the 50-MW/25-MWh, 100-MW/50-MWh, and 150-MW/75-MWh systems supplying contingency reserve provided, respectively, savings of 9.6, 15.6, and 18.3 million USD over system year 2018. Over the range of fuel prices tested, these cost savings were found to be directly proportional to the cost of fuel. Lastly, as the focus is the operational benefit of BESS capacity, the capacity value of the BESS was not included in benefit calculations.« less

  10. The benefits of grid-scale storage on Oahu

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ellison, James F.; Rashkin, Lee J.; Serio, Joseph

    The Hawaiian Electric Company intends to procure grid-scale Battery Energy Storage System (“BESS”) capacity. The purpose of this study is to determine whether providing contingency reserve or time-of-day shifting is of more benefit to the Oahu grid, and to better understand the relationship between BESS size and level of benefit. This is an independent study by Sandia, and is not being used to support the regulatory case for BESS capacity by Hawaiian Electric. The study team created a production cost model of the Oahu grid using data primarily from the Hawaiian Electric Company. The proposed BESS supplied contingency reserve inmore » one set of runs and time-of-day shifting in another. Supplying contingency reserve led to larger savings than time-of-day energy shifting. Assuming a renewable reserve and a quick-start reserve, and $15/MMBtu for Low-Sulphur Fuel Oil, the 50-MW/25-MWh, 100-MW/50-MWh, and 150-MW/75-MWh systems supplying contingency reserve provided, respectively, savings of 9.6, 15.6, and 18.3 million USD over system year 2018. Over the range of fuel prices tested, these cost savings were found to be directly proportional to the cost of fuel. Lastly, as the focus is the operational benefit of BESS capacity, the capacity value of the BESS was not included in benefit calculations.« less

  11. Variational treatment of electron-polyatomic-molecule scattering calculations using adaptive overset grids

    NASA Astrophysics Data System (ADS)

    Greenman, Loren; Lucchese, Robert R.; McCurdy, C. William

    2017-11-01

    The complex Kohn variational method for electron-polyatomic-molecule scattering is formulated using an overset-grid representation of the scattering wave function. The overset grid consists of a central grid and multiple dense atom-centered subgrids that allow the simultaneous spherical expansions of the wave function about multiple centers. Scattering boundary conditions are enforced by using a basis formed by the repeated application of the free-particle Green's function and potential Ĝ0+V ̂ on the overset grid in a Born-Arnoldi solution of the working equations. The theory is shown to be equivalent to a specific Padé approximant to the T matrix and has rapid convergence properties, in both the number of numerical basis functions employed and the number of partial waves employed in the spherical expansions. The method is demonstrated in calculations on methane and CF4 in the static-exchange approximation and compared in detail with calculations performed with the numerical Schwinger variational approach based on single-center expansions. An efficient procedure for operating with the free-particle Green's function and exchange operators (to which no approximation is made) is also described.

  12. Mapped grid methods for long-range molecules and cold collisions

    NASA Astrophysics Data System (ADS)

    Willner, K.; Dulieu, O.; Masnou-Seeuws, F.

    2004-01-01

    The paper discusses ways of improving the accuracy of numerical calculations for vibrational levels of diatomic molecules close to the dissociation limit or for ultracold collisions, in the framework of a grid representation. In order to avoid the implementation of very large grids, Kokoouline et al. [J. Chem. Phys. 110, 9865 (1999)] have proposed a mapping procedure through introduction of an adaptive coordinate x subjected to the variation of the local de Broglie wavelength as a function of the internuclear distance R. Some unphysical levels ("ghosts") then appear in the vibrational series computed via a mapped Fourier grid representation. In the present work the choice of the basis set is reexamined, and two alternative expansions are discussed: Sine functions and Hardy functions. It is shown that use of a basis set with fixed nodes at both grid ends is efficient to eliminate "ghost" solutions. It is further shown that the Hamiltonian matrix in the sine basis can be calculated very accurately by using an auxiliary basis of cosine functions, overcoming the problems arising from numerical calculation of the Jacobian J(x) of the R→x coordinate transformation.

  13. A WENO-solver combined with adaptive momentum discretization for the Wigner transport equation and its application to resonant tunneling diodes

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Dorda, Antonius, E-mail: dorda@tugraz.at; Schürrer, Ferdinand, E-mail: ferdinand.schuerrer@tugraz.at

    2015-03-01

    We present a novel numerical scheme for the deterministic solution of the Wigner transport equation, especially suited to deal with situations in which strong quantum effects are present. The unique feature of the algorithm is the expansion of the Wigner function in local basis functions, similar to finite element or finite volume methods. This procedure yields a discretization of the pseudo-differential operator that conserves the particle density on arbitrarily chosen grids. The high flexibility in refining the grid spacing together with the weighted essentially non-oscillatory (WENO) scheme for the advection term allows for an accurate and well-resolved simulation of themore » phase space dynamics. A resonant tunneling diode is considered as test case and a detailed convergence study is given by comparing the results to a non-equilibrium Green's functions calculation. The impact of the considered domain size and of the grid spacing is analyzed. The obtained convergence of the results towards a quasi-exact agreement of the steady state Wigner and Green's functions computations demonstrates the accuracy of the scheme, as well as the high flexibility to adjust to different physical situations.« less

  14. A WENO-solver combined with adaptive momentum discretization for the Wigner transport equation and its application to resonant tunneling diodes

    PubMed Central

    Dorda, Antonius; Schürrer, Ferdinand

    2015-01-01

    We present a novel numerical scheme for the deterministic solution of the Wigner transport equation, especially suited to deal with situations in which strong quantum effects are present. The unique feature of the algorithm is the expansion of the Wigner function in local basis functions, similar to finite element or finite volume methods. This procedure yields a discretization of the pseudo-differential operator that conserves the particle density on arbitrarily chosen grids. The high flexibility in refining the grid spacing together with the weighted essentially non-oscillatory (WENO) scheme for the advection term allows for an accurate and well-resolved simulation of the phase space dynamics. A resonant tunneling diode is considered as test case and a detailed convergence study is given by comparing the results to a non-equilibrium Green's functions calculation. The impact of the considered domain size and of the grid spacing is analyzed. The obtained convergence of the results towards a quasi-exact agreement of the steady state Wigner and Green's functions computations demonstrates the accuracy of the scheme, as well as the high flexibility to adjust to different physical situations. PMID:25892748

  15. A WENO-solver combined with adaptive momentum discretization for the Wigner transport equation and its application to resonant tunneling diodes.

    PubMed

    Dorda, Antonius; Schürrer, Ferdinand

    2015-03-01

    We present a novel numerical scheme for the deterministic solution of the Wigner transport equation, especially suited to deal with situations in which strong quantum effects are present. The unique feature of the algorithm is the expansion of the Wigner function in local basis functions, similar to finite element or finite volume methods. This procedure yields a discretization of the pseudo-differential operator that conserves the particle density on arbitrarily chosen grids. The high flexibility in refining the grid spacing together with the weighted essentially non-oscillatory (WENO) scheme for the advection term allows for an accurate and well-resolved simulation of the phase space dynamics. A resonant tunneling diode is considered as test case and a detailed convergence study is given by comparing the results to a non-equilibrium Green's functions calculation. The impact of the considered domain size and of the grid spacing is analyzed. The obtained convergence of the results towards a quasi-exact agreement of the steady state Wigner and Green's functions computations demonstrates the accuracy of the scheme, as well as the high flexibility to adjust to different physical situations.

  16. An estimate of field size distributions for selected sites in the major grain producing countries

    NASA Technical Reports Server (NTRS)

    Podwysocki, M. H.

    1977-01-01

    The field size distributions for the major grain producing countries of the World were estimated. LANDSAT-1 and 2 images were evaluated for two areas each in the United States, People's Republic of China, and the USSR. One scene each was evaluated for France, Canada, and India. Grid sampling was done for representative sub-samples of each image, measuring the long and short axes of each field; area was then calculated. Each of the resulting data sets was computer analyzed for their frequency distributions. Nearly all frequency distributions were highly peaked and skewed (shifted) towards small values, approaching that of either a Poisson or log-normal distribution. The data were normalized by a log transformation, creating a Gaussian distribution which has moments readily interpretable and useful for estimating the total population of fields. Resultant predictors of the field size estimates are discussed.

  17. Calculating depths to shallow magnetic sources using aeromagnetic data from the Tucson Basin

    USGS Publications Warehouse

    Casto, Daniel W.

    2001-01-01

    Using gridded high-resolution aeromagnetic data, the performance of several automated 3-D depth-to-source methods was evaluated over shallow control sources based on how close their depth estimates came to the actual depths to the tops of the sources. For all three control sources, only the simple analytic signal method, the local wavenumber method applied to the vertical integral of the magnetic field, and the horizontal gradient method applied to the pseudo-gravity field provided median depth estimates that were close (-11% to +14% error) to the actual depths. Careful attention to data processing was required in order to calculate a sufficient number of depth estimates and to reduce the occurrence of false depth estimates. For example, to eliminate sampling bias, high-frequency noise and interference from deeper sources, it was necessary to filter the data before calculating derivative grids and subsequent depth estimates. To obtain smooth spatial derivative grids using finite differences, the data had to be gridded at intervals less than one percent of the anomaly wavelength. Before finding peak values in the derived signal grids, it was necessary to remove calculation noise by applying a low-pass filter in the grid-line directions and to re-grid at an interval that enabled the search window to encompass only the peaks of interest. Using the methods that worked best over the control sources, depth estimates over geologic sites of interest suggested the possible occurrence of volcanics nearly 170 meters beneath a city landfill. Also, a throw of around 2 kilometers was determined for a detachment fault that has a displacement of roughly 6 kilometers.

  18. Fatigue and Fracture Characterization of GlasGridRTM Reinforced Asphalt Concrete Pavement

    NASA Astrophysics Data System (ADS)

    Safavizadeh, Seyed Amirshayan

    The purpose of this research is to develop an experimental and analytical framework for describing, modeling, and predicting the reflective cracking patterns and crack growth rates in GlasGridRTM-reinforced asphalt pavements. In order to fulfill this objective, the effects of different interfacial conditions (mixture and tack coat type, and grid opening size) on reflective cracking-related failure mechanisms and the fatigue and fracture characteristics of fiberglass grid-reinforced asphalt concrete beams were studied by means of four- and threepoint bending notched beam fatigue tests (NBFTs) and cyclic and monotonic interface shear tests. The digital image correlation (DIC) technique was utilized for obtaining the displacement and strain contours of specimen surfaces during each test. The DIC analysis results were used to develop crack tip detection methods that were in turn used to determine interfacial crack lengths in the shear tests, and vertical and horizontal (interfacial) crack lengths in the notched beam fatigue tests. Linear elastic fracture mechanics (LEFM) principles were applied to the crack length data to describe the crack growth. In the case of the NBFTs, a finite element (FE) code was developed and used for modeling each beam at different stages of testing and back-calculating the stress intensity factors (SIFs) for the vertical and horizontal cracks. The local effect of reinforcement on the stiffness of the system at a vertical crack-interface intersection or the resistance of the grid system to the deflection differential at the joint/crack (hereinafter called joint stiffness) for GlasGrid-reinforced asphalt concrete beams was determined by implementing a joint stiffness parameter into the finite element code. The strain level dependency of the fatigue and fracture characteristics of the GlasGrid-reinforced beams was studied by performing four-point bending notched beam fatigue tests at strain levels of 600, 750, and 900 microstrain. These beam tests were conducted at 15°C, 20°C, and 23°C, with the main focus being to find the characteristics at 20°C. The results obtained from the tests at the different temperatures were used to investigate the effects of temperature on the reflective cracking performance of the gridreinforced beam specimens. The temperature tests were also used to investigate the validity of the time-temperature superposition (t-TS) principle in shear and the beam fatigue performance of the grid-reinforced specimens. The NBFT results suggest that different interlayer conditions do not reflect a unique failure mechanism, and thus, in order to predict and model the performance of grid-reinforced pavement, all the mechanisms involved in weakening its structural integrity, including damage within the asphalt layers and along the interface, must be considered. The shear and beam fatigue test results suggest that the grid opening size, interfacial bond quality, and mixture type play important roles in the reflective cracking performance of GlasGrid-reinforced asphalt pavements. According to the NBTF results, GlasGrid reinforcement retards reflective crack growth by stiffening the composite system and introducing a joint stiffness parameter. The results also show that the higher the bond strength and interlayer stiffness values, the higher the joint stiffness and retardation effects. The t-TS studies proved the validity of this principle in terms of the reflective crack growth of the grid-reinforced beam specimens and the shear modulus and shear strength of the grid-reinforced interfaces.

  19. On unstructured grids and solvers

    NASA Technical Reports Server (NTRS)

    Barth, T. J.

    1990-01-01

    The fundamentals and the state-of-the-art technology for unstructured grids and solvers are highlighted. Algorithms and techniques pertinent to mesh generation are discussed. It is shown that grid generation and grid manipulation schemes rely on fast multidimensional searching. Flow solution techniques for the Euler equations, which can be derived from the integral form of the equations are discussed. Sample calculations are also provided.

  20. Nonhydrostatic simulation of hyperpycnal river plumes on sloping continental shelves: Flow structures and nonhydrostatic effect

    NASA Astrophysics Data System (ADS)

    Tseng, Chien-Yung; Chou, Yi-Ju

    2018-04-01

    A three-dimensional nonhydrostatic coastal model SUNTANS is used to study hyperpycnal plumes on sloping continental shelves with idealized domain setup. The study aims to examine the nonhydrostatic effect of the plunging hyperpycnal plume and the associated flow structures on different shelf slopes. The unstructured triangular grid in SUNTANS allows for local refinement of the grid size for regions in which the flow varies abruptly, while retaining low-cost computation using the coarse grid resolution for regions in which the flow is more uniform. These nonhydrostatic simulations reveal detailed three-dimensional flow structures in both transient and steady states. Via comparison with the hydrostatic simulation, we show that the nonhydrostatic effect is particularly important before plunging, when the plume is subject to significant changes in both the along-shore and vertical directions. After plunging, where the plume becomes an undercurrent that is more spatially uniform, little difference is found between the hydrostatic and nonhydrostatic simulations in the present gentle- and mild-slope cases. A grid-dependence study shows that the nonhydrostatic effect can be seen only when the grid resolution is sufficiently fine that the calculation is not overly diffusive. A depth-integrated momentum budget analysis is then conducted to show that the flow convergence due to plunging is an important factor in the three-dimensional flow structures. Moreover, it shows that the nonhydrostatic effect becomes more important as the slope increases, and in the steep-slope case, neglect of transport of the vertical momentum during plunging in the hydrostatic case further leads to an erroneous prediction for the undercurrent.

  1. RCS of fundamental scatterers in the HF band by wire-grid modelling

    NASA Astrophysics Data System (ADS)

    Trueman, C. W.; Kubina, S. J.

    To extract the maximum information from the return of a radar target such as an aircraft, the target's scattering properties must be well known. Wire grid modeling allows a detailed representation of the surface of a complex scatterer such as an aircraft, in the frequency range where the aircraft size is comparable to a wavelength. A moment method analysis determines the currents on the wires of the grid including the interactions between all parts of the structure. Wire grid models of fundamental scatterers (plates, strips, cubes, and spheres) of sizes comparable to the wavelength in the 2-30 MHz range are analyzed. The study of the radar cross section (RCS) of wire grids in comparison with measured RCS data helps to establish guidelines for building wire grid models, specifying such parameters as where to locate wires, how short the segments must be, and what radius to use. The guidelines so developed can then be applied to build wire grid models of much more complex bodies such as aircraft with much greater confidence.

  2. School Finance and Technology: A Case Study Using Grid and Group Theory to Explore the Connections

    ERIC Educational Resources Information Center

    Case, Stephoni; Harris, Edward L.

    2014-01-01

    Using grid and group theory (Douglas 1982, 2011), the study described in this article examined the intersections of technology and school finance in four schools located in districts differing in size, wealth, and commitment to technology integration. In grid and group theory, grid refers to the degree to which policies and role prescriptions…

  3. An Eulerian/Lagrangian method for computing blade/vortex impingement

    NASA Technical Reports Server (NTRS)

    Steinhoff, John; Senge, Heinrich; Yonghu, Wenren

    1991-01-01

    A combined Eulerian/Lagrangian approach to calculating helicopter rotor flows with concentrated vortices is described. The method computes a general evolving vorticity distribution without any significant numerical diffusion. Concentrated vortices can be accurately propagated over long distances on relatively coarse grids with cores only several grid cells wide. The method is demonstrated for a blade/vortex impingement case in 2D and 3D where a vortex is cut by a rotor blade, and the results are compared to previous 2D calculations involving a fifth-order Navier-Stokes solver on a finer grid.

  4. GridTool: A surface modeling and grid generation tool

    NASA Technical Reports Server (NTRS)

    Samareh-Abolhassani, Jamshid

    1995-01-01

    GridTool is designed around the concept that the surface grids are generated on a set of bi-linear patches. This type of grid generation is quite easy to implement, and it avoids the problems associated with complex CAD surface representations and associated surface parameterizations. However, the resulting surface grids are close to but not on the original CAD surfaces. This problem can be alleviated by projecting the resulting surface grids onto the original CAD surfaces. GridTool is designed primary for unstructured grid generation systems. Currently, GridTool supports VGRID and FELISA systems, and it can be easily extended to support other unstructured grid generation systems. The data in GridTool is stored parametrically so that once the problem is set up, one can modify the surfaces and the entire set of points, curves and patches will be updated automatically. This is very useful in a multidisciplinary design and optimization process. GridTool is written entirely in ANSI 'C', the interface is based on the FORMS library, and the graphics is based on the GL library. The code has been tested successfully on IRIS workstations running IRIX4.0 and above. The memory is allocated dynamically, therefore, memory size will depend on the complexity of geometry/grid. GridTool data structure is based on a link-list structure which allows the required memory to expand and contract dynamically according to the user's data size and action. Data structure contains several types of objects such as points, curves, patches, sources and surfaces. At any given time, there is always an active object which is drawn in magenta, or in their highlighted colors as defined by the resource file which will be discussed later.

  5. Noniterative three-dimensional grid generation using parabolic partial differential equations

    NASA Technical Reports Server (NTRS)

    Edwards, T. A.

    1985-01-01

    A new algorithm for generating three-dimensional grids has been developed and implemented which numerically solves a parabolic partial differential equation (PDE). The solution procedure marches outward in two coordinate directions, and requires inversion of a scalar tridiagonal system in the third. Source terms have been introduced to control the spacing and angle of grid lines near the grid boundaries, and to control the outer boundary point distribution. The method has been found to generate grids about 100 times faster than comparable grids generated via solution of elliptic PDEs, and produces smooth grids for finite-difference flow calculations.

  6. Hybrid PV/Wind Power Systems Incorporating Battery Storage and Considering the Stochastic Nature of Renewable Resources

    NASA Astrophysics Data System (ADS)

    Barnawi, Abdulwasa Bakr

    Hybrid power generation system and distributed generation technology are attracting more investments due to the growing demand for energy nowadays and the increasing awareness regarding emissions and their environmental impacts such as global warming and pollution. The price fluctuation of crude oil is an additional reason for the leading oil producing countries to consider renewable resources as an alternative. Saudi Arabia as the top oil exporter country in the word announced the "Saudi Arabia Vision 2030" which is targeting to generate 9.5 GW of electricity from renewable resources. Two of the most promising renewable technologies are wind turbines (WT) and photovoltaic cells (PV). The integration or hybridization of photovoltaics and wind turbines with battery storage leads to higher adequacy and redundancy for both autonomous and grid connected systems. This study presents a method for optimal generation unit planning by installing a proper number of solar cells, wind turbines, and batteries in such a way that the net present value (NPV) is minimized while the overall system redundancy and adequacy is maximized. A new renewable fraction technique (RFT) is used to perform the generation unit planning. RFT was tested and validated with particle swarm optimization and HOMER Pro under the same conditions and environment. Renewable resources and load randomness and uncertainties are considered. Both autonomous and grid-connected system designs were adopted in the optimal generation units planning process. An uncertainty factor was designed and incorporated in both autonomous and grid connected system designs. In the autonomous hybrid system design model, the strategy including an additional amount of operation reserve as a percent of the hourly load was considered to deal with resource uncertainty since the battery storage system is the only backup. While in the grid-connected hybrid system design model, demand response was incorporated to overcome the impact of uncertainty and perform energy trading between the hybrid grid utility and main grid utility in addition to the designed uncertainty factor. After the generation unit planning was carried out and component sizing was determined, adequacy evaluation was conducted by calculating the loss of load expectation adequacy index for different contingency criteria considering probability of equipment failure. Finally, a microgrid planning was conducted by finding the proper size and location to install distributed generation units in a radial distribution network.

  7. The use of solution adaptive grids in solving partial differential equations

    NASA Technical Reports Server (NTRS)

    Anderson, D. A.; Rai, M. M.

    1982-01-01

    The grid point distribution used in solving a partial differential equation using a numerical method has a substantial influence on the quality of the solution. An adaptive grid which adjusts as the solution changes provides the best results when the number of grid points available for use during the calculation is fixed. Basic concepts used in generating and applying adaptive grids are reviewed in this paper, and examples illustrating applications of these concepts are presented.

  8. Recombination of the steering vector of the triangle grid array in quaternions and the reduction of the MUSIC algorithm

    NASA Astrophysics Data System (ADS)

    Bai, Chen; Han, Dongjuan

    2018-04-01

    MUSIC is widely used on DOA estimation. Triangle grid is a common kind of the arrangement of array, but it is more complicated than rectangular array in calculation of steering vector. In this paper, the quaternions algorithm can reduce dimension of vector and make the calculation easier.

  9. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Nutaro, James; Kuruganti, Teja

    Numerical simulations of the wave equation that are intended to provide accurate time domain solutions require a computational mesh with grid points separated by a distance less than the wavelength of the source term and initial data. However, calculations of radio signal pathloss generally do not require accurate time domain solutions. This paper describes an approach for calculating pathloss by using the finite difference time domain and transmission line matrix models of wave propagation on a grid with points separated by distances much greater than the signal wavelength. The calculated pathloss can be kept close to the true value formore » freespace propagation with an appropriate selection of initial conditions. This method can also simulate diffraction with an error governed by the ratio of the signal wavelength to the grid spacing.« less

  10. MrGrid: A Portable Grid Based Molecular Replacement Pipeline

    PubMed Central

    Reboul, Cyril F.; Androulakis, Steve G.; Phan, Jennifer M. N.; Whisstock, James C.; Goscinski, Wojtek J.; Abramson, David; Buckle, Ashley M.

    2010-01-01

    Background The crystallographic determination of protein structures can be computationally demanding and for difficult cases can benefit from user-friendly interfaces to high-performance computing resources. Molecular replacement (MR) is a popular protein crystallographic technique that exploits the structural similarity between proteins that share some sequence similarity. But the need to trial permutations of search models, space group symmetries and other parameters makes MR time- and labour-intensive. However, MR calculations are embarrassingly parallel and thus ideally suited to distributed computing. In order to address this problem we have developed MrGrid, web-based software that allows multiple MR calculations to be executed across a grid of networked computers, allowing high-throughput MR. Methodology/Principal Findings MrGrid is a portable web based application written in Java/JSP and Ruby, and taking advantage of Apple Xgrid technology. Designed to interface with a user defined Xgrid resource the package manages the distribution of multiple MR runs to the available nodes on the Xgrid. We evaluated MrGrid using 10 different protein test cases on a network of 13 computers, and achieved an average speed up factor of 5.69. Conclusions MrGrid enables the user to retrieve and manage the results of tens to hundreds of MR calculations quickly and via a single web interface, as well as broadening the range of strategies that can be attempted. This high-throughput approach allows parameter sweeps to be performed in parallel, improving the chances of MR success. PMID:20386612

  11. New method: calculation of magnification factor from an intracardiac marker.

    PubMed

    Cha, S D; Incarvito, J; Maranhao, V

    1983-01-01

    In order to calculate a magnification factor (MF), an intracardiac marker (pigtail catheter with markers) was evaluated using a new formula and correlated with the conventional grid method. By applying the Pythagorean theorem and trigonometry, a new formula was developed, which is (formula; see text) In an experimental study, MF by the intracardiac markers was 0.71 +/- 0.15 (M +/- SD) and one by the grid method was 0.72 +/- 0.15, with a correlation coefficient of 0.96. In patients study, MF by the intracardiac markers was 0.77 +/- 0.06 and one by the grid method was 0.77 +/- 0.05. We conclude that this new method is simple and the results were comparable to the conventional grid method at mid-chest level.

  12. CheckDen, a program to compute quantum molecular properties on spatial grids.

    PubMed

    Pacios, Luis F; Fernandez, Alberto

    2009-09-01

    CheckDen, a program to compute quantum molecular properties on a variety of spatial grids is presented. The program reads as unique input wavefunction files written by standard quantum packages and calculates the electron density rho(r), promolecule and density difference function, gradient of rho(r), Laplacian of rho(r), information entropy, electrostatic potential, kinetic energy densities G(r) and K(r), electron localization function (ELF), and localized orbital locator (LOL) function. These properties can be calculated on a wide range of one-, two-, and three-dimensional grids that can be processed by widely used graphics programs to render high-resolution images. CheckDen offers also other options as extracting separate atom contributions to the property computed, converting grid output data into CUBE and OpenDX volumetric data formats, and perform arithmetic combinations with grid files in all the recognized formats.

  13. Security and Stability Analysis of Wind Farms Integration into Distribution Network

    NASA Astrophysics Data System (ADS)

    Guan-yang, Li; Hongzhao, Wang; Guanglei, Li; Yamei, Cheng; Hong-zheng, Liu; Yi, Sun

    2017-05-01

    With the increasing share of the wind power in the power system, wind power fluctuations will cause obvious negative impacts on weak local grid. This paper firstly establish electromechanical transient simulation model for doubly fed induction wind turbine, then use Matlab/Simulink to achieve power flow calculation and transient simulation of power system including wind farms, the local synchronous generator, load, etc, finally analyze wind power on the impact of the local power grid under typical circumstances. The actual calculated results indicate that wind mutation causes little effect on the power grid, but when the three-phase short circuit fault happens, active power of wind power decreases sharply and the voltage of location of wind power into the grid also drop sharply, finally wind farm split from power system. This situation is not conducive to security and stability of the local power grid. It is necessary to develop security and stability measures in the future.

  14. Grid of Supergiant B[e] Models from HDUST Radiative Transfer

    NASA Astrophysics Data System (ADS)

    Domiciano de Souza, A.; Carciofi, A. C.

    2012-12-01

    By using the Monte Carlo radiative transfer code HDUST (developed by A. C. Carciofi and J..E. Bjorkman) we have built a grid of models for stars presenting the B[e] phenomenon and a bimodal outflowing envelope. The models are particularly adapted to the study of B[e] supergiants and FS CMa type stars. The adopted physical parameters of the calculated models make the grid well adapted to interpret high angular and high spectral observations, in particular spectro-interferometric data from ESO-VLTI instruments AMBER (near-IR at low and medium spectral resolution) and MIDI (mid-IR at low spectral resolution). The grid models include, for example, a central B star with different effective temperatures, a gas (hydrogen) and silicate dust circumstellar envelope with a bimodal mass loss presenting dust in the denser equatorial regions. The HDUST grid models were pre-calculated using the high performance parallel computing facility Mésocentre SIGAMM, located at OCA, France.

  15. Simulation of wave propagation in three-dimensional random media

    NASA Astrophysics Data System (ADS)

    Coles, Wm. A.; Filice, J. P.; Frehlich, R. G.; Yadlowsky, M.

    1995-04-01

    Quantitative error analyses for the simulation of wave propagation in three-dimensional random media, when narrow angular scattering is assumed, are presented for plane-wave and spherical-wave geometry. This includes the errors that result from finite grid size, finite simulation dimensions, and the separation of the two-dimensional screens along the propagation direction. Simple error scalings are determined for power-law spectra of the random refractive indices of the media. The effects of a finite inner scale are also considered. The spatial spectra of the intensity errors are calculated and compared with the spatial spectra of

  16. Spectral nudging to eliminate the effects of domain position and geometry in regional climate model simulations

    NASA Astrophysics Data System (ADS)

    Miguez-Macho, Gonzalo; Stenchikov, Georgiy L.; Robock, Alan

    2004-07-01

    It is well known that regional climate simulations are sensitive to the size and position of the domain chosen for calculations. Here we study the physical mechanisms of this sensitivity. We conducted simulations with the Regional Atmospheric Modeling System (RAMS) for June 2000 over North America at 50 km horizontal resolution using a 7500 km × 5400 km grid and NCEP/NCAR reanalysis as boundary conditions. The position of the domain was displaced in several directions, always maintaining the U.S. in the interior, out of the buffer zone along the lateral boundaries. Circulation biases developed a large scale structure, organized by the Rocky Mountains, resulting from a systematic shifting of the synoptic wave trains that crossed the domain. The distortion of the large-scale circulation was produced by interaction of the modeled flow with the lateral boundaries of the nested domain and varied when the position of the grid was altered. This changed the large-scale environment among the different simulations and translated into diverse conditions for the development of the mesoscale processes that produce most of precipitation for the Great Plains in the summer season. As a consequence, precipitation results varied, sometimes greatly, among the experiments with the different grid positions. To eliminate the dependence of results on the position of the domain, we used spectral nudging of waves longer than 2500 km above the boundary layer. Moisture was not nudged at any level. This constrained the synoptic scales to follow reanalysis while allowing the model to develop the small-scale dynamics responsible for the rainfall. Nudging of the large scales successfully eliminated the variation of precipitation results when the grid was moved. We suggest that this technique is necessary for all downscaling studies with regional models with domain sizes of a few thousand kilometers and larger embedded in global models.

  17. Combined effect of pulse density and grid cell size on predicting and mapping aboveground carbon in fast-growing Eucalyptus forest plantation using airborne LiDAR data.

    PubMed

    Silva, Carlos Alberto; Hudak, Andrew Thomas; Klauberg, Carine; Vierling, Lee Alexandre; Gonzalez-Benecke, Carlos; de Padua Chaves Carvalho, Samuel; Rodriguez, Luiz Carlos Estraviz; Cardil, Adrián

    2017-12-01

    LiDAR remote sensing is a rapidly evolving technology for quantifying a variety of forest attributes, including aboveground carbon (AGC). Pulse density influences the acquisition cost of LiDAR, and grid cell size influences AGC prediction using plot-based methods; however, little work has evaluated the effects of LiDAR pulse density and cell size for predicting and mapping AGC in fast-growing Eucalyptus forest plantations. The aim of this study was to evaluate the effect of LiDAR pulse density and grid cell size on AGC prediction accuracy at plot and stand-levels using airborne LiDAR and field data. We used the Random Forest (RF) machine learning algorithm to model AGC using LiDAR-derived metrics from LiDAR collections of 5 and 10 pulses m -2 (RF5 and RF10) and grid cell sizes of 5, 10, 15 and 20 m. The results show that LiDAR pulse density of 5 pulses m -2 provides metrics with similar prediction accuracy for AGC as when using a dataset with 10 pulses m -2 in these fast-growing plantations. Relative root mean square errors (RMSEs) for the RF5 and RF10 were 6.14 and 6.01%, respectively. Equivalence tests showed that the predicted AGC from the training and validation models were equivalent to the observed AGC measurements. The grid cell sizes for mapping ranging from 5 to 20 also did not significantly affect the prediction accuracy of AGC at stand level in this system. LiDAR measurements can be used to predict and map AGC across variable-age Eucalyptus plantations with adequate levels of precision and accuracy using 5 pulses m -2 and a grid cell size of 5 m. The promising results for AGC modeling in this study will allow for greater confidence in comparing AGC estimates with varying LiDAR sampling densities for Eucalyptus plantations and assist in decision making towards more cost effective and efficient forest inventory.

  18. Near-Body Grid Adaption for Overset Grids

    NASA Technical Reports Server (NTRS)

    Buning, Pieter G.; Pulliam, Thomas H.

    2016-01-01

    A solution adaption capability for curvilinear near-body grids has been implemented in the OVERFLOW overset grid computational fluid dynamics code. The approach follows closely that used for the Cartesian off-body grids, but inserts refined grids in the computational space of original near-body grids. Refined curvilinear grids are generated using parametric cubic interpolation, with one-sided biasing based on curvature and stretching ratio of the original grid. Sensor functions, grid marking, and solution interpolation tasks are implemented in the same fashion as for off-body grids. A goal-oriented procedure, based on largest error first, is included for controlling growth rate and maximum size of the adapted grid system. The adaption process is almost entirely parallelized using MPI, resulting in a capability suitable for viscous, moving body simulations. Two- and three-dimensional examples are presented.

  19. Uncertainties in planned dose due to the limited voxel size of the planning CT when treating lung tumors with proton therapy

    NASA Astrophysics Data System (ADS)

    España, Samuel; Paganetti, Harald

    2011-07-01

    Dose calculation for lung tumors can be challenging due to the low density and the fine structure of the geometry. The latter is not fully considered in the CT image resolution used in treatment planning causing the prediction of a more homogeneous tissue distribution. In proton therapy, this could result in predicting an unrealistically sharp distal dose falloff, i.e. an underestimation of the distal dose falloff degradation. The goal of this work was the quantification of such effects. Two computational phantoms resembling a two-dimensional heterogeneous random lung geometry and a swine lung were considered applying a variety of voxel sizes for dose calculation. Monte Carlo simulations were used to compare the dose distributions predicted with the voxel size typically used for the treatment planning procedure with those expected to be delivered using the finest resolution. The results show, for example, distal falloff position differences of up to 4 mm between planned and expected dose at the 90% level for the heterogeneous random lung (assuming treatment plan on a 2 × 2 × 2.5 mm3 grid). For the swine lung, differences of up to 38 mm were seen when airways are present in the beam path when the treatment plan was done on a 0.8 × 0.8 × 2.4 mm3 grid. The two-dimensional heterogeneous random lung phantom apparently does not describe the impact of the geometry adequately because of the lack of heterogeneities in the axial direction. The differences observed in the swine lung between planned and expected dose are presumably due to the poor axial resolution of the CT images used in clinical routine. In conclusion, when assigning margins for treatment planning for lung cancer, proton range uncertainties due to the heterogeneous lung geometry and CT image resolution need to be considered.

  20. JPARSS: A Java Parallel Network Package for Grid Computing

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Chen, Jie; Akers, Walter; Chen, Ying

    2002-03-01

    The emergence of high speed wide area networks makes grid computinga reality. However grid applications that need reliable data transfer still have difficulties to achieve optimal TCP performance due to network tuning of TCP window size to improve bandwidth and to reduce latency on a high speed wide area network. This paper presents a Java package called JPARSS (Java Parallel Secure Stream (Socket)) that divides data into partitions that are sent over several parallel Java streams simultaneously and allows Java or Web applications to achieve optimal TCP performance in a grid environment without the necessity of tuning TCP window size.more » This package enables single sign-on, certificate delegation and secure or plain-text data transfer using several security components based on X.509 certificate and SSL. Several experiments will be presented to show that using Java parallelstreams is more effective than tuning TCP window size. In addition a simple architecture using Web services« less

  1. Effects of Stencil Width on Surface Ocean Geostrophic Velocity and Vorticity Estimation from Gridded Satellite Altimeter Data

    DTIC Science & Technology

    2012-03-17

    Texas at Austin, Austin, Texas, USA. g dq ’Departement de Physique and LPO, Universite de Bretagne V _ /" r5r’ Occidental, Brest ...grid points are used in the calculation, so that the grid spacing is 8 times larger than on the original grid. The 3-point stencil differences are sig...that the difference between narrow and wide stencil estimates increases over that found on the original higher resolution grid. Interpolation of the

  2. A grid amplifier

    NASA Technical Reports Server (NTRS)

    Kim, Moonil; Weikle, Robert M., II; Hacker, Jonathan B.; Delisio, Michael P.; Rutledge, David B.; Rosenberg, James J.; Smith, R. P.

    1991-01-01

    A 50-MESFET grid amplifier is reported that has a gain of 11 dB at 3.3 GHz. The grid isolates the input from the output by using vertical polarization for the input beam and horizontal polarization for the transmitted output beam. The grid unit cell is a two-MESFET differential amplifier. A simple calibration procedure allows the gain to be calculated from a relative power measurement. This grid is a hybrid circuit, but the structure is suitable for fabrication as a monolithic wafer-scale integrated circuit, particularly at millimeter wavelengths.

  3. PARADIGM USING JOINT DETERMINISTIC GRID MODELING AND SUB-GRID VARIABILITY STOCHASTIC DESCRIPTION AS A TEMPLATE FOR MODEL EVALUATION

    EPA Science Inventory

    The goal of achieving verisimilitude of air quality simulations to observations is problematic. Chemical transport models such as the Community Multi-Scale Air Quality (CMAQ) modeling system produce volume averages of pollutant concentration fields. When grid sizes are such tha...

  4. Microwave Frequency Polarizers

    NASA Technical Reports Server (NTRS)

    Ha, Vien The; Mirel, Paul; Kogut, Alan J.

    2013-01-01

    This article describes the fabrication and analysis of microwave frequency polarizing grids. The grids are designed to measure polarization from the cosmic microwave background. It is effective in the range of 500 to 1500 micron wavelength. It is cryogenic compatible and highly robust to high load impacts. Each grid is fabricated using an array of different assembly processes which vary in the types of tension mechanisms to the shape and size of the grids. We provide a comprehensive study on the analysis of the grids' wire heights, diameters, and spacing.

  5. Rapid inundation estimates at harbor scale using tsunami wave heights offshore simulation and Green's law approach

    NASA Astrophysics Data System (ADS)

    Gailler, Audrey; Hébert, Hélène; Loevenbruck, Anne

    2013-04-01

    Improvements in the availability of sea-level observations and advances in numerical modeling techniques are increasing the potential for tsunami warnings to be based on numerical model forecasts. Numerical tsunami propagation and inundation models are well developed and have now reached an impressive level of accuracy, especially in locations such as harbors where the tsunami waves are mostly amplified. In the framework of tsunami warning under real-time operational conditions, the main obstacle for the routine use of such numerical simulations remains the slowness of the numerical computation, which is strengthened when detailed grids are required for the precise modeling of the coastline response on the scale of an individual harbor. In fact, when facing the problem of the interaction of the tsunami wavefield with a shoreline, any numerical simulation must be performed over an increasingly fine grid, which in turn mandates a reduced time step, and the use of a fully non-linear code. Such calculations become then prohibitively time-consuming, which is clearly unacceptable in the framework of real-time warning. Thus only tsunami offshore propagation modeling tools using a single sparse bathymetric computation grid are presently included within the French Tsunami Warning Center (CENALT), providing rapid estimation of tsunami wave heights in high seas, and tsunami warning maps at western Mediterranean and NE Atlantic basins scale. We present here a preliminary work that performs quick estimates of the inundation at individual harbors from these deep wave heights simulations. The method involves an empirical correction relation derived from Green's law, expressing conservation of wave energy flux to extend the gridded wave field into the harbor with respect to the nearby deep-water grid node. The main limitation of this method is that its application to a given coastal area would require a large database of previous observations, in order to define the empirical parameters of the correction equation. As no such data (i.e., historical tide gage records of significant tsunamis) are available for the western Mediterranean and NE Atlantic basins, a set of synthetic mareograms is calculated for both fake and well-known historical tsunamigenic earthquakes in the area. This synthetic dataset is obtained through accurate numerical tsunami propagation and inundation modeling by using several nested bathymetric grids characterized by a coarse resolution over deep water regions and an increasingly fine resolution close to the shores (down to a grid cell size of 3m in some Mediterranean harbors). This synthetic dataset is then used to approximate the empirical parameters of the correction equation. Results of inundation estimates in several french Mediterranean harbors obtained with the fast "Green's law - derived" method are presented and compared with values given by time-consuming nested grids simulations.

  6. Morphometric analysis of Russian Plain's small lakes on the base of accurate digital bathymetric models

    NASA Astrophysics Data System (ADS)

    Naumenko, Mikhail; Guzivaty, Vadim; Sapelko, Tatiana

    2016-04-01

    Lake morphometry refers to physical factors (shape, size, structure, etc) that determine the lake depression. Morphology has a great influence on lake ecological characteristics especially on water thermal conditions and mixing depth. Depth analyses, including sediment measurement at various depths, volumes of strata and shoreline characteristics are often critical to the investigation of biological, chemical and physical properties of fresh waters as well as theoretical retention time. Management techniques such as loading capacity for effluents and selective removal of undesirable components of the biota are also dependent on detailed knowledge of the morphometry and flow characteristics. During the recent years a lake bathymetric surveys were carried out by using echo sounder with a high bottom depth resolution and GPS coordinate determination. Few digital bathymetric models have been created with 10*10 m spatial grid for some small lakes of Russian Plain which the areas not exceed 1-2 sq. km. The statistical characteristics of the depth and slopes distribution of these lakes calculated on an equidistant grid. It will provide the level-surface-volume variations of small lakes and reservoirs, calculated through combination of various satellite images. We discuss the methodological aspects of creating of morphometric models of depths and slopes of small lakes as well as the advantages of digital models over traditional methods.

  7. Off-Grid Direction of Arrival Estimation Based on Joint Spatial Sparsity for Distributed Sparse Linear Arrays

    PubMed Central

    Liang, Yujie; Ying, Rendong; Lu, Zhenqi; Liu, Peilin

    2014-01-01

    In the design phase of sensor arrays during array signal processing, the estimation performance and system cost are largely determined by array aperture size. In this article, we address the problem of joint direction-of-arrival (DOA) estimation with distributed sparse linear arrays (SLAs) and propose an off-grid synchronous approach based on distributed compressed sensing to obtain larger array aperture. We focus on the complex source distribution in the practical applications and classify the sources into common and innovation parts according to whether a signal of source can impinge on all the SLAs or a specific one. For each SLA, we construct a corresponding virtual uniform linear array (ULA) to create the relationship of random linear map between the signals respectively observed by these two arrays. The signal ensembles including the common/innovation sources for different SLAs are abstracted as a joint spatial sparsity model. And we use the minimization of concatenated atomic norm via semidefinite programming to solve the problem of joint DOA estimation. Joint calculation of the signals observed by all the SLAs exploits their redundancy caused by the common sources and decreases the requirement of array size. The numerical results illustrate the advantages of the proposed approach. PMID:25420150

  8. Localized strain measurements of the intervertebral disc annulus during biaxial tensile testing.

    PubMed

    Karakolis, Thomas; Callaghan, Jack P

    2015-01-01

    Both inter-lamellar and intra-lamellar failures of the annulus have been described as potential modes of disc herniation. Attempts to characterize initial lamellar failure of the annulus have involved tensile testing of small tissue samples. The purpose of this study was to evaluate a method of measuring local surface strains through image analysis of a tensile test conducted on an isolated sample of annular tissue in order to enhance future studies of intervertebral disc failure. An annulus tissue sample was biaxial strained to 10%. High-resolution images captured the tissue surface throughout testing. Three test conditions were evaluated: submerged, non-submerged and marker. Surface strains were calculated for the two non-marker conditions based on motion of virtual tracking points. Tracking algorithm parameters (grid resolution and template size) were varied to determine the effect on estimated strains. Accuracy of point tracking was assessed through a comparison of the non-marker conditions to a condition involving markers placed on tissue surface. Grid resolution had a larger effect on local strain than template size. Average local strain error ranged from 3% to 9.25% and 0.1% to 2.0%, for the non-submerged and submerged conditions, respectively. Local strain estimation has a relatively high potential for error. Submerging the tissue provided superior strain estimates.

  9. Optimizing the Terzaghi Estimator of the 3D Distribution of Rock Fracture Orientations

    NASA Astrophysics Data System (ADS)

    Tang, Huiming; Huang, Lei; Juang, C. Hsein; Zhang, Junrong

    2017-08-01

    Orientation statistics are prone to bias when surveyed with the scanline mapping technique in which the observed probabilities differ, depending on the intersection angle between the fracture and the scanline. This bias leads to 1D frequency statistical data that are poorly representative of the 3D distribution. A widely accessible estimator named after Terzaghi was developed to estimate 3D frequencies from 1D biased observations, but the estimation accuracy is limited for fractures at narrow intersection angles to scanlines (termed the blind zone). Although numerous works have concentrated on accuracy with respect to the blind zone, accuracy outside the blind zone has rarely been studied. This work contributes to the limited investigations of accuracy outside the blind zone through a qualitative assessment that deploys a mathematical derivation of the Terzaghi equation in conjunction with a quantitative evaluation that uses fractures simulation and verification of natural fractures. The results show that the estimator does not provide a precise estimate of 3D distributions and that the estimation accuracy is correlated with the grid size adopted by the estimator. To explore the potential for improving accuracy, the particular grid size producing maximum accuracy is identified from 168 combinations of grid sizes and two other parameters. The results demonstrate that the 2° × 2° grid size provides maximum accuracy for the estimator in most cases when applied outside the blind zone. However, if the global sample density exceeds 0.5°-2, then maximum accuracy occurs at a grid size of 1° × 1°.

  10. Radioactive Pollution Estimate for Fukushima Nuclear Power Plant by a Particle Model

    NASA Astrophysics Data System (ADS)

    Saito, Keisuke; Ogawa, Susumu

    2016-06-01

    On Mar 12, 2011, very wide radioactive pollution occurred by a hydrogen explosion in Fukushima Nuclear Power Plant. A large amount of radioisotopes started with four times of explosions. With traditional atmospheric diffusion models could not reconstruct radioactive pollution in Fukushima. Then, with a particle model, this accident was reconstructed from meteorological archive and Radar- AMeDAS. Calculations with the particle model were carried out for Mar 12, 15, 18 and 20 when east southeast winds blew for five hours continuously. Meteorological archive is expressed by wind speeds and directions in five-km grid every hour with eight classes of height till 3000 m. Radar- AMeDAS is precipitation data in one-km grid every thirty minutes. Particles are ten scales of 0.01 to 0.1 mm in diameter with specific weight of 2.65 and vertical speeds given by Stokes equation. But, on Mar 15, it rained from 16:30 and then the particles fell down at a moment as wet deposit in calculation. On the other hand, the altitudes on the ground were given by DEM with 1 km-grid. The spatial dose by emitted radioisotopes was referred to the observation data at monitoring posts of Tokyo Electric Power Company. The falling points of radioisotopes were expressed on the map using the particle model. As a result, the same distributions were obtained as the surface spatial dose of radioisotopes in aero-monitoring by Ministry of Education, Culture, Sports, Science and Technology. Especially, on Mar 15, the simulated pollution fitted to the observation, which extended to the northwest of Fukushima Daiichi Nuclear Power Plant and caused mainly sever pollution. By the particle model, the falling positions on the ground were estimated each particle size. Particles with more than 0.05 mm of size were affected by the topography and blocked by the mountains with the altitudes of more than 700 m. The particle model does not include the atmospheric stability, the source height, and deposit speeds. The present assignment is how to express the difference of deposition each nucleus.

  11. Fabrication and characterization of a deep ultraviolet wire grid polarizer with a chromium-oxide subwavelength grating.

    PubMed

    Asano, Kosuke; Yokoyama, Satoshi; Kemmochi, Atsushi; Yatagai, Toyohiko

    2014-05-01

    A wire grid polarizer comprised of chromium oxide is designed for a micro-lithography system using an ArF excimer laser. Optical properties for some material candidates are calculated using a rigorous coupled-wave analysis. The chromium oxide wire grid polarizer with a 90 nm period is fabricated by a double-patterning technique using KrF lithography and dry etching. The extinction ratio of the grating is greater than 20 dB (100:1) at a wavelength of 193 nm. Differences between the calculated and experimental results are discussed.

  12. Analytical Computation of Effective Grid Parameters for the Finite-Difference Seismic Waveform Modeling With the PREM, IASP91, SP6, and AK135

    NASA Astrophysics Data System (ADS)

    Toyokuni, G.; Takenaka, H.

    2007-12-01

    We propose a method to obtain effective grid parameters for the finite-difference (FD) method with standard Earth models using analytical ways. In spite of the broad use of the heterogeneous FD formulation for seismic waveform modeling, accurate treatment of material discontinuities inside the grid cells has been a serious problem for many years. One possible way to solve this problem is to introduce effective grid elastic moduli and densities (effective parameters) calculated by the volume harmonic averaging of elastic moduli and volume arithmetic averaging of density in grid cells. This scheme enables us to put a material discontinuity into an arbitrary position in the spatial grids. Most of the methods used for synthetic seismogram calculation today receives the blessing of the standard Earth models, such as the PREM, IASP91, SP6, and AK135, represented as functions of normalized radius. For the FD computation of seismic waveform with such models, we first need accurate treatment of material discontinuities in radius. This study provides a numerical scheme for analytical calculations of the effective parameters for an arbitrary spatial grids in radial direction as to these major four standard Earth models making the best use of their functional features. This scheme can analytically obtain the integral volume averages through partial fraction decompositions (PFDs) and integral formulae. We have developed a FORTRAN subroutine to perform the computations, which is opened to utilization in a large variety of FD schemes ranging from 1-D to 3-D, with conventional- and staggered-grids. In the presentation, we show some numerical examples displaying the accuracy of the FD synthetics simulated with the analytical effective parameters.

  13. Multi-off-grid methods in multi-step integration of ordinary differential equations

    NASA Technical Reports Server (NTRS)

    Beaudet, P. R.

    1974-01-01

    Description of methods of solving first- and second-order systems of differential equations in which all derivatives are evaluated at off-grid locations in order to circumvent the Dahlquist stability limitation on the order of on-grid methods. The proposed multi-off-grid methods require off-grid state predictors for the evaluation of the n derivatives at each step. Progressing forward in time, the off-grid states are predicted using a linear combination of back on-grid state values and off-grid derivative evaluations. A comparison is made between the proposed multi-off-grid methods and the corresponding Adams and Cowell on-grid integration techniques in integrating systems of ordinary differential equations, showing a significant reduction in the error at larger step sizes in the case of the multi-off-grid integrator.

  14. An efficient basis set representation for calculating electrons in molecules

    DOE PAGES

    Jones, Jeremiah R.; Rouet, Francois -Henry; Lawler, Keith V.; ...

    2016-04-27

    The method of McCurdy, Baertschy, and Rescigno, is generalised to obtain a straightforward, surprisingly accurate, and scalable numerical representation for calculating the electronic wave functions of molecules. It uses a basis set of product sinc functions arrayed on a Cartesian grid, and yields 1 kcal/mol precision for valence transition energies with a grid resolution of approximately 0.1 bohr. The Coulomb matrix elements are replaced with matrix elements obtained from the kinetic energy operator. A resolution-of-the-identity approximation renders the primitive one- and two-electron matrix elements diagonal; in other words, the Coulomb operator is local with respect to the grid indices. Themore » calculation of contracted two-electron matrix elements among orbitals requires only O( Nlog (N)) multiplication operations, not O( N 4), where N is the number of basis functions; N = n 3 on cubic grids. The representation not only is numerically expedient, but also produces energies and properties superior to those calculated variationally. Absolute energies, absorption cross sections, transition energies, and ionisation potentials are reported for 1- (He +, H + 2), 2- (H 2, He), 10- (CH 4), and 56-electron (C 8H 8) systems.« less

  15. Program Calculates Forces in Bolted Structural Joints

    NASA Technical Reports Server (NTRS)

    Buder, Daniel A.

    2005-01-01

    FORTRAN 77 computer program calculates forces in bolts in the joints of structures. This program is used in conjunction with the NASTRAN finite-element structural-analysis program. A mathematical model of a structure is first created by approximating its load-bearing members with representative finite elements, then NASTRAN calculates the forces and moments that each finite element contributes to grid points located throughout the structure. The user selects the finite elements that correspond to structural members that contribute loads to the joints of interest, and identifies the grid point nearest to each such joint. This program reads the pertinent NASTRAN output, combines the forces and moments from the contributing elements to determine the resultant force and moment acting at each proximate grid point, then transforms the forces and moments from these grid points to the centroids of the affected joints. Then the program uses these joint loads to obtain the axial and shear forces in the individual bolts. The program identifies which bolts bear the greatest axial and/or shear loads. The program also performs a fail-safe analysis in which the foregoing calculations are repeated for a sequence of cases in which each fastener, in turn, is assumed not to transmit an axial force.

  16. Horizontal Residual Mean Circulation: Evaluation of Spatial Correlations in Coarse Resolution Ocean Models

    NASA Astrophysics Data System (ADS)

    Li, Y.; McDougall, T. J.

    2016-02-01

    Coarse resolution ocean models lack knowledge of spatial correlations between variables on scales smaller than the grid scale. Some researchers have shown that these spatial correlations play a role in the poleward heat flux. In order to evaluate the poleward transport induced by the spatial correlations at a fixed horizontal position, an equation is obtained to calculate the approximate transport from velocity gradients. The equation involves two terms that can be added to the quasi-Stokes streamfunction (based on temporal correlations) to incorporate the contribution of spatial correlations. Moreover, these new terms do not need to be parameterized and is ready to be evaluated by using model data directly. In this study, data from a high resolution ocean model have been used to estimate the accuracy of this HRM approach for improving the horizontal property fluxes in coarse-resolution ocean models. A coarse grid is formed by sub-sampling and box-car averaging the fine grid scale. The transport calculated on the coarse grid is then compared to the transport on original high resolution grid scale accumulated over a corresponding number of grid boxes. The preliminary results have shown that the estimate on coarse resolution grids roughly match the corresponding transports on high resolution grids.

  17. Satellite-derived ice data sets no. 2: Arctic monthly average microwave brightness temperatures and sea ice concentrations, 1973-1976

    NASA Technical Reports Server (NTRS)

    Parkinson, C. L.; Comiso, J. C.; Zwally, H. J.

    1987-01-01

    A summary data set for four years (mid 70's) of Arctic sea ice conditions is available on magnetic tape. The data include monthly and yearly averaged Nimbus 5 electrically scanning microwave radiometer (ESMR) brightness temperatures, an ice concentration parameter derived from the brightness temperatures, monthly climatological surface air temperatures, and monthly climatological sea level pressures. All data matrices are applied to 293 by 293 grids that cover a polar stereographic map enclosing the 50 deg N latitude circle. The grid size varies from about 32 X 32 km at the poles to about 28 X 28 km at 50 deg N. The ice concentration parameter is calculated assuming that the field of view contains only open water and first-year ice with an ice emissivity of 0.92. To account for the presence of multiyear ice, a nomogram is provided relating the ice concentration parameter, the total ice concentration, and the fraction of the ice cover which is multiyear ice.

  18. Grid removal and impact on population dose in full-field digital mammography.

    PubMed

    Gennaro, Gisella; Katz, Luc; Souchay, Henri; Klausz, Remy; Alberelli, Claudio; di Maggio, Cosimo

    2007-02-01

    The study purpose was to determine the impact of anti-scatter grid removal on patient dose, in full field digital mammography. Dose saving, phantom based, was evaluated with the constraint that images acquired with and without grid would provide the same contrast-to-noise ratio (CNR). The digital equipment employed a flat panel detector with cesium iodide for x-ray to light conversion, 100 microm pixel size; the x-ray source was a dual-track tube with selectable filtration. Poly(methyl-emathocrylate) (PMMA) layers in the range 20-70 mm were used to simulate the absorption of different breast thickness, while two Al foils, 0.1 and 0.2 mm thick were used to provide a certain CNR. Images with grid were acquired with the same beam quality as selected in full automatic exposure mode and the mAs levels as close as possible, and the CNR measured for each thickness between 20 and 70 mm. Phantom images without grid were acquired in manual exposure mode, by selecting the same anode/filter combination and kVp as the image with grid at the same thickness, but varying mAs from 10 to 200. For each thickness, an image without aluminum was acquired for each mAs value, in order to obtain a flat image to be used to subtract the scatter nonuniformity from the phantom images. After scatter subtraction, the CNR was measured on images without grid. The mAs value that should be set to acquire a phantom image without grid so that it has the same CNR as the corresponding grid image was calculated. Therefore, mAs reduction percentage was determined versus phantom thickness. Results showed that dose saving was lower than 30% for PMMA equivalent breast thinner than 40 mm, decreased below 10% for intermediate thickness (45-50 mm), but there was no dose gain for thickness beyond 60 mm. By applying the mAs reduction factors to a clinical population derived from a data base of 4622 breasts, dose benefit was quantified in terms of population dose. On the average, the overall dose reduction was about 8%. It was considered small, not sufficient to justify a clinical implementation, and the anti-scatter grid was maintained.

  19. Recent Developments in Grid Generation and Force Integration Technology for Overset Grids

    NASA Technical Reports Server (NTRS)

    Chan, William M.; VanDalsem, William R. (Technical Monitor)

    1994-01-01

    Recent developments in algorithms and software tools for generating overset grids for complex configurations are described. These include the overset surface grid generation code SURGRD and version 2.0 of the hyperbolic volume grid generation code HYPGEN. The SURGRD code is in beta test mode where the new features include the capability to march over a collection of panel networks, a variety of ways to control the side boundaries and the marching step sizes and distance, a more robust projection scheme and an interpolation option. New features in version 2.0 of HYPGEN include a wider range of boundary condition types. The code also allows the user to specify different marching step sizes and distance for each point on the surface grid. A scheme that takes into account of the overlapped zones on the body surface for the purpose of forces and moments computation is also briefly described, The process involves the following two software modules: MIXSUR - a composite grid generation module to produce a collection of quadrilaterals and triangles on which pressure and viscous stresses are to be integrated, and OVERINT - a forces and moments integration module.

  20. Optimal moving grids for time-dependent partial differential equations

    NASA Technical Reports Server (NTRS)

    Wathen, A. J.

    1989-01-01

    Various adaptive moving grid techniques for the numerical solution of time-dependent partial differential equations were proposed. The precise criterion for grid motion varies, but most techniques will attempt to give grids on which the solution of the partial differential equation can be well represented. Moving grids are investigated on which the solutions of the linear heat conduction and viscous Burgers' equation in one space dimension are optimally approximated. Precisely, the results of numerical calculations of optimal moving grids for piecewise linear finite element approximation of partial differential equation solutions in the least squares norm.

  1. Optimal moving grids for time-dependent partial differential equations

    NASA Technical Reports Server (NTRS)

    Wathen, A. J.

    1992-01-01

    Various adaptive moving grid techniques for the numerical solution of time-dependent partial differential equations were proposed. The precise criterion for grid motion varies, but most techniques will attempt to give grids on which the solution of the partial differential equation can be well represented. Moving grids are investigated on which the solutions of the linear heat conduction and viscous Burgers' equation in one space dimension are optimally approximated. Precisely, the results of numerical calculations of optimal moving grids for piecewise linear finite element approximation of PDE solutions in the least-squares norm are reported.

  2. Efficient Load Balancing and Data Remapping for Adaptive Grid Calculations

    NASA Technical Reports Server (NTRS)

    Oliker, Leonid; Biswas, Rupak

    1997-01-01

    Mesh adaption is a powerful tool for efficient unstructured- grid computations but causes load imbalance among processors on a parallel machine. We present a novel method to dynamically balance the processor workloads with a global view. This paper presents, for the first time, the implementation and integration of all major components within our dynamic load balancing strategy for adaptive grid calculations. Mesh adaption, repartitioning, processor assignment, and remapping are critical components of the framework that must be accomplished rapidly and efficiently so as not to cause a significant overhead to the numerical simulation. Previous results indicated that mesh repartitioning and data remapping are potential bottlenecks for performing large-scale scientific calculations. We resolve these issues and demonstrate that our framework remains viable on a large number of processors.

  3. Recommended GIS Analysis Methods for Global Gridded Population Data

    NASA Astrophysics Data System (ADS)

    Frye, C. E.; Sorichetta, A.; Rose, A.

    2017-12-01

    When using geographic information systems (GIS) to analyze gridded, i.e., raster, population data, analysts need a detailed understanding of several factors that affect raster data processing, and thus, the accuracy of the results. Global raster data is most often provided in an unprojected state, usually in the WGS 1984 geographic coordinate system. Most GIS functions and tools evaluate data based on overlay relationships (area) or proximity (distance). Area and distance for global raster data can be either calculated directly using the various earth ellipsoids or after transforming the data to equal-area/equidistant projected coordinate systems to analyze all locations equally. However, unlike when projecting vector data, not all projected coordinate systems can support such analyses equally, and the process of transforming raster data from one coordinate space to another often results unmanaged loss of data through a process called resampling. Resampling determines which values to use in the result dataset given an imperfect locational match in the input dataset(s). Cell size or resolution, registration, resampling method, statistical type, and whether the raster represents continuous or discreet information potentially influence the quality of the result. Gridded population data represent estimates of population in each raster cell, and this presentation will provide guidelines for accurately transforming population rasters for analysis in GIS. Resampling impacts the display of high resolution global gridded population data, and we will discuss how to properly handle pyramid creation using the Aggregate tool with the sum option to create overviews for mosaic datasets.

  4. Computation of high Reynolds number internal/external flows

    NASA Technical Reports Server (NTRS)

    Cline, M. C.; Wilmoth, R. G.

    1981-01-01

    A general, user oriented computer program, called VNAP2, has been developed to calculate high Reynolds number, internal/external flows. VNAP2 solves the two-dimensional, time-dependent Navier-Stokes equations. The turbulence is modeled with either a mixing-length, a one transport equation, or a two transport equation model. Interior grid points are computed using the explicit MacCormack scheme with special procedures to speed up the calculation in the fine grid. All boundary conditions are calculated using a reference plane characteristic scheme with the viscous terms treated as source terms. Several internal, and internal/external flow calculations are presented.

  5. Computation of high Reynolds number internal/external flows

    NASA Technical Reports Server (NTRS)

    Cline, M. C.; Wilmoth, R. G.

    1981-01-01

    A general, user oriented computer program, called VNAP2, was developed to calculate high Reynolds number, internal/ external flows. The VNAP2 program solves the two dimensional, time dependent Navier-Stokes equations. The turbulence is modeled with either a mixing-length, a one transport equation, or a two transport equation model. Interior grid points are computed using the explicit MacCormack Scheme with special procedures to speed up the calculation in the fine grid. All boundary conditions are calculated using a reference plane characteristic scheme with the viscous terms treated as source terms. Several internal, external, and internal/external flow calculations are presented.

  6. Computation of high Reynolds number internal/external flows

    NASA Technical Reports Server (NTRS)

    Cline, M. C.; Wilmoth, R. G.

    1981-01-01

    A general, user oriented computer program, called VNAF2, developed to calculate high Reynolds number internal/external flows is described. The program solves the two dimensional, time dependent Navier-Stokes equations. Turbulence is modeled with either a mixing length, a one transport equation, or a two transport equation model. Interior grid points are computed using the explicit MacCormack scheme with special procedures to speed up the calculation in the fine grid. All boundary conditions are calculated using a reference plane characteristic scheme with the viscous terms treated as source terms. Several internal, external, and internal/external flow calculations are presented.

  7. Residential scene classification for gridded population sampling in developing countries using deep convolutional neural networks on satellite imagery.

    PubMed

    Chew, Robert F; Amer, Safaa; Jones, Kasey; Unangst, Jennifer; Cajka, James; Allpress, Justine; Bruhn, Mark

    2018-05-09

    Conducting surveys in low- and middle-income countries is often challenging because many areas lack a complete sampling frame, have outdated census information, or have limited data available for designing and selecting a representative sample. Geosampling is a probability-based, gridded population sampling method that addresses some of these issues by using geographic information system (GIS) tools to create logistically manageable area units for sampling. GIS grid cells are overlaid to partition a country's existing administrative boundaries into area units that vary in size from 50 m × 50 m to 150 m × 150 m. To avoid sending interviewers to unoccupied areas, researchers manually classify grid cells as "residential" or "nonresidential" through visual inspection of aerial images. "Nonresidential" units are then excluded from sampling and data collection. This process of manually classifying sampling units has drawbacks since it is labor intensive, prone to human error, and creates the need for simplifying assumptions during calculation of design-based sampling weights. In this paper, we discuss the development of a deep learning classification model to predict whether aerial images are residential or nonresidential, thus reducing manual labor and eliminating the need for simplifying assumptions. On our test sets, the model performs comparable to a human-level baseline in both Nigeria (94.5% accuracy) and Guatemala (96.4% accuracy), and outperforms baseline machine learning models trained on crowdsourced or remote-sensed geospatial features. Additionally, our findings suggest that this approach can work well in new areas with relatively modest amounts of training data. Gridded population sampling methods like geosampling are becoming increasingly popular in countries with outdated or inaccurate census data because of their timeliness, flexibility, and cost. Using deep learning models directly on satellite images, we provide a novel method for sample frame construction that identifies residential gridded aerial units. In cases where manual classification of satellite images is used to (1) correct for errors in gridded population data sets or (2) classify grids where population estimates are unavailable, this methodology can help reduce annotation burden with comparable quality to human analysts.

  8. User's Manual for FOMOCO Utilities-Force and Moment Computation Tools for Overset Grids

    NASA Technical Reports Server (NTRS)

    Chan, William M.; Buning, Pieter G.

    1996-01-01

    In the numerical computations of flows around complex configurations, accurate calculations of force and moment coefficients for aerodynamic surfaces are required. When overset grid methods are used, the surfaces on which force and moment coefficients are sought typically consist of a collection of overlapping surface grids. Direct integration of flow quantities on the overlapping grids would result in the overlapped regions being counted more than once. The FOMOCO Utilities is a software package for computing flow coefficients (force, moment, and mass flow rate) on a collection of overset surfaces with accurate accounting of the overlapped zones. FOMOCO Utilities can be used in stand-alone mode or in conjunction with the Chimera overset grid compressible Navier-Stokes flow solver OVERFLOW. The software package consists of two modules corresponding to a two-step procedure: (1) hybrid surface grid generation (MIXSUR module), and (2) flow quantities integration (OVERINT module). Instructions on how to use this software package are described in this user's manual. Equations used in the flow coefficients calculation are given in Appendix A.

  9. The Impacts of Bowtie Effect and View Angle Discontinuity on MODIS Swath Data Gridding

    NASA Technical Reports Server (NTRS)

    Wang, Yujie; Lyapustin, Alexei

    2007-01-01

    We have analyzed two effects of the MODIS viewing geometry on the quality of gridded imagery. First, the fact that the MODIS scans a swath of the Earth 10 km wide at nadir, causes abrupt change of the view azimuth angle at the boundary of adjacent scans. This discontinuity appears as striping of the image clearly visible in certain cases with viewing geometry close to principle plane over the snow of the glint area of water. The striping is a true surface Bi-directional Reflectance Factor (BRF) effect and should be preserved during gridding. Second, due to bowtie effect, the observations in adjacent scans overlap each other. Commonly used method of calculating grid cell value by averaging all overlapping observations may result in smearing of the image. This paper describes a refined gridding algorithm that takes the above two effects into account. By calculating the grid cell value by averaging the overlapping observations from a single scan, the new algorithm preserves the measured BRF signal and enhances sharpness of the image.

  10. DPW-VI Results Using FUN3D with Focus on k-kL-MEAH2015 (k-kL) Turbulence Model

    NASA Technical Reports Server (NTRS)

    Abdol-Hamid, K. S.; Carlson, Jan-Renee; Rumsey, Christopher L.; Lee-Rausch, Elizabeth M.; Park, Michael A.

    2017-01-01

    The Common Research Model wing-body configuration is investigated with the k-kL-MEAH2015 turbulence model implemented in FUN3D. This includes results presented at the Sixth Drag Prediction Workshop and additional results generated after the workshop with a nonlinear Quadratic Constitutive Relation (QCR) variant of the same turbulence model. The workshop provided grids are used, and a uniform grid refinement study is performed at the design condition. A large variation between results with and without a reconstruction limiter is exhibited on "medium" grid sizes, indicating that the medium grid size is too coarse for drawing conclusions in comparison with experiment. This variation is reduced with grid refinement. At a fixed angle of attack near design conditions, the QCR variant yielded decreased lift and drag compared with the linear eddy-viscosity model by an amount that was approximately constant with grid refinement. The k-kL-MEAH2015 turbulence model produced wing root junction flow behavior consistent with wind tunnel observations.

  11. Modeling Potential Tephra Dispersal at Yucca Mountain, Nevada

    NASA Astrophysics Data System (ADS)

    Hooper, D.; Franklin, N.; Adams, N.; Basu, D.

    2006-12-01

    Quaternary basaltic volcanoes exist within 20 km [12 mi] of the potential radioactive waste repository at Yucca Mountain, Nevada, and future basaltic volcanism at the repository is considered a low-probability, potentially high-consequence event. If radioactive waste was entrained in the conduit of a future volcanic event, tephra and waste could be transported in the resulting eruption plume. During an eruption, basaltic tephra would be dispersed primarily according to the height of the eruption column, particle-size distribution, and structure of the winds aloft. Following an eruption, contaminated tephra-fall deposits would be affected by surface redistribution processes. The Center for Nuclear Waste Regulatory Analyses developed the computer code TEPHRA to calculate atmospheric dispersion and subsequent deposition of tephra and spent nuclear fuel from a potential eruption at Yucca Mountain and to help prepare the U.S. Nuclear Regulatory Commission to review a potential U.S. Department of Energy license application. The TEPHRA transport code uses the Suzuki model to simulate the thermo-fluid dynamics of atmospheric tephra dispersion. TEPHRA models the transport of airborne pyroclasts based on particle diffusion from an eruption column, horizontal diffusion of particles by atmospheric and plume turbulence, horizontal advection by atmospheric circulation, and particle settling by gravity. More recently, TEPHRA was modified to calculate potential tephra deposit distributions using stratified wind fields based on upper atmosphere data from the Nevada Test Site. Wind data are binned into 1-km [0.62-mi]-high intervals with coupled distributions of wind speed and direction produced for each interval. Using this stratified wind field and discretization with respect to height, TEPHRA calculates particle fall and lateral displacement for each interval. This implementation permits modeling of split wind fields. We use a parallel version of the code to calculate expected tephra and high-level waste accumulation at specified points on a two-dimensional spatial grid, thereby simulating a three- dimensional initial deposit. To assess subsequent tephra and high-level waste redistribution and resuspension, modeling grids were devised to measure deposition in eolian and fluvial source regions. The eolian grid covers an area of 2,600 km2 [1,000 mi2] and the fluvial grid encompasses 318 km2 [123 mi2] of the southernmost portion of the Fortymile Wash catchment basin. Because each realization is independent, distributions of tephra and high-level waste reflect anticipated variations in source-term and transport characteristics. This abstract is an independent product of the Center for Nuclear Waste Regulatory Analyses and does not necessarily reflect the view or regulatory position of the U.S. Nuclear Regulatory Commission.

  12. Population dynamics of Microtus pennsylvanicus in corridor-linked patches

    USGS Publications Warehouse

    Coffman, C.J.; Nichols, J.D.; Pollock, K.H.

    2001-01-01

    Corridors have become a key issue in the discussion of conservation planning: however, few empirical data exist on the use of corridors and their effects on population dynamics. The objective of this replicated, population level, capture-re-capture experiment on meadow voles was to estimate and compare population characteristics of voles between (1) corridor-linked fragments, (2) isolated or non-linked fragments, and (3) unfragmented areas. We conducted two field experiments involving 22600 captures of 5700 individuals. In the first, the maintained corridor study, corridors were maintained at the time of fragmentation, and in the second, the constructed corridor study, we constructed corridors between patches that had been fragmented for some period of time. We applied multistate capture-recapture models with the robust design to estimate adult movement and survival rates, population size, temporal variation in population size, recruitment, and juvenile survival rates. Movement rates increased to a greater extent on constructed corridor-linked grids than on the unfragmented or non-linked fragmented grids between the pre- and post-treatment periods. We found significant differences in local survival on the treated (corridor-linked) grids compared to survival on the fragmented and unfragmented grids between the pre- and post-treatment periods. We found no clear pattern of treatment effects on population size or recruitment in either study. However, in both studies, we found that unfragmented grids were more stable than the fragmented grids based on lower temporal variability in population size. To our knowledge, this is the first experimental study demonstrating that corridors constructed between existing fragmented populations can indeed cause increases in movement and associated changes in demography, supporting the use of constructed corridors for this purpose in conservation biology.

  13. The influence of model resolution on ozone in industrial volatile organic compound plumes.

    PubMed

    Henderson, Barron H; Jeffries, Harvey E; Kim, Byeong-Uk; Vizuete, William G

    2010-09-01

    Regions with concentrated petrochemical industrial activity (e.g., Houston or Baton Rouge) frequently experience large, localized releases of volatile organic compounds (VOCs). Aircraft measurements suggest these released VOCs create plumes with ozone (O3) production rates 2-5 times higher than typical urban conditions. Modeling studies found that simulating high O3 productions requires superfine (1-km) horizontal grid cell size. Compared with fine modeling (4-kmin), the superfine resolution increases the peak O3 concentration by as much as 46%. To understand this drastic O3 change, this study quantifies model processes for O3 and "odd oxygen" (Ox) in both resolutions. For the entire plume, the superfine resolution increases the maximum O3 concentration 3% but only decreases the maximum Ox concentration 0.2%. The two grid sizes produce approximately equal Ox mass but by different reaction pathways. Derived sensitivity to oxides of nitrogen (NOx) and VOC emissions suggests resolution-specific sensitivity to NOx and VOC emissions. Different sensitivity to emissions will result in different O3 responses to subsequently encountered emissions (within the city or downwind). Sensitivity of O3 to emission changes also results in different simulated O3 responses to the same control strategies. Sensitivity of O3 to NOx and VOC emission changes is attributed to finer resolved Eulerian grid and finer resolved NOx emissions. Urban NOx concentration gradients are often caused by roadway mobile sources that would not typically be addressed with Plume-in-Grid models. This study shows that grid cell size (an artifact of modeling) influences simulated control strategies and could bias regulatory decisions. Understanding the dynamics of VOC plume dependence on grid size is the first step toward providing more detailed guidance for resolution. These results underscore VOC and NOx resolution interdependencies best addressed by finer resolution. On the basis of these results, the authors suggest a need for quantitative metrics for horizontal grid resolution in future model guidance.

  14. Application of FUN3D Solver for Aeroacoustics Simulation of a Nose Landing Gear Configuration

    NASA Technical Reports Server (NTRS)

    Vatsa, Veer N.; Lockard, David P.; Khorrami, Mehdi R.

    2011-01-01

    Numerical simulations have been performed for a nose landing gear configuration corresponding to the experimental tests conducted in the Basic Aerodynamic Research Tunnel at NASA Langley Research Center. A widely used unstructured grid code, FUN3D, is examined for solving the unsteady flow field associated with this configuration. A series of successively finer unstructured grids has been generated to assess the effect of grid refinement. Solutions have been obtained on purely tetrahedral grids as well as mixed element grids using hybrid RANS/LES turbulence models. The agreement of FUN3D solutions with experimental data on the same size mesh is better on mixed element grids compared to pure tetrahedral grids, and in general improves with grid refinement.

  15. Vascularized networks with two optimized channel sizes

    NASA Astrophysics Data System (ADS)

    Wang, K.-M.; Lorente, S.; Bejan, A.

    2006-07-01

    This paper reports the development of optimal vascularization for supplying self-healing smart materials with liquid that fills and seals the cracks that may occur throughout their volume. The vascularization consists of two-dimensional grids of interconnected orthogonal channels with two hydraulic diameters (D1, D2). The smallest square loop is designed to match the size (d) of the smallest crack. The network is sealed with respect to the outside and is filled with pressurized liquid. In this work, the crack site is modelled as a small spherical volume of diameter d. When a crack is formed, fluid flows from neighbouring channels to the crack site. This volume-to-point flow is optimized using two formulations: (1) incompressible liquid from steady constant-strength sources located in every node of the grid and from sources located equidistantly on the perimeter of the vascularized body of length scale L and (2) slightly compressible liquid from an initially pressurized grid discharging in time-dependent fashion into one crack site. The flow in every channel is laminar and fully developed. The objectives are (a) to minimize the global resistance to the flow from the grid to the crack site and (b) to minimize the time of discharge from the pressurized grid to the crack site. It is shown that methods (a) and (b) yield similar results. There is an optimal ratio of channel diameters D2/D1 < 1, and it decreases as the grid fineness (L/d) increases. The global flow resistance of the grid with optimized ratio of diameters is approximately half of the resistance of the corresponding grid with one channel size (D1 = D2). The optimized ratio of diameters and the minimized global resistance depend on how the grid intersects the crack site: this effect is minor and stresses the robustness of the vascularized design.

  16. Progress in Visualizing Atomic Size Effects with DFT-Chemical Pressure Analysis: From Isolated Atoms to Trends in AB5 Intermetallics.

    PubMed

    Berns, Veronica M; Engelkemier, Joshua; Guo, Yiming; Kilduff, Brandon J; Fredrickson, Daniel C

    2014-08-12

    The notion of atomic size poses an important challenge to chemical theory: empirical evidence has long established that atoms have spatial requirements, which are summarized in tables of covalent, ionic, metallic, and van der Waals radii. Considerations based on these radii play a central role in the design and interpretation of experiments, but few methods are available to directly support arguments based on atomic size using electronic structure methods. Recently, we described an approach to elucidating atomic size effects using theoretical calculations: the DFT-Chemical Pressure analysis, which visualizes the local pressures arising in crystal structures from the interactions of atomic size and electronic effects. Using this approach, a variety of structural phenomena in intermetallic phases have already been understood in terms that provide guidance to new synthetic experiments. However, the applicability of the DFT-CP method to the broad range of the structures encountered in the solid state is limited by two issues: (1) the difficulty of interpreting the intense pressure features that appear in atomic core regions and (2) the need to divide space among pairs of interacting atoms in a meaningful way. In this article, we describe general solutions to these issues. In addressing the first issue, we explore the CP analysis of a test case in which no core pressures would be expected to arise: isolated atoms in large boxes. Our calculations reveal that intense core pressures do indeed arise in these virtually pressure-less model systems and allow us to trace the issue to the shifts in the voxel positions relative to atomic centers upon expanding and contracting the unit cell. A compensatory grid unwarping procedure is introduced to remedy this artifact. The second issue revolves around the difficulty of interpreting the pressure map in terms of interatomic interactions in a way that respects the size differences of the atoms and avoids artificial geometrical constraints. In approaching this challenge, we have developed a scheme for allocating the grid pressures to contacts inspired by the Hirshfeld charge analysis. Here, each voxel is allocated to the contact between the two atoms whose free atom electron densities show the largest values at that position. In this way, the differing sizes of atoms are naturally included in the division of space without resorting to empirical radii. The use of the improved DFT-CP method is illustrated through analyses of the applicability of radius ratio arguments to Laves phase structures and the structural preferences of AB5 intermetallics between the CaCu5 and AuBe5 structure types.

  17. West Antarctic Balance Fluxes: Impact of Smoothing, Algorithm and Topography.

    NASA Astrophysics Data System (ADS)

    Le Brocq, A.; Payne, A. J.; Siegert, M. J.; Bamber, J. L.

    2004-12-01

    Grid-based calculations of balance flux and velocity have been widely used to understand the large-scale dynamics of ice masses and as indicators of their state of balance. This research investigates a number of issues relating to their calculation for the West Antarctic Ice Sheet (see below for further details): 1) different topography smoothing techniques; 2) different grid based flow-apportioning algorithms; 3) the source of the flow direction, whether from smoothed topography, or smoothed gravitational driving stress; 4) different flux routing techniques and 5) the impact of different topographic datasets. The different algorithms described below lead to significant differences in both ice stream margins and values of fluxes within them. This encourages caution in the use of grid-based balance flux/velocity distributions and values, especially when considering the state of balance of individual ice streams. 1) Most previous calculations have used the same numerical scheme (Budd and Warner, 1996) applied to a smoothed topography in order to incorporate the longitudinal stresses that smooth ice flow. There are two options to consider when smoothing the topography, the size of the averaging filter and the shape of the averaging function. However, this is not a physically-based approach to incorporating smoothed ice flow and also introduces significant flow artefacts when using a variable weighting function. 2) Different algorithms to apportion flow are investigated; using 4 or 8 neighbours, and apportioning flow to all down-slope cells or only 2 (based on derived flow direction). 3) A theoretically more acceptable approach of incorporating smoothed ice flow is to use the smoothed gravitational driving stress in x and y components to derive a flow direction. The flux can then be apportioned using the flow direction approach used above. 4) The original scheme (Budd and Warner, 1996) uses an elevation sort technique to calculate the balance flux contribution from all cells to each individual cell. However, elevation sort is only successful when ice cannot flow uphill. Other possible techniques include using a recursive call for each neighbour or using a sparse matrix solution. 5) Two digital elevation models are used as input data, which have significant differences in coastal and mountainous areas and therefore lead to different calculations. Of particular interest is the difference in the Rutford Ice Stream/Carlson Inlet and Kamb Ice Stream (Ice Stream C) fluxes.

  18. Effects of downscaled high-resolution meteorological data on the PSCF identification of emission sources

    DOE PAGES

    Cheng, Meng -Dawn; Kabela, Erik D.

    2016-04-30

    The Potential Source Contribution Function (PSCF) model has been successfully used for identifying regions of emission source at a long distance in this study, the PSCF model relies on backward trajectories calculated by the Hybrid Single-Particle Lagrangian Integrated Trajectory (HYSPLIT) model. In this study, we investigated the impacts of grid resolution and Planetary Boundary Layer (PBL) parameterization (e.g., turbulent transport of pollutants) on the PSCF analysis. The Mellor-Yamada-Janjic (MYJ) and Yonsei University (YUS) parameterization schemes were selected to model the turbulent transport in the PBL within the Weather Research and Forecasting (WRF version 3.6) model. Two separate domain grid sizesmore » (83 and 27 km) were chosen in the WRF downscaling in generating the wind data for driving the HYSPLIT calculation. The effects of grid size and PBL parameterization are important in incorporating the influ- ence of regional and local meteorological processes such as jet streaks, blocking patterns, Rossby waves, and terrain-induced convection on the transport of pollutants by a wind trajectory. We found high resolution PSCF did discover and locate source areas more precisely than that with lower resolution meteorological inputs. The lack of anticipated improvement could also be because a PBL scheme chosen to produce the WRF data was only a local parameterization and unable to faithfully duplicate the real atmosphere on a global scale. The MYJ scheme was able to replicate PSCF source identification by those using the Reanalysis and discover additional source areas that was not identified by the Reanalysis data. In conclusion, a potential benefit for using high-resolution wind data in the PSCF modeling is that it could discover new source location in addition to those identified by using the Reanalysis data input.« less

  19. Application of a predator-prey overlap metric to determine the impact of sub-grid scale feeding dynamics on ecosystem productivity

    NASA Astrophysics Data System (ADS)

    Greer, A. T.; Woodson, C. B.

    2016-02-01

    Because of the complexity and extremely large size of marine ecosystems, research attention has a strong focus on modelling the system through space and time to elucidate processes driving ecosystem state. One of the major weaknesses of current modelling approaches is the reliance on a particular grid cell size (usually 10's of km in the horizontal & water column mean) to capture the relevant processes, even though empirical research has shown that marine systems are highly structured on fine scales, and this structure can persist over relatively long time scales (days to weeks). Fine-scale features can have a strong influence on the predator-prey interactions driving trophic transfer. Here we apply a statistic, the AB ratio, used to quantify increased predator production due to predator-prey overlap on fine scales in a manner that is computationally feasible for larger scale models. We calculated the AB ratio for predator-prey distributions throughout the scientific literature, as well as for data obtained with a towed plankton imaging system, demonstrating that averaging across a typical model grid cell neglects the fine-scale predator-prey overlap that is an essential component of ecosystem productivity. Organisms from a range of trophic levels and oceanographic regions tended to overlap with their prey both in the horizontal and vertical dimensions. When predator swimming over a diel cycle was incorporated, the amount of production indicated by the AB ratio increased substantially. For the plankton image data, the AB ratio was higher with increasing sampling resolution, especially when prey were highly aggregated. We recommend that ecosystem models incorporate more fine-scale information both to more accurately capture trophic transfer processes and to capitalize on the increasing sampling resolution and data volume from empirical studies.

  20. About rats and jackfruit trees: modeling the carrying capacity of a Brazilian Atlantic Forest spiny-rat Trinomys dimidiatus (Günther, 1877) - Rodentia, Echimyidae - population with varying jackfruit tree (Artocarpus heterophyllus L.) abundances.

    PubMed

    Mello, J H F; Moulton, T P; Raíces, D S L; Bergallo, H G

    2015-01-01

    We carried out a six-year study aimed at evaluating if and how a Brazilian Atlantic Forest small mammal community responded to the presence of the invasive exotic species Artocarpus heterophyllus, the jackfruit tree. In the surroundings of Vila Dois Rios, Ilha Grande, RJ, 18 grids were established, 10 where the jackfruit tree was present and eight were it was absent. Previous results indicated that the composition and abundance of this small mammal community were altered by the presence and density of A. heterophyllus. One observed effect was the increased population size of the spiny-rat Trinomys dimidiatus within the grids where the jackfruit trees were present. Therefore we decided to create a mathematical model for this species, based on the Verhulst-Pearl logistic equation. Our objectives were i) to calculate the carrying capacity K based on real data of the involved species and the environment; ii) propose and evaluate a mathematical model to estimate the population size of T. dimidiatus based on the monthly seed production of jackfruit tree, Artocarpus heterophyllus and iii) determinate the minimum jackfruit tree seed production to maintain at least two T. dimidiatus individuals in one study grid. Our results indicated that the predicted values by the model for the carrying capacity K were significantly correlated with real data. The best fit was found considering 20~35% energy transfer efficiency between trophic levels. Within the scope of assumed premises, our model showed itself to be an adequate simulator for Trinomys dimidiatus populations where the invasive jackfruit tree is present.

  1. Particle Morphology and Size Results from the Smoke Aerosol Measurement Experiment-2

    NASA Technical Reports Server (NTRS)

    Urban, David L.; Ruff, Gary A.; Greenberg, Paul S.; Fischer, David; Meyer, Marit; Mulholland, George; Yuan, Zeng-Guang; Bryg, Victoria; Cleary, Thomas; Yang, Jiann

    2012-01-01

    Results are presented from the Reflight of the Smoke Aerosol Measurement Experiment (SAME-2) which was conducted during Expedition 24 (July-September 2010). The reflight experiment built upon the results of the original flight during Expedition 15 by adding diagnostic measurements and expanding the test matrix. Five different materials representative of those found in spacecraft (Teflon, Kapton, cotton, silicone rubber and Pyrell) were heated to temperatures below the ignition point with conditions controlled to provide repeatable sample surface temperatures and air flow. The air flow past the sample during the heating period ranged from quiescent to 8 cm/s. The smoke was initially collected in an aging chamber to simulate the transport time from the smoke source to the detector. This effective transport time was varied by holding the smoke in the aging chamber for times ranging from 11 to 1800 s. Smoke particle samples were collected on Transmission Electron Microscope (TEM) grids for post-flight analysis. The TEM grids were analyzed to observe the particle morphology and size parameters. The diagnostics included a prototype two-moment smoke detector and three different measures of moments of the particle size distribution. These moment diagnostics were used to determine the particle number concentration (zeroth moment), the diameter concentration (first moment), and the mass concentration (third moment). These statistics were combined to determine the diameter of average mass and the count mean diameter and, by assuming a log-normal distribution, the geometric mean diameter and the geometric standard deviations can also be calculated. Overall the majority of the average smoke particle sizes were found to be in the 200 nm to 400 nm range with the quiescent cases producing some cases with substantially larger particles.

  2. Evaluation of Low-Gravity Smoke Particulate for Spacecraft Fire Detection

    NASA Technical Reports Server (NTRS)

    Urban, David; Ruff, Gary A.; Mulholland George; Meyer, Marit; Yuan, Zeng guang; Cleary, Thomas; Yang, Jiann; Greenberg, Paul; Bryg, Victoria

    2013-01-01

    Tests were conducted on the International Space Station to evaluate the smoke particulate size from materials and conditions that are typical of those expected in spacecraft fires. Five different materials representative of those found in spacecraft (Teflon, Kapton, cotton, silicone rubber and Pyrell) were heated to temperatures below the ignition point with conditions controlled to provide repeatable sample surface temperatures and air flow. The air flow past the sample during the heating period ranged from quiescent to 8 cm/s. The effective transport time to the measurement instruments was varied from 11 to 800 seconds to simulate different smoke transport conditions in spacecraft. The resultant aerosol was evaluated by three instruments which measured different moments of the particle size distribution. These moment diagnostics were used to determine the particle number concentration (zeroth moment), the diameter concentration (first moment), and the mass concentration (third moment). These statistics were combined to determine the diameter of average mass and the count mean diameter and by assuming a log-normal distribution, the geometric mean diameter and the geometric standard deviations were also calculated. Smoke particle samples were collected on TEM grids using a thermal precipitator for post flight analysis. The TEM grids were analyzed to determine the particle morphology and shape parameters. The different materials produced particles with significantly different morphologies. Overall the majority of the average smoke particle sizes were found to be in the 200 to 400 nanometer range with the quiescent cases and the cases with increased transport time typically producing with substantially larger particles. The results varied between materials but the smoke particles produced in low gravity were typically twice the size of particles produced in normal gravity. These results can be used to establish design requirements for future spacecraft smoke detectors.

  3. Energy production estimation for Kosh-Agach grid-tie photovoltaic power plant for different photovoltaic module types

    NASA Astrophysics Data System (ADS)

    Gabderakhmanova, T. S.; Kiseleva, S. V.; Frid, S. E.; Tarasenko, A. B.

    2016-11-01

    This paper is devoted to calculation of yearly energy production, demanded area and capital costs for first Russian 5 MW grid-tie photovoltaic (PV) plant in Altay Republic that is named Kosh-Agach. Simple linear calculation model, involving average solar radiation and temperature data, grid-tie inverter power-efficiency dependence and PV modules parameters is proposed. Monthly and yearly energy production, equipment costs and demanded area for PV plant are estimated for mono-, polycrystalline and amorphous modules. Calculation includes three types of initial radiation and temperature data—average day for every month from NASA SSE, average radiation and temperature for each day of the year from NASA POWER and typical meteorology year generated from average data for every month. The peculiarities for each type of initial data and their influence on results are discussed.

  4. Multigrid method for stability problems

    NASA Technical Reports Server (NTRS)

    Ta'asan, Shlomo

    1988-01-01

    The problem of calculating the stability of steady state solutions of differential equations is addressed. Leading eigenvalues of large matrices that arise from discretization are calculated, and an efficient multigrid method for solving these problems is presented. The resulting grid functions are used as initial approximations for appropriate eigenvalue problems. The method employs local relaxation on all levels together with a global change on the coarsest level only, which is designed to separate the different eigenfunctions as well as to update their corresponding eigenvalues. Coarsening is done using the FAS formulation in a nonstandard way in which the right-hand side of the coarse grid equations involves unknown parameters to be solved on the coarse grid. This leads to a new multigrid method for calculating the eigenvalues of symmetric problems. Numerical experiments with a model problem are presented which demonstrate the effectiveness of the method.

  5. A Wall-Distance-Free k-ω SST Turbulence Model

    NASA Astrophysics Data System (ADS)

    Gleize, Vincent; Burnley, Victor

    2001-11-01

    In the calculation of flows around aircraft and aerodynamic bodies, the Shear-Stress Transport (SST) model by Menter has been used extensively due to its good prediction of flows with adverse pressure gradients. One main drawback of this model is the need to calculate the distance from the wall. While this is not a serious drawback for steady state calculations on non-moving grids, this calculation can become very cumbersome and expensive for unsteady simulations, especially when using unstructured grids. In this case, the wall-distance needs to be determined after each iteration. To avoid this problem, a new model is proposed which provides the benefits of the SST correction and avoids the freestream dependency of the solution, while not requiring the wall-distance. The first results for a wide range of test cases show that this model produces very good agreement with experimental data for flows with adverse pressure gradients, separation zones and shock-boundary layer interactions, closely matching the results obtained with the original SST model. This model should be very useful for unsteady calculations, such as store separation, grid adaptation, and other practical flows.

  6. Output-Adaptive Tetrahedral Cut-Cell Validation for Sonic Boom Prediction

    NASA Technical Reports Server (NTRS)

    Park, Michael A.; Darmofal, David L.

    2008-01-01

    A cut-cell approach to Computational Fluid Dynamics (CFD) that utilizes the median dual of a tetrahedral background grid is described. The discrete adjoint is also calculated, which permits adaptation based on improving the calculation of a specified output (off-body pressure signature) in supersonic inviscid flow. These predicted signatures are compared to wind tunnel measurements on and off the configuration centerline 10 body lengths below the model to validate the method for sonic boom prediction. Accurate mid-field sonic boom pressure signatures are calculated with the Euler equations without the use of hybrid grid or signature propagation methods. Highly-refined, shock-aligned anisotropic grids were produced by this method from coarse isotropic grids created without prior knowledge of shock locations. A heuristic reconstruction limiter provided stable flow and adjoint solution schemes while producing similar signatures to Barth-Jespersen and Venkatakrishnan limiters. The use of cut-cells with an output-based adaptive scheme completely automated this accurate prediction capability after a triangular mesh is generated for the cut surface. This automation drastically reduces the manual intervention required by existing methods.

  7. View of Pakistan Atomic Energy Commission towards SMPR's in the light of KANUPP performance

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Huseini, S.D.

    1985-01-01

    The developing countries in general do not have grid capacities adequate enough to incorporate standard size, economic but rather large nuclear power plants for maximum advantage. Therefore, small and medium size reactors (SMPR) have been and still are, of particular interest to the developing countries in spite of certain known problems with these reactors. Pakistan Atomic Energy Commission (PAEC) has been operating a CANDU type of a small PHWR plant since 1971 when it was connected to the local Karachi grid. This paper describes PAEC's view in the light of KANUPP performance with respect to such factors associated with SMPR'smore » as selection of suitable reactor size and type, its operation in a grid of small capacity, flexibility of operation and its role as a reliable source of electrical power.« less

  8. Comments on numerical solution of boundary value problems of the Laplace equation and calculation of eigenvalues by the grid method

    NASA Technical Reports Server (NTRS)

    Lyusternik, L. A.

    1980-01-01

    The mathematics involved in numerically solving for the plane boundary value of the Laplace equation by the grid method is developed. The approximate solution of a boundary value problem for the domain of the Laplace equation by the grid method consists of finding u at the grid corner which satisfies the equation at the internal corners (u=Du) and certain boundary value conditions at the boundary corners.

  9. Function-Space-Based Solution Scheme for the Size-Modified Poisson-Boltzmann Equation in Full-Potential DFT.

    PubMed

    Ringe, Stefan; Oberhofer, Harald; Hille, Christoph; Matera, Sebastian; Reuter, Karsten

    2016-08-09

    The size-modified Poisson-Boltzmann (MPB) equation is an efficient implicit solvation model which also captures electrolytic solvent effects. It combines an account of the dielectric solvent response with a mean-field description of solvated finite-sized ions. We present a general solution scheme for the MPB equation based on a fast function-space-oriented Newton method and a Green's function preconditioned iterative linear solver. In contrast to popular multigrid solvers, this approach allows us to fully exploit specialized integration grids and optimized integration schemes. We describe a corresponding numerically efficient implementation for the full-potential density-functional theory (DFT) code FHI-aims. We show that together with an additional Stern layer correction the DFT+MPB approach can describe the mean activity coefficient of a KCl aqueous solution over a wide range of concentrations. The high sensitivity of the calculated activity coefficient on the employed ionic parameters thereby suggests to use extensively tabulated experimental activity coefficients of salt solutions for a systematic parametrization protocol.

  10. Convergence of the Bouguer-Beer law for radiation extinction in particulate media

    NASA Astrophysics Data System (ADS)

    Frankel, A.; Iaccarino, G.; Mani, A.

    2016-10-01

    Radiation transport in particulate media is a common physical phenomenon in natural and industrial processes. Developing predictive models of these processes requires a detailed model of the interaction between the radiation and the particles. Resolving the interaction between the radiation and the individual particles in a very large system is impractical, whereas continuum-based representations of the particle field lend themselves to efficient numerical techniques based on the solution of the radiative transfer equation. We investigate radiation transport through discrete and continuum-based representations of a particle field. Exact solutions for radiation extinction are developed using a Monte Carlo model in different particle distributions. The particle distributions are then projected onto a concentration field with varying grid sizes, and the Bouguer-Beer law is applied by marching across the grid. We show that the continuum-based solution approaches the Monte Carlo solution under grid refinement, but quickly diverges as the grid size approaches the particle diameter. This divergence is attributed to the homogenization error of an individual particle across a whole grid cell. We remark that the concentration energy spectrum of a point-particle field does not approach zero, and thus the concentration variance must also diverge under infinite grid refinement, meaning that no grid-converged solution of the radiation transport is possible.

  11. Composite sizing and ply orientation for stiffness requirements using a large finite element structural model

    NASA Technical Reports Server (NTRS)

    Radovcich, N. A.; Gentile, D. P.

    1989-01-01

    A NASTRAN bulk dataset preprocessor was developed to facilitate the integration of filamentary composite laminate properties into composite structural resizing for stiffness requirements. The NASCOMP system generates delta stiffness and delta mass matrices for input to the flutter derivative program. The flutter baseline analysis, derivative calculations, and stiffness and mass matrix updates are controlled by engineer defined processes under an operating system called CBUS. A multi-layered design variable grid system permits high fidelity resizing without excessive computer cost. The NASCOMP system uses ply layup drawings for basic input. The aeroelastic resizing for stiffness capability was used during an actual design exercise.

  12. Computationally efficient method for Fourier transform of highly chirped pulses for laser and parametric amplifier modeling.

    PubMed

    Andrianov, Alexey; Szabo, Aron; Sergeev, Alexander; Kim, Arkady; Chvykov, Vladimir; Kalashnikov, Mikhail

    2016-11-14

    We developed an improved approach to calculate the Fourier transform of signals with arbitrary large quadratic phase which can be efficiently implemented in numerical simulations utilizing Fast Fourier transform. The proposed algorithm significantly reduces the computational cost of Fourier transform of a highly chirped and stretched pulse by splitting it into two separate transforms of almost transform limited pulses, thereby reducing the required grid size roughly by a factor of the pulse stretching. The application of our improved Fourier transform algorithm in the split-step method for numerical modeling of CPA and OPCPA shows excellent agreement with standard algorithms.

  13. Boundary conditions for the solution of compressible Navier-Stokes equations by an implicit factored method

    NASA Technical Reports Server (NTRS)

    Shih, T. I.-P.; Smith, G. E.; Springer, G. S.; Rimon, Y.

    1983-01-01

    A method is presented for formulating the boundary conditions in implicit finite-difference form needed for obtaining solutions to the compressible Navier-Stokes equations by the Beam and Warming implicit factored method. The usefulness of the method was demonstrated (a) by establishing the boundary conditions applicable to the analysis of the flow inside an axisymmetric piston-cylinder configuration and (b) by calculating velocities and mass fractions inside the cylinder for different geometries and different operating conditions. Stability, selection of time step and grid sizes, and computer time requirements are discussed in reference to the piston-cylinder problem analyzed.

  14. Computational time analysis of the numerical solution of 3D electrostatic Poisson's equation

    NASA Astrophysics Data System (ADS)

    Kamboh, Shakeel Ahmed; Labadin, Jane; Rigit, Andrew Ragai Henri; Ling, Tech Chaw; Amur, Khuda Bux; Chaudhary, Muhammad Tayyab

    2015-05-01

    3D Poisson's equation is solved numerically to simulate the electric potential in a prototype design of electrohydrodynamic (EHD) ion-drag micropump. Finite difference method (FDM) is employed to discretize the governing equation. The system of linear equations resulting from FDM is solved iteratively by using the sequential Jacobi (SJ) and sequential Gauss-Seidel (SGS) methods, simulation results are also compared to examine the difference between the results. The main objective was to analyze the computational time required by both the methods with respect to different grid sizes and parallelize the Jacobi method to reduce the computational time. In common, the SGS method is faster than the SJ method but the data parallelism of Jacobi method may produce good speedup over SGS method. In this study, the feasibility of using parallel Jacobi (PJ) method is attempted in relation to SGS method. MATLAB Parallel/Distributed computing environment is used and a parallel code for SJ method is implemented. It was found that for small grid size the SGS method remains dominant over SJ method and PJ method while for large grid size both the sequential methods may take nearly too much processing time to converge. Yet, the PJ method reduces computational time to some extent for large grid sizes.

  15. Rapid Airplane Parametric Input Design (RAPID)

    NASA Technical Reports Server (NTRS)

    Smith, Robert E.

    1995-01-01

    RAPID is a methodology and software system to define a class of airplane configurations and directly evaluate surface grids, volume grids, and grid sensitivity on and about the configurations. A distinguishing characteristic which separates RAPID from other airplane surface modellers is that the output grids and grid sensitivity are directly applicable in CFD analysis. A small set of design parameters and grid control parameters govern the process which is incorporated into interactive software for 'real time' visual analysis and into batch software for the application of optimization technology. The computed surface grids and volume grids are suitable for a wide range of Computational Fluid Dynamics (CFD) simulation. The general airplane configuration has wing, fuselage, horizontal tail, and vertical tail components. The double-delta wing and tail components are manifested by solving a fourth order partial differential equation (PDE) subject to Dirichlet and Neumann boundary conditions. The design parameters are incorporated into the boundary conditions and therefore govern the shapes of the surfaces. The PDE solution yields a smooth transition between boundaries. Surface grids suitable for CFD calculation are created by establishing an H-type topology about the configuration and incorporating grid spacing functions in the PDE equation for the lifting components and the fuselage definition equations. User specified grid parameters govern the location and degree of grid concentration. A two-block volume grid about a configuration is calculated using the Control Point Form (CPF) technique. The interactive software, which runs on Silicon Graphics IRIS workstations, allows design parameters to be continuously varied and the resulting surface grid to be observed in real time. The batch software computes both the surface and volume grids and also computes the sensitivity of the output grid with respect to the input design parameters by applying the precompiler tool ADIFOR to the grid generation program. The output of ADIFOR is a new source code containing the old code plus expressions for derivatives of specified dependent variables (grid coordinates) with respect to specified independent variables (design parameters). The RAPID methodology and software provide a means of rapidly defining numerical prototypes, grids, and grid sensitivity of a class of airplane configurations. This technology and software is highly useful for CFD research for preliminary design and optimization processes.

  16. NAS Grid Benchmarks: A Tool for Grid Space Exploration

    NASA Technical Reports Server (NTRS)

    Frumkin, Michael; VanderWijngaart, Rob F.; Biegel, Bryan (Technical Monitor)

    2001-01-01

    We present an approach for benchmarking services provided by computational Grids. It is based on the NAS Parallel Benchmarks (NPB) and is called NAS Grid Benchmark (NGB) in this paper. We present NGB as a data flow graph encapsulating an instance of an NPB code in each graph node, which communicates with other nodes by sending/receiving initialization data. These nodes may be mapped to the same or different Grid machines. Like NPB, NGB will specify several different classes (problem sizes). NGB also specifies the generic Grid services sufficient for running the bench-mark. The implementor has the freedom to choose any specific Grid environment. However, we describe a reference implementation in Java, and present some scenarios for using NGB.

  17. SteamTablesGrid: An ActiveX control for thermodynamic properties of pure water

    NASA Astrophysics Data System (ADS)

    Verma, Mahendra P.

    2011-04-01

    An ActiveX control, steam tables grid ( StmTblGrd) to speed up the calculation of the thermodynamic properties of pure water is developed. First, it creates a grid (matrix) for a specified range of temperature (e.g. 400-600 K with 40 segments) and pressure (e.g. 100,000-20,000,000 Pa with 40 segments). Using the ActiveX component SteamTables, the values of selected properties of water for each element (nodal point) of the 41×41 matrix are calculated. The created grid can be saved in a file for its reuse. A linear interpolation within an individual phase, vapor or liquid is implemented to calculate the properties at a given value of temperature and pressure. A demonstration program to illustrate the functionality of StmTblGrd is written in Visual Basic 6.0. Similarly, a methodology is presented to explain the use of StmTblGrd in MS-Excel 2007. In an Excel worksheet, the enthalpy of 1000 random datasets for temperature and pressure is calculated using StmTblGrd and SteamTables. The uncertainty in the enthalpy calculated with StmTblGrd is within ±0.03%. The calculations were performed on a personal computer that has a "Pentium(R) 4 CPU 3.2 GHz, RAM 1.0 GB" processor and Windows XP. The total execution time for the calculation with StmTblGrd was 0.3 s, while it was 60.0 s for SteamTables. Thus, the ActiveX control approach is reliable, accurate and efficient for the numerical simulation of complex systems that demand the thermodynamic properties of water at several values of temperature and pressure like steam flow in a geothermal pipeline network.

  18. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wiley, J.C.

    The author describes a general `hp` finite element method with adaptive grids. The code was based on the work of Oden, et al. The term `hp` refers to the method of spatial refinement (h), in conjunction with the order of polynomials used as a part of the finite element discretization (p). This finite element code seems to handle well the different mesh grid sizes occuring between abuted grids with different resolutions.

  19. A new solution-adaptive grid generation method for transonic airfoil flow calculations

    NASA Technical Reports Server (NTRS)

    Nakamura, S.; Holst, T. L.

    1981-01-01

    The clustering algorithm is controlled by a second-order, ordinary differential equation which uses the airfoil surface density gradient as a forcing function. The solution to this differential equation produces a surface grid distribution which is automatically clustered in regions with large gradients. The interior grid points are established from this surface distribution by using an interpolation scheme which is fast and retains the desirable properties of the original grid generated from the standard elliptic equation approach.

  20. Development of a hybrid 3-D hydrological model to simulate hillslopes and the regional unconfined aquifer system in Earth system models

    NASA Astrophysics Data System (ADS)

    Hazenberg, P.; Broxton, P. D.; Brunke, M.; Gochis, D.; Niu, G. Y.; Pelletier, J. D.; Troch, P. A. A.; Zeng, X.

    2015-12-01

    The terrestrial hydrological system, including surface and subsurface water, is an essential component of the Earth's climate system. Over the past few decades, land surface modelers have built one-dimensional (1D) models resolving the vertical flow of water through the soil column for use in Earth system models (ESMs). These models generally have a relatively coarse model grid size (~25-100 km) and only account for sub-grid lateral hydrological variations using simple parameterization schemes. At the same time, hydrologists have developed detailed high-resolution (~0.1-10 km grid size) three dimensional (3D) models and showed the importance of accounting for the vertical and lateral redistribution of surface and subsurface water on soil moisture, the surface energy balance and ecosystem dynamics on these smaller scales. However, computational constraints have limited the implementation of the high-resolution models for continental and global scale applications. The current work presents a hybrid-3D hydrological approach is presented, where the 1D vertical soil column model (available in many ESMs) is coupled with a high-resolution lateral flow model (h2D) to simulate subsurface flow and overland flow. H2D accounts for both local-scale hillslope and regional-scale unconfined aquifer responses (i.e. riparian zone and wetlands). This approach was shown to give comparable results as those obtained by an explicit 3D Richards model for the subsurface, but improves runtime efficiency considerably. The h3D approach is implemented for the Delaware river basin, where Noah-MP land surface model (LSM) is used to calculated vertical energy and water exchanges with the atmosphere using a 10km grid resolution. Noah-MP was coupled within the WRF-Hydro infrastructure with the lateral 1km grid resolution h2D model, for which the average depth-to-bedrock, hillslope width function and soil parameters were estimated from digital datasets. The ability of this h3D approach to simulate the hydrological dynamics of the Delaware River basin will be assessed by comparing the model results (both hydrological performance and numerical efficiency) with the standard setup of the NOAH-MP model and a high-resolution (1km) version of NOAH-MP, which also explicitly accounts for lateral subsurface and overland flow.

  1. SU-E-T-405: Evaluation of the Raystation Electron Monte Carlo Algorithm for Varian Linear Accelerators

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sansourekidou, P; Allen, C

    2015-06-15

    Purpose: To evaluate the Raystation v4.51 Electron Monte Carlo algorithm for Varian Trilogy, IX and 2100 series linear accelerators and commission for clinical use. Methods: Seventy two water and forty air scans were acquired with a water tank in the form of profiles and depth doses, as requested by vendor. Data was imported into Rayphysics beam modeling module. Energy spectrum was modeled using seven parameters. Contamination photons were modeled using five parameters. Source phase space was modeled using six parameters. Calculations were performed in clinical version 4.51 and percent depth dose curves and profiles were extracted to be compared tomore » water tank measurements. Sensitivity tests were performed for all parameters. Grid size and particle histories were evaluated per energy for statistical uncertainty performance. Results: Model accuracy for air profiles is poor in the shoulder and penumbra region. However, model accuracy for water scans is acceptable. All energies and cones are within 2%/2mm for 90% of the points evaluated. Source phase space parameters have a cumulative effect. To achieve distributions with satisfactory smoothness level a 0.1cm grid and 3,000,000 particle histories were used for commissioning calculations. Calculation time was approximately 3 hours per energy. Conclusion: Raystation electron Monte Carlo is acceptable for clinical use for the Varian accelerators listed. Results are inferior to Elekta Electron Monte Carlo modeling. Known issues were reported to Raysearch and will be resolved in upcoming releases. Auto-modeling is limited to open cone depth dose curves and needs expansion.« less

  2. Noise pollution mapping approach and accuracy on landscape scales.

    PubMed

    Iglesias Merchan, Carlos; Diaz-Balteiro, Luis

    2013-04-01

    Noise mapping allows the characterization of environmental variables, such as noise pollution or soundscape, depending on the task. Strategic noise mapping (as per Directive 2002/49/EC, 2002) is a tool intended for the assessment of noise pollution at the European level every five years. These maps are based on common methods and procedures intended for human exposure assessment in the European Union that could be also be adapted for assessing environmental noise pollution in natural parks. However, given the size of such areas, there could be an alternative approach to soundscape characterization rather than using human noise exposure procedures. It is possible to optimize the size of the mapping grid used for such work by taking into account the attributes of the area to be studied and the desired outcome. This would then optimize the mapping time and the cost. This type of optimization is important in noise assessment as well as in the study of other environmental variables. This study compares 15 models, using different grid sizes, to assess the accuracy of the noise mapping of the road traffic noise at a landscape scale, with respect to noise and landscape indicators. In a study area located in the Manzanares High River Basin Regional Park in Spain, different accuracy levels (Kappa index values from 0.725 to 0.987) were obtained depending on the terrain and noise source properties. The time taken for the calculations and the noise mapping accuracy results reveal the potential for setting the map resolution in line with decision-makers' criteria and budget considerations. Copyright © 2013 Elsevier B.V. All rights reserved.

  3. Accurate, robust and reliable calculations of Poisson-Boltzmann binding energies

    PubMed Central

    Nguyen, Duc D.; Wang, Bao

    2017-01-01

    Poisson-Boltzmann (PB) model is one of the most popular implicit solvent models in biophysical modeling and computation. The ability of providing accurate and reliable PB estimation of electrostatic solvation free energy, ΔGel, and binding free energy, ΔΔGel, is important to computational biophysics and biochemistry. In this work, we investigate the grid dependence of our PB solver (MIBPB) with SESs for estimating both electrostatic solvation free energies and electrostatic binding free energies. It is found that the relative absolute error of ΔGel obtained at the grid spacing of 1.0 Å compared to ΔGel at 0.2 Å averaged over 153 molecules is less than 0.2%. Our results indicate that the use of grid spacing 0.6 Å ensures accuracy and reliability in ΔΔGel calculation. In fact, the grid spacing of 1.1 Å appears to deliver adequate accuracy for high throughput screening. PMID:28211071

  4. Ionization correction factors for H II regions in blue compact dwarf galaxies

    NASA Astrophysics Data System (ADS)

    Holovatyi, V. V.; Melekh, B. Ya.

    2002-08-01

    Energy distributions in the spectra of the ionizing nuclei of H II regions beyond λ <= 91.2 nm were calculated. A grid of photoionization models of 270 H II regions was constructed. The free parameters of the model grid are the hydrogen density nH in the nebular gas, filling factor, energy Lc-spectrum of ionizing nuclei, and metallicity. The chemical composition from the studies of Izotov et al. were used for model grid initialization. The integral linear spectra calculated for the photoionization models were used to determine the concentration ne, temperatures Te of electrons, and ionic concentrations n(A+i)/n(H+) by the nebular gas diagnostic method. The averaged relative ionic abundances n(A+i)/n(H+) thus calculated were used to determine new expressions for ionization correction factors which we recommend for the determination of abundances in the H II regions of blue compact dwarf galaxies.

  5. GRAMS: A Grid of RSG and AGB Models

    NASA Astrophysics Data System (ADS)

    Srinivasan, S.; Sargent, B. A.; Meixner, M.

    2011-09-01

    We present a grid of oxygen- and carbon-rich circumstellar dust radiative transfer models for asymptotic giant branch (AGB) and red supergiant (RSG) stars. The grid samples a large region of the relevant parameter space, and it allows for a quick calculation of bolometric fluxes and dust mass-loss rates from multi-wavelength photometry. This method of fitting observed spectral energy distributions (SEDs) is preferred over detailed radiative transfer calculations, especially for large data sets such as the SAGE (Surveying the Agents of a Galaxy's Evolution) survey of the Magellanic Clouds. The mass-loss rates calculated for SAGE data will allow us to quantify the dust returned to the interstellar medium (ISM) by the entire AGB population. The total injection rate provides an important constraint for models of galactic chemical evolution. Here, we discuss our carbon star models and compare the results to SAGE observations in the Large Magellanic Cloud (LMC).

  6. The equal load-sharing model of cascade failures in power grids

    NASA Astrophysics Data System (ADS)

    Scala, Antonio; De Sanctis Lucentini, Pier Giorgio

    2016-11-01

    Electric power-systems are one of the most important critical infrastructures. In recent years, they have been exposed to extreme stress due to the increasing power demand, the introduction of distributed renewable energy sources, and the development of extensive interconnections. We investigate the phenomenon of abrupt breakdown of an electric power-system under two scenarios: load growth (mimicking the ever-increasing customer demand) and power fluctuations (mimicking the effects of renewable sources). Our results indicate that increasing the system size causes breakdowns to become more abrupt; in fact, mapping the system to a solvable statistical-physics model indicates the occurrence of a first order transition in the large size limit. Such an enhancement for the systemic risk failures (black-outs) with increasing network size is an effect that should be considered in the current projects aiming to integrate national power-grids into ;super-grids;.

  7. Abruptness of Cascade Failures in Power Grids

    NASA Astrophysics Data System (ADS)

    Pahwa, Sakshi; Scoglio, Caterina; Scala, Antonio

    2014-01-01

    Electric power-systems are one of the most important critical infrastructures. In recent years, they have been exposed to extreme stress due to the increasing demand, the introduction of distributed renewable energy sources, and the development of extensive interconnections. We investigate the phenomenon of abrupt breakdown of an electric power-system under two scenarios: load growth (mimicking the ever-increasing customer demand) and power fluctuations (mimicking the effects of renewable sources). Our results on real, realistic and synthetic networks indicate that increasing the system size causes breakdowns to become more abrupt; in fact, mapping the system to a solvable statistical-physics model indicates the occurrence of a first order transition in the large size limit. Such an enhancement for the systemic risk failures (black-outs) with increasing network size is an effect that should be considered in the current projects aiming to integrate national power-grids into ``super-grids''.

  8. Abruptness of cascade failures in power grids.

    PubMed

    Pahwa, Sakshi; Scoglio, Caterina; Scala, Antonio

    2014-01-15

    Electric power-systems are one of the most important critical infrastructures. In recent years, they have been exposed to extreme stress due to the increasing demand, the introduction of distributed renewable energy sources, and the development of extensive interconnections. We investigate the phenomenon of abrupt breakdown of an electric power-system under two scenarios: load growth (mimicking the ever-increasing customer demand) and power fluctuations (mimicking the effects of renewable sources). Our results on real, realistic and synthetic networks indicate that increasing the system size causes breakdowns to become more abrupt; in fact, mapping the system to a solvable statistical-physics model indicates the occurrence of a first order transition in the large size limit. Such an enhancement for the systemic risk failures (black-outs) with increasing network size is an effect that should be considered in the current projects aiming to integrate national power-grids into "super-grids".

  9. Analysis of turbine-grid interaction of grid-connected wind turbine using HHT

    NASA Astrophysics Data System (ADS)

    Chen, A.; Wu, W.; Miao, J.; Xie, D.

    2018-05-01

    This paper processes the output power of the grid-connected wind turbine with the denoising and extracting method based on Hilbert Huang transform (HHT) to discuss the turbine-grid interaction. At first, the detailed Empirical Mode Decomposition (EMD) and the Hilbert Transform (HT) are introduced. Then, on the premise of decomposing the output power of the grid-connected wind turbine into a series of Intrinsic Mode Functions (IMFs), energy ratio and power volatility are calculated to detect the unessential components. Meanwhile, combined with vibration function of turbine-grid interaction, data fitting of instantaneous amplitude and phase of each IMF is implemented to extract characteristic parameters of different interactions. Finally, utilizing measured data of actual parallel-operated wind turbines in China, this work accurately obtains the characteristic parameters of turbine-grid interaction of grid-connected wind turbine.

  10. Fully Automated Single-Zone Elliptic Grid Generation for Mars Science Laboratory (MSL) Aeroshell and Canopy Geometries

    NASA Technical Reports Server (NTRS)

    kaul, Upender K.

    2008-01-01

    A procedure for generating smooth uniformly clustered single-zone grids using enhanced elliptic grid generation has been demonstrated here for the Mars Science Laboratory (MSL) geometries such as aeroshell and canopy. The procedure obviates the need for generating multizone grids for such geometries, as reported in the literature. This has been possible because the enhanced elliptic grid generator automatically generates clustered grids without manual prescription of decay parameters needed with the conventional approach. In fact, these decay parameters are calculated as decay functions as part of the solution, and they are not constant over a given boundary. Since these decay functions vary over a given boundary, orthogonal grids near any arbitrary boundary can be clustered automatically without having to break up the boundaries and the corresponding interior domains into various zones for grid generation.

  11. Advanced Unstructured Grid Generation for Complex Aerodynamic Applications

    NASA Technical Reports Server (NTRS)

    Pirzadeh, Shahyar Z.

    2008-01-01

    A new approach for distribution of grid points on the surface and in the volume has been developed and implemented in the NASA unstructured grid generation code VGRID. In addition to the point and line sources of prior work, the new approach utilizes surface and volume sources for automatic curvature-based grid sizing and convenient point distribution in the volume. A new exponential growth function produces smoother and more efficient grids and provides superior control over distribution of grid points in the field. All types of sources support anisotropic grid stretching which not only improves the grid economy but also provides more accurate solutions for certain aerodynamic applications. The new approach does not require a three-dimensional background grid as in the previous methods. Instead, it makes use of an efficient bounding-box auxiliary medium for storing grid parameters defined by surface sources. The new approach is less memory-intensive and more efficient computationally. The grids generated with the new method either eliminate the need for adaptive grid refinement for certain class of problems or provide high quality initial grids that would enhance the performance of many adaptation methods.

  12. Review of Strategies and Technologies for Demand-Side Management on Isolated Mini-Grids

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Harper, Meg

    This review provides an overview of strategies and currently available technologies used for demandside management (DSM) on mini-grids throughout the world. For the purposes of this review, mini-grids are defined as village-scale electricity distribution systems powered by small local generation sources and not connected to a main grid.1 Mini-grids range in size from less than 1 kW to several hundred kW of installed generation capacity and may utilize different generation technologies, such as micro-hydro, biomass gasification, solar, wind, diesel generators, or a hybrid combination of any of these. This review will primarily refer to AC mini-grids, though much of themore » discussion could apply to DC grids as well. Many mini-grids include energy storage, though some rely solely on real-time generation.« less

  13. Spatio-temporal modelling of residential exposure to particulate matter and gaseous pollutants for the Heinz Nixdorf Recall Cohort

    NASA Astrophysics Data System (ADS)

    Nonnemacher, Michael; Jakobs, Hermann; Viehmann, Anja; Vanberg, Irene; Kessler, Christoph; Moebus, Susanne; Möhlenkamp, Stefan; Erbel, Raimund; Hoffmann, Barbara; Memmesheimer, Michael

    2014-07-01

    For the simultaneous analysis of short- and long-term effects of air pollution in the Heinz Nixdorf Recall Cohort a sophisticated exposure modelling was performed. The dispersion and chemistry transport model EURAD (European Air Pollution Dispersion) was used for the estimation of hourly concentrations of a number of pollutants for a horizontal grid with a cell size of 1 km² covering the whole study area (three large adjacent cities in a highly urbanized region in Western Germany) for the years 2000-2003 and 2006-2008. For each 1 km² cell we estimated the mean concentration by calculating daily means from the hourly concentrations modelled by the EURAD process. The modelled concentrations showed an overall tendency to decrease from 2001 to 2008 whereas the trend in the single grid cells and study period was inhomogeneous. Participant-related exposure slightly increased from 2001 to 2003 followed by a decrease from 2006 to 2008. The exposure modelling enables a very flexible exposure assessment compared to conventional modelling approaches which either use central monitoring or temporally static spatial contrasts. The modelling allows the calculation of an average exposure concentration for any place and time within the study region and study period with a high spatial and temporal resolution. This is important for the assessment of short-, medium and long-term effects of air pollution on human health in epidemiological studies.

  14. The Surface Pressure Response of a NACA 0015 Airfoil Immersed in Grid Turbulence. Volume 1; Characteristics of the Turbulence

    NASA Technical Reports Server (NTRS)

    Bereketab, Semere; Wang, Hong-Wei; Mish, Patrick; Devenport, William J.

    2000-01-01

    Two grids have been developed for the Virginia Tech 6 ft x 6 ft Stability wind tunnel for the purpose of generating homogeneous isotropic turbulent flows for the study of unsteady airfoil response. The first, a square bi-planar grid with a 12" mesh size and an open area ratio of 69.4%, was mounted in the wind tunnel contraction. The second grid, a metal weave with a 1.2 in. mesh size and an open area ratio of 68.2% was mounted in the tunnel test section. Detailed statistical and spectral measurements of the turbulence generated by the two grids are presented for wind tunnel free stream speeds of 10, 20, 30 and 40 m/s. These measurements show the flows to be closely homogeneous and isotropic. Both grids produce flows with a turbulence intensity of about 4% at the location planned for the airfoil leading edge. Turbulence produced by the large grid has an integral scale of some 3.2 inches here. Turbulence produced by the small grid is an order of magnitude smaller. For wavenumbers below the upper limit of the inertial subrange, the spectra and correlations measured with both grids at all speeds can be represented using the von Karman interpolation formula with a single velocity and length scale. The spectra maybe accurately represented over the entire wavenumber range by a modification of the von Karman interpolation formula that includes the effects of dissipation. These models are most accurate at the higher speeds (30 and 40 m/s).

  15. Elimination of artificial grid distortion and hourglass-type motions by means of Lagrangian subzonal masses and pressures

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Caramana, E.J.; Shashkov, M.J.

    1997-12-31

    The bane of Lagrangian hydrodynamics calculations is premature breakdown of the grid topology that results in severe degradation of accuracy and run termination often long before the assumption of Lagrangian zonal mass ceased to be valid. At short spatial grid scales this is usually referred to by the terms hourglass mode or keystone motion associated in particular with underconstrained grids such as quadrilaterals and hexahedrons in two and three dimensions, respectively. At longer spatial scales relative to the grid spacing there is what is referred to ubiquitously as spurious vorticity, or the long-thin zone problem. In both cases the resultmore » is anomalous grid distortion and tangling that has nothing to do with the actual solution, as would be the case for turbulent flow. In this work the authors show how such motions can be eliminated by the proper use of subzonal Lagrangian masses, and associated densities and pressures. These subzonal masses arise in a natural way from the fact that they require the mass associated with the nodal grid point to be constant in time. This is addition to the usual assumption of constant, Lagrangian zonal mass in staggered grid hydrodynamics scheme. The authors show that with proper discretization of subzonal forces resulting from subzonal pressures, hourglass motion and spurious vorticity can be eliminated for a very large range of problems. Finally the authors are presenting results of calculations of many test problems.« less

  16. Numerical prediction of the interference drag of a streamlined strut intersecting a surface in transonic flow

    NASA Astrophysics Data System (ADS)

    Tetrault, Philippe-Andre

    2000-10-01

    In transonic flow, the aerodynamic interference that occurs on a strut-braced wing airplane, pylons, and other applications is significant. The purpose of this work is to provide relationships to estimate the interference drag of wing-strut, wing-pylon, and wing-body arrangements. Those equations are obtained by fitting a curve to the results obtained from numerous Computational Fluid Dynamics (CFD) calculations using state-of-the-art codes that employ the Spalart-Allmaras turbulence model. In order to estimate the effect of the strut thickness, the Reynolds number of the flow, and the angle made by the strut with an adjacent surface, inviscid and viscous calculations are performed on a symmetrical strut at an angle between parallel walls. The computations are conducted at a Mach number of 0.85 and Reynolds numbers of 5.3 and 10.6 million based on the strut chord. The interference drag is calculated as the drag increment of the arrangement compared to an equivalent two-dimensional strut of the same cross-section. The results show a rapid increase of the interference drag as the angle of the strut deviates from a position perpendicular to the wall. Separation regions appear for low intersection angles, but the viscosity generally provides a positive effect in alleviating the strength of the shock near the junction and thus the drag penalty. When the thickness-to-chord ratio of the strut is reduced, the flowfield is disturbed only locally at the intersection of the strut with the wall. This study provides an equation to estimate the interference drag of simple intersections in transonic flow. In the course of performing the calculations associated with this work, an unstructured flow solver was utilized. Accurate drag prediction requires a very fine grid and this leads to problems associated with the grid generator. Several challenges facing the unstructured grid methodology are discussed: slivers, grid refinement near the leading edge and at the trailing edge, grid convergence studies, volume grid generation, and other practical matters concerning such calculations.

  17. Real-time digital holographic microscopy using the graphic processing unit.

    PubMed

    Shimobaba, Tomoyoshi; Sato, Yoshikuni; Miura, Junya; Takenouchi, Mai; Ito, Tomoyoshi

    2008-08-04

    Digital holographic microscopy (DHM) is a well-known powerful method allowing both the amplitude and phase of a specimen to be simultaneously observed. In order to obtain a reconstructed image from a hologram, numerous calculations for the Fresnel diffraction are required. The Fresnel diffraction can be accelerated by the FFT (Fast Fourier Transform) algorithm. However, real-time reconstruction from a hologram is difficult even if we use a recent central processing unit (CPU) to calculate the Fresnel diffraction by the FFT algorithm. In this paper, we describe a real-time DHM system using a graphic processing unit (GPU) with many stream processors, which allows use as a highly parallel processor. The computational speed of the Fresnel diffraction using the GPU is faster than that of recent CPUs. The real-time DHM system can obtain reconstructed images from holograms whose size is 512 x 512 grids in 24 frames per second.

  18. Development of computer-aided design system of elastic sensitive elements of automatic metering devices

    NASA Astrophysics Data System (ADS)

    Kalinkina, M. E.; Kozlov, A. S.; Labkovskaia, R. I.; Pirozhnikova, O. I.; Tkalich, V. L.; Shmakov, N. A.

    2018-05-01

    The object of research is the element base of devices of control and automation systems, including in its composition annular elastic sensitive elements, methods of their modeling, calculation algorithms and software complexes for automation of their design processes. The article is devoted to the development of the computer-aided design system of elastic sensitive elements used in weight- and force-measuring automation devices. Based on the mathematical modeling of deformation processes in a solid, as well as the results of static and dynamic analysis, the calculation of elastic elements is given using the capabilities of modern software systems based on numerical simulation. In the course of the simulation, the model was a divided hexagonal grid of finite elements with a maximum size not exceeding 2.5 mm. The results of modal and dynamic analysis are presented in this article.

  19. Improved image quality of cone beam CT scans for radiotherapy image guidance using fiber-interspaced antiscatter grid

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Stankovic, Uros; Herk, Marcel van; Ploeger, Lennert S.

    Purpose: Medical linear accelerator mounted cone beam CT (CBCT) scanner provides useful soft tissue contrast for purposes of image guidance in radiotherapy. The presence of extensive scattered radiation has a negative effect on soft tissue visibility and uniformity of CBCT scans. Antiscatter grids (ASG) are used in the field of diagnostic radiography to mitigate the scatter. They usually do increase the contrast of the scan, but simultaneously increase the noise. Therefore, and considering other scatter mitigation mechanisms present in a CBCT scanner, the applicability of ASGs with aluminum interspacing for a wide range of imaging conditions has been inconclusive inmore » previous studies. In recent years, grids using fiber interspacers have appeared, providing grids with higher scatter rejection while maintaining reasonable transmission of primary radiation. The purpose of this study was to evaluate the impact of one such grid on CBCT image quality. Methods: The grid used (Philips Medical Systems) had ratio of 21:1, frequency 36 lp/cm, and nominal selectivity of 11.9. It was mounted on the kV flat panel detector of an Elekta Synergy linear accelerator and tested in a phantom and a clinical study. Due to the flex of the linac and presence of gridline artifacts an angle dependent gain correction algorithm was devised to mitigate resulting artifacts. Scan reconstruction was performed using XVI4.5 augmented with inhouse developed image lag correction and Hounsfield unit calibration. To determine the necessary parameters for Hounsfield unit calibration and software scatter correction parameters, the Catphan 600 (The Phantom Laboratory) phantom was used. Image quality parameters were evaluated using CIRS CBCT Image Quality and Electron Density Phantom (CIRS) in two different geometries: one modeling head and neck and other pelvic region. Phantoms were acquired with and without the grid and reconstructed with and without software correction which was adapted for the different acquisition scenarios. Parameters used in the phantom study weret{sub cup} for nonuniformity and contrast-to-noise ratio (CNR) for soft tissue visibility. Clinical scans were evaluated in an observer study in which four experienced radiotherapy technologists rated soft tissue visibility and uniformity of scans with and without the grid. Results: The proposed angle dependent gain correction algorithm suppressed the visible ring artifacts. Grid had a beneficial impact on nonuniformity, contrast to noise ratio, and Hounsfield unit accuracy for both scanning geometries. The nonuniformity reduced by 90% for head sized object and 91% for pelvic-sized object. CNR improved compared to no corrections on average by a factor 2.8 for the head sized object, and 2.2 for the pelvic sized phantom. Grid outperformed software correction alone, but adding additional software correction to the grid was overall the best strategy. In the observer study, a significant improvement was found in both soft tissue visibility and nonuniformity of scans when grid is used. Conclusions: The evaluated fiber-interspaced grid improved the image quality of the CBCT system for broad range of imaging conditions. Clinical scans show significant improvement in soft tissue visibility and uniformity without the need to increase the imaging dose.« less

  20. Improved image quality of cone beam CT scans for radiotherapy image guidance using fiber-interspaced antiscatter grid.

    PubMed

    Stankovic, Uros; van Herk, Marcel; Ploeger, Lennert S; Sonke, Jan-Jakob

    2014-06-01

    Medical linear accelerator mounted cone beam CT (CBCT) scanner provides useful soft tissue contrast for purposes of image guidance in radiotherapy. The presence of extensive scattered radiation has a negative effect on soft tissue visibility and uniformity of CBCT scans. Antiscatter grids (ASG) are used in the field of diagnostic radiography to mitigate the scatter. They usually do increase the contrast of the scan, but simultaneously increase the noise. Therefore, and considering other scatter mitigation mechanisms present in a CBCT scanner, the applicability of ASGs with aluminum interspacing for a wide range of imaging conditions has been inconclusive in previous studies. In recent years, grids using fiber interspacers have appeared, providing grids with higher scatter rejection while maintaining reasonable transmission of primary radiation. The purpose of this study was to evaluate the impact of one such grid on CBCT image quality. The grid used (Philips Medical Systems) had ratio of 21:1, frequency 36 lp/cm, and nominal selectivity of 11.9. It was mounted on the kV flat panel detector of an Elekta Synergy linear accelerator and tested in a phantom and a clinical study. Due to the flex of the linac and presence of gridline artifacts an angle dependent gain correction algorithm was devised to mitigate resulting artifacts. Scan reconstruction was performed using XVI4.5 augmented with inhouse developed image lag correction and Hounsfield unit calibration. To determine the necessary parameters for Hounsfield unit calibration and software scatter correction parameters, the Catphan 600 (The Phantom Laboratory) phantom was used. Image quality parameters were evaluated using CIRS CBCT Image Quality and Electron Density Phantom (CIRS) in two different geometries: one modeling head and neck and other pelvic region. Phantoms were acquired with and without the grid and reconstructed with and without software correction which was adapted for the different acquisition scenarios. Parameters used in the phantom study were t(cup) for nonuniformity and contrast-to-noise ratio (CNR) for soft tissue visibility. Clinical scans were evaluated in an observer study in which four experienced radiotherapy technologists rated soft tissue visibility and uniformity of scans with and without the grid. The proposed angle dependent gain correction algorithm suppressed the visible ring artifacts. Grid had a beneficial impact on nonuniformity, contrast to noise ratio, and Hounsfield unit accuracy for both scanning geometries. The nonuniformity reduced by 90% for head sized object and 91% for pelvic-sized object. CNR improved compared to no corrections on average by a factor 2.8 for the head sized object, and 2.2 for the pelvic sized phantom. Grid outperformed software correction alone, but adding additional software correction to the grid was overall the best strategy. In the observer study, a significant improvement was found in both soft tissue visibility and nonuniformity of scans when grid is used. The evaluated fiber-interspaced grid improved the image quality of the CBCT system for broad range of imaging conditions. Clinical scans show significant improvement in soft tissue visibility and uniformity without the need to increase the imaging dose.

  1. Advanced Unstructured Grid Generation for Complex Aerodynamic Applications

    NASA Technical Reports Server (NTRS)

    Pirzadeh, Shahyar

    2010-01-01

    A new approach for distribution of grid points on the surface and in the volume has been developed. In addition to the point and line sources of prior work, the new approach utilizes surface and volume sources for automatic curvature-based grid sizing and convenient point distribution in the volume. A new exponential growth function produces smoother and more efficient grids and provides superior control over distribution of grid points in the field. All types of sources support anisotropic grid stretching which not only improves the grid economy but also provides more accurate solutions for certain aerodynamic applications. The new approach does not require a three-dimensional background grid as in the previous methods. Instead, it makes use of an efficient bounding-box auxiliary medium for storing grid parameters defined by surface sources. The new approach is less memory-intensive and more efficient computationally. The grids generated with the new method either eliminate the need for adaptive grid refinement for certain class of problems or provide high quality initial grids that would enhance the performance of many adaptation methods.

  2. Efficient algorithm for locating and sizing series compensation devices in large power transmission grids: II. Solutions and applications

    DOE PAGES

    Frolov, Vladimir; Backhaus, Scott; Chertkov, Misha

    2014-10-01

    In a companion manuscript, we developed a novel optimization method for placement, sizing, and operation of Flexible Alternating Current Transmission System (FACTS) devices to relieve transmission network congestion. Specifically, we addressed FACTS that provide Series Compensation (SC) via modification of line inductance. In this manuscript, this heuristic algorithm and its solutions are explored on a number of test cases: a 30-bus test network and a realistically-sized model of the Polish grid (~ 2700 nodes and ~ 3300 lines). The results on the 30-bus network are used to study the general properties of the solutions including non-locality and sparsity. The Polishmore » grid is used as a demonstration of the computational efficiency of the heuristics that leverages sequential linearization of power flow constraints and cutting plane methods that take advantage of the sparse nature of the SC placement solutions. Using these approaches, the algorithm is able to solve an instance of Polish grid in tens of seconds. We explore the utility of the algorithm by analyzing transmission networks congested by (a) uniform load growth, (b) multiple overloaded configurations, and (c) sequential generator retirements.« less

  3. Efficient Algorithm for Locating and Sizing Series Compensation Devices in Large Transmission Grids: Solutions and Applications (PART II)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Frolov, Vladimir; Backhaus, Scott N.; Chertkov, Michael

    2014-01-14

    In a companion manuscript, we developed a novel optimization method for placement, sizing, and operation of Flexible Alternating Current Transmission System (FACTS) devices to relieve transmission network congestion. Specifically, we addressed FACTS that provide Series Compensation (SC) via modification of line inductance. In this manuscript, this heuristic algorithm and its solutions are explored on a number of test cases: a 30-bus test network and a realistically-sized model of the Polish grid (~2700 nodes and ~3300 lines). The results on the 30-bus network are used to study the general properties of the solutions including non-locality and sparsity. The Polish grid ismore » used as a demonstration of the computational efficiency of the heuristics that leverages sequential linearization of power flow constraints and cutting plane methods that take advantage of the sparse nature of the SC placement solutions. Using these approaches, the algorithm is able to solve an instance of Polish grid in tens of seconds. We explore the utility of the algorithm by analyzing transmission networks congested by (a) uniform load growth, (b) multiple overloaded configurations, and (c) sequential generator retirements« less

  4. On Accuracy of Adaptive Grid Methods for Captured Shocks

    NASA Technical Reports Server (NTRS)

    Yamaleev, Nail K.; Carpenter, Mark H.

    2002-01-01

    The accuracy of two grid adaptation strategies, grid redistribution and local grid refinement, is examined by solving the 2-D Euler equations for the supersonic steady flow around a cylinder. Second- and fourth-order linear finite difference shock-capturing schemes, based on the Lax-Friedrichs flux splitting, are used to discretize the governing equations. The grid refinement study shows that for the second-order scheme, neither grid adaptation strategy improves the numerical solution accuracy compared to that calculated on a uniform grid with the same number of grid points. For the fourth-order scheme, the dominant first-order error component is reduced by the grid adaptation, while the design-order error component drastically increases because of the grid nonuniformity. As a result, both grid adaptation techniques improve the numerical solution accuracy only on the coarsest mesh or on very fine grids that are seldom found in practical applications because of the computational cost involved. Similar error behavior has been obtained for the pressure integral across the shock. A simple analysis shows that both grid adaptation strategies are not without penalties in the numerical solution accuracy. Based on these results, a new grid adaptation criterion for captured shocks is proposed.

  5. pySeismicFMM: Python based Travel Time Calculation in Regular 2D and 3D Grids in Cartesian and Geographic Coordinates using Fast Marching Method

    NASA Astrophysics Data System (ADS)

    Wilde-Piorko, M.; Polkowski, M.

    2016-12-01

    Seismic wave travel time calculation is the most common numerical operation in seismology. The most efficient is travel time calculation in 1D velocity model - for given source, receiver depths and angular distance time is calculated within fraction of a second. Unfortunately, in most cases 1D is not enough to encounter differentiating local and regional structures. Whenever possible travel time through 3D velocity model has to be calculated. It can be achieved using ray calculation or time propagation in space. While single ray path calculation is quick it is complicated to find the ray path that connects source with the receiver. Time propagation in space using Fast Marching Method seems more efficient in most cases, especially when there are multiple receivers. In this presentation final release of a Python module pySeismicFMM is presented - simple and very efficient tool for calculating travel time from sources to receivers. Calculation requires regular 2D or 3D velocity grid either in Cartesian or geographic coordinates. On desktop class computer calculation speed is 200k grid cells per second. Calculation has to be performed once for every source location and provides travel time to all receivers. pySeismicFMM is free and open source. Development of this tool is a part of authors PhD thesis. Source code of pySeismicFMM will be published before Fall Meeting. National Science Centre Poland provided financial support for this work via NCN grant DEC-2011/02/A/ST10/00284.

  6. On the use of Schwarz-Christoffel conformal mappings to the grid generation for global ocean models

    NASA Astrophysics Data System (ADS)

    Xu, S.; Wang, B.; Liu, J.

    2015-10-01

    In this article we propose two grid generation methods for global ocean general circulation models. Contrary to conventional dipolar or tripolar grids, the proposed methods are based on Schwarz-Christoffel conformal mappings that map areas with user-prescribed, irregular boundaries to those with regular boundaries (i.e., disks, slits, etc.). The first method aims at improving existing dipolar grids. Compared with existing grids, the sample grid achieves a better trade-off between the enlargement of the latitudinal-longitudinal portion and the overall smooth grid cell size transition. The second method addresses more modern and advanced grid design requirements arising from high-resolution and multi-scale ocean modeling. The generated grids could potentially achieve the alignment of grid lines to the large-scale coastlines, enhanced spatial resolution in coastal regions, and easier computational load balance. Since the grids are orthogonal curvilinear, they can be easily utilized by the majority of ocean general circulation models that are based on finite difference and require grid orthogonality. The proposed grid generation algorithms can also be applied to the grid generation for regional ocean modeling where complex land-sea distribution is present.

  7. Cloud cover estimation: Use of GOES imagery in development of cloud cover data base for insolation assessment

    NASA Technical Reports Server (NTRS)

    Huning, J. R.; Logan, T. L.; Smith, J. H.

    1982-01-01

    The potential of using digital satellite data to establish a cloud cover data base for the United States, one that would provide detailed information on the temporal and spatial variability of cloud development are studied. Key elements include: (1) interfacing GOES data from the University of Wisconsin Meteorological Data Facility with the Jet Propulsion Laboratory's VICAR image processing system and IBIS geographic information system; (2) creation of a registered multitemporal GOES data base; (3) development of a simple normalization model to compensate for sun angle; (4) creation of a variable size georeference grid that provides detailed cloud information in selected areas and summarized information in other areas; and (5) development of a cloud/shadow model which details the percentage of each grid cell that is cloud and shadow covered, and the percentage of cloud or shadow opacity. In addition, comparison of model calculations of insolation with measured values at selected test sites was accomplished, as well as development of preliminary requirements for a large scale data base of cloud cover statistics.

  8. SU-E-T-36: A GPU-Accelerated Monte-Carlo Dose Calculation Platform and Its Application Toward Validating a ViewRay Beam Model

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wang, Y; Mazur, T; Green, O

    Purpose: To build a fast, accurate and easily-deployable research platform for Monte-Carlo dose calculations. We port the dose calculation engine PENELOPE to C++, and accelerate calculations using GPU acceleration. Simulations of a Co-60 beam model provided by ViewRay demonstrate the capabilities of the platform. Methods: We built software that incorporates a beam model interface, CT-phantom model, GPU-accelerated PENELOPE engine, and GUI front-end. We rewrote the PENELOPE kernel in C++ (from Fortran) and accelerated the code on a GPU. We seamlessly integrated a Co-60 beam model (obtained from ViewRay) into our platform. Simulations of various field sizes and SSDs using amore » homogeneous water phantom generated PDDs, dose profiles, and output factors that were compared to experiment data. Results: With GPU acceleration using a dated graphics card (Nvidia Tesla C2050), a highly accurate simulation – including 100*100*100 grid, 3×3×3 mm3 voxels, <1% uncertainty, and 4.2×4.2 cm2 field size – runs 24 times faster (20 minutes versus 8 hours) than when parallelizing on 8 threads across a new CPU (Intel i7-4770). Simulated PDDs, profiles and output ratios for the commercial system agree well with experiment data measured using radiographic film or ionization chamber. Based on our analysis, this beam model is precise enough for general applications. Conclusions: Using a beam model for a Co-60 system provided by ViewRay, we evaluate a dose calculation platform that we developed. Comparison to measurements demonstrates the promise of our software for use as a research platform for dose calculations, with applications including quality assurance and treatment plan verification.« less

  9. Multiscale computations with a wavelet-adaptive algorithm

    NASA Astrophysics Data System (ADS)

    Rastigejev, Yevgenii Anatolyevich

    A wavelet-based adaptive multiresolution algorithm for the numerical solution of multiscale problems governed by partial differential equations is introduced. The main features of the method include fast algorithms for the calculation of wavelet coefficients and approximation of derivatives on nonuniform stencils. The connection between the wavelet order and the size of the stencil is established. The algorithm is based on the mathematically well established wavelet theory. This allows us to provide error estimates of the solution which are used in conjunction with an appropriate threshold criteria to adapt the collocation grid. The efficient data structures for grid representation as well as related computational algorithms to support grid rearrangement procedure are developed. The algorithm is applied to the simulation of phenomena described by Navier-Stokes equations. First, we undertake the study of the ignition and subsequent viscous detonation of a H2 : O2 : Ar mixture in a one-dimensional shock tube. Subsequently, we apply the algorithm to solve the two- and three-dimensional benchmark problem of incompressible flow in a lid-driven cavity at large Reynolds numbers. For these cases we show that solutions of comparable accuracy as the benchmarks are obtained with more than an order of magnitude reduction in degrees of freedom. The simulations show the striking ability of the algorithm to adapt to a solution having different scales at different spatial locations so as to produce accurate results at a relatively low computational cost.

  10. Harmonics analysis of the ITER poloidal field converter based on a piecewise method

    NASA Astrophysics Data System (ADS)

    Xudong, WANG; Liuwei, XU; Peng, FU; Ji, LI; Yanan, WU

    2017-12-01

    Poloidal field (PF) converters provide controlled DC voltage and current to PF coils. The many harmonics generated by the PF converter flow into the power grid and seriously affect power systems and electric equipment. Due to the complexity of the system, the traditional integral operation in Fourier analysis is complicated and inaccurate. This paper presents a piecewise method to calculate the harmonics of the ITER PF converter. The relationship between the grid input current and the DC output current of the ITER PF converter is deduced. The grid current is decomposed into the sum of some simple functions. By calculating simple function harmonics based on the piecewise method, the harmonics of the PF converter under different operation modes are obtained. In order to examine the validity of the method, a simulation model is established based on Matlab/Simulink and a relevant experiment is implemented in the ITER PF integration test platform. Comparative results are given. The calculated results are found to be consistent with simulation and experiment. The piecewise method is proved correct and valid for calculating the system harmonics.

  11. Errors in retarding potential analyzers caused by nonuniformity of the grid-plane potential.

    NASA Technical Reports Server (NTRS)

    Hanson, W. B.; Frame, D. R.; Midgley, J. E.

    1972-01-01

    One aspect of the degradation in performance of retarding potential analyzers caused by potential depressions in the retarding grid is quantitatively estimated from laboratory measurements and theoretical calculations. A simple expression is obtained that permits the use of laboratory measurements of grid properties to make first-order corrections to flight data. Systematic positive errors in ion temperature of approximately 16% for the Ogo 4 instrument and 3% for the Ogo 6 instrument are deduced. The effects of the transverse electric fields arising from the grid potential depressions are not treated.

  12. VizieR Online Data Catalog: Giant HII regions BOND abundances (Vale Asari+, 2016)

    NASA Astrophysics Data System (ADS)

    Vale Asari, N.; Stasinska, G.; Morisset, C.; Cid Fernandes, R.

    2017-10-01

    BOND determines nitrogen and oxygen gas-phase abundances by using strong and semistrong lines and comparing them to a grid of photoionization models in a Bayesian framework. The code is written in python and its source is publicly available at http://bond.ufsc.br. The grid of models presented here is included in the 3MdB data base (Morisset, Delgado-Inglada & Flores-Fajardo 2015RMxAA..51..103M, see https://sites.google.com/site/mexicanmillionmodels/) under the reference 'BOND'. The Bayesian posterior probability calculated by bond stands on two pillars: our grid of models and our choice of observational constraints (from which we calculate our likelihoods). We discuss each of these in turn. (2 data files).

  13. SU-E-T-650: Quantification and Modeling of the Dosimetric Impact of the IBEAM Evo Treatment Couchtop EP (Elekta) in VMAT

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhang, R; Mannheim Medical Center, Mannheim, Baden-Wurttemberg; Bai, W

    2015-06-15

    Purpose: quantification and modelling of the dosimetric impact of the treatment couch in Monaco Treatment Planning System. Methods: The attenuation characteristics of couchtop EP was evaluated for two different photon acceleration potentials (6MV and 10MV) for a field size of (10×10) cm2. Phantom positions in A-B direction: on the left half, in the center and on the right half of the couch. Dose measurements of couch attenuation were performed at gantry angles from 180° to 122°, using a 0.125cc semiflex ionization chamber isocentrically placed in the center of a homogeneous cylindric sliced RW3 phantom. Each experimental setup was first measuredmore » on the LINAC and then reproduced in the TPS. By adjusting the relative-to-water electron density (ED) values of the couch, the measured attenuation was replicated. The simulated results were evaluated by comparing the measurements and simulations. Results: Without the couch model included the maximum difference between measured and calculated dose was 5.5% (5.1%) and 6.6% (6.1%) for 2 mm and 5 mm voxel size, when the phantom was positioned on the left (center). The couch model was included in the TPS with a uniform ED of 0.18 or a 2 component model with a fiber ED= 0.6 and foam core ED= 0.1. After including the treatment couch, the mean dose attenuation was reduced from 2.8% without couch included to (0.0, 0.8, −0.2, 0.6)%. The 4 different values represent the 1 and 2 components model and 2 and 5 mm voxel grid size. Conclusion: For a uniform relative-to-water couch electron density of 0.18 a good agreement between measured and calculated dose distributions was obtained for all different energies, voxel grid spacings and gantry angles. Therefore, we conclude that the Monaco couch model accurately describes the dose perturbations due to the presence of the patient couch and should therefore be used during treatment planning. This project is supported by Technology Foundation for Selected Overseas Chinese Scholar, Ministry of Hebei Personnel of China.« less

  14. Effect of Orthene on an unconfined population of the meadow vole (Microtus pennsylvanicus)

    USGS Publications Warehouse

    Jett, David A.; Nichols, James D.; Hines, James E.

    1986-01-01

    The possible impact on Microtus pennsylvanicus of ground applications of Orthene® insecticide was investigated in old-field habitats in northern Maryland during 1982 and 1983. The treatment grids in 1982 and 1983 were sprayed at 0.62 and 0.82 kg active ingredient/ha, respectively. A capture–recapture design robust to unequal capture probabilities was utilized to estimate population size, survival, and recruitment. Data on reproductive activity and relative weight change were also collected to investigate the effect of the insecticide treatment. There were no significant differences in population size or recruitment between control and treatment grids which could be directly related to the treatment. Survival rate was significantly lower on the treatment grid than on the control grid after spraying in 1983; however, survival rate was higher on the treatment grid after spraying in 1982. Significantly fewer pregnant adults were found on the treatment grid after spraying in 1982, whereas the proportions of voles lactating or with perforate vaginas or open pubic symphyses were slightly higher or remained unchanged during this period. Relative weight change was not affected by the treatment. Results do not indicate any pattern of inhibitory effects from the insecticide treatment. Field application of Orthene® did not have an adverse effect on this Microtus population.

  15. Sources of spurious force oscillations from an immersed boundary method for moving-body problems

    NASA Astrophysics Data System (ADS)

    Lee, Jongho; Kim, Jungwoo; Choi, Haecheon; Yang, Kyung-Soo

    2011-04-01

    When a discrete-forcing immersed boundary method is applied to moving-body problems, it produces spurious force oscillations on a solid body. In the present study, we identify two sources of these force oscillations. One source is from the spatial discontinuity in the pressure across the immersed boundary when a grid point located inside a solid body becomes that of fluid with a body motion. The addition of mass source/sink together with momentum forcing proposed by Kim et al. [J. Kim, D. Kim, H. Choi, An immersed-boundary finite volume method for simulations of flow in complex geometries, Journal of Computational Physics 171 (2001) 132-150] reduces the spurious force oscillations by alleviating this pressure discontinuity. The other source is from the temporal discontinuity in the velocity at the grid points where fluid becomes solid with a body motion. The magnitude of velocity discontinuity decreases with decreasing the grid spacing near the immersed boundary. Four moving-body problems are simulated by varying the grid spacing at a fixed computational time step and at a constant CFL number, respectively. It is found that the spurious force oscillations decrease with decreasing the grid spacing and increasing the computational time step size, but they depend more on the grid spacing than on the computational time step size.

  16. Point Cloud Oriented Shoulder Line Extraction in Loess Hilly Area

    NASA Astrophysics Data System (ADS)

    Min, Li; Xin, Yang; Liyang, Xiong

    2016-06-01

    Shoulder line is the significant line in hilly area of Loess Plateau in China, dividing the surface into positive and negative terrain (P-N terrains). Due to the point cloud vegetation removal methods of P-N terrains are different, there is an imperative need for shoulder line extraction. In this paper, we proposed an automatic shoulder line extraction method based on point cloud. The workflow is as below: (i) ground points were selected by using a grid filter in order to remove most of noisy points. (ii) Based on DEM interpolated by those ground points, slope was mapped and classified into two classes (P-N terrains), using Natural Break Classified method. (iii) The common boundary between two slopes is extracted as shoulder line candidate. (iv) Adjust the filter gird size and repeat step i-iii until the shoulder line candidate matches its real location. (v) Generate shoulder line of the whole area. Test area locates in Madigou, Jingbian County of Shaanxi Province, China. A total of 600 million points are acquired in the test area of 0.23km2, using Riegl VZ400 3D Laser Scanner in August 2014. Due to the limit Granted computing performance, the test area is divided into 60 blocks and 13 of them around the shoulder line were selected for filter grid size optimizing. The experiment result shows that the optimal filter grid size varies in diverse sample area, and a power function relation exists between filter grid size and point density. The optimal grid size was determined by above relation and shoulder lines of 60 blocks were then extracted. Comparing with the manual interpretation results, the accuracy of the whole result reaches 85%. This method can be applied to shoulder line extraction in hilly area, which is crucial for point cloud denoising and high accuracy DEM generation.

  17. Chimera Grid Tools

    NASA Technical Reports Server (NTRS)

    Chan, William M.; Rogers, Stuart E.; Nash, Steven M.; Buning, Pieter G.; Meakin, Robert

    2005-01-01

    Chimera Grid Tools (CGT) is a software package for performing computational fluid dynamics (CFD) analysis utilizing the Chimera-overset-grid method. For modeling flows with viscosity about geometrically complex bodies in relative motion, the Chimera-overset-grid method is among the most computationally cost-effective methods for obtaining accurate aerodynamic results. CGT contains a large collection of tools for generating overset grids, preparing inputs for computer programs that solve equations of flow on the grids, and post-processing of flow-solution data. The tools in CGT include grid editing tools, surface-grid-generation tools, volume-grid-generation tools, utility scripts, configuration scripts, and tools for post-processing (including generation of animated images of flows and calculating forces and moments exerted on affected bodies). One of the tools, denoted OVERGRID, is a graphical user interface (GUI) that serves to visualize the grids and flow solutions and provides central access to many other tools. The GUI facilitates the generation of grids for a new flow-field configuration. Scripts that follow the grid generation process can then be constructed to mostly automate grid generation for similar configurations. CGT is designed for use in conjunction with a computer-aided-design program that provides the geometry description of the bodies, and a flow-solver program.

  18. Elliptic generation of composite three-dimensional grids about realistic aircraft

    NASA Technical Reports Server (NTRS)

    Sorenson, R. L.

    1986-01-01

    An elliptic method for generating composite grids about realistic aircraft is presented. A body-conforming grid is first generated about the entire aircraft by the solution of Poisson's differential equation. This grid has relatively coarse spacing, and it covers the entire physical domain. At boundary surfaces, cell size is controlled and cell skewness is nearly eliminated by inhomogeneous terms, which are found automatically by the program. Certain regions of the grid in which high gradients are expected, and which map into rectangular solids in the computational domain, are then designated for zonal refinement. Spacing in the zonal grids is reduced by adding points with a simple, algebraic scheme. Details of the grid generation method are presented along with results of the present application, a wing-body configuration based on the F-16 fighter aircraft.

  19. Influence of multiple brain metastases’ size and number on the quality of SRS - VMAT dose delivery

    NASA Astrophysics Data System (ADS)

    Prentou, G.; Koutsouveli, E.; Pantelis, E.; Papagiannis, P.; Georgiou, E.; Karaiskos, P.

    2017-11-01

    Stereotactic radiosurgery with volumetric modulated arc therapy (SRS-VMAT) has recently been introduced for treatment of multiple brain metastases with a single isocenter. The technique’s high efficiency is nevertheless dependent of metastatic tumors’ characteristics such as size and number. In this work the impact of the metastases’ size and number on the plan quality indices clinically used for plan evaluation and acceptance is investigated. Fifteen targets with a diameter of 1 cm and average volume of 0.7 cm3 and ten targets with a diameter of 2 cm and average volume of 6.5 cm3 were contoured on an anonymized patient CT dataset, in Monaco (Elekta) treatment planning system. VMAT plans for different target volumes (1 and 2 cm in diameter) and various target numbers (1-15) were generated using four non-coplanar arcs and the Agility (Elekta) linear accelerator (5 mm MLC width) using a Monte Carlo dose calculation algorithm and 1mm dose calculation grid resolution. Conformity index (CI), gradient index (GI) and heterogeneity index (HI) were determined for each target. High quality plans were created for both 1 cm and 2 cm in diameter targets for limited (<6) number of targets per plan. For increased number of irradiated targets (>6) both CI and GI, clinically used for plan evaluation and acceptance, were found to deteriorate.

  20. Statistical variability and confidence intervals for planar dose QA pass rates

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bailey, Daniel W.; Nelms, Benjamin E.; Attwood, Kristopher

    Purpose: The most common metric for comparing measured to calculated dose, such as for pretreatment quality assurance of intensity-modulated photon fields, is a pass rate (%) generated using percent difference (%Diff), distance-to-agreement (DTA), or some combination of the two (e.g., gamma evaluation). For many dosimeters, the grid of analyzed points corresponds to an array with a low areal density of point detectors. In these cases, the pass rates for any given comparison criteria are not absolute but exhibit statistical variability that is a function, in part, on the detector sampling geometry. In this work, the authors analyze the statistics ofmore » various methods commonly used to calculate pass rates and propose methods for establishing confidence intervals for pass rates obtained with low-density arrays. Methods: Dose planes were acquired for 25 prostate and 79 head and neck intensity-modulated fields via diode array and electronic portal imaging device (EPID), and matching calculated dose planes were created via a commercial treatment planning system. Pass rates for each dose plane pair (both centered to the beam central axis) were calculated with several common comparison methods: %Diff/DTA composite analysis and gamma evaluation, using absolute dose comparison with both local and global normalization. Specialized software was designed to selectively sample the measured EPID response (very high data density) down to discrete points to simulate low-density measurements. The software was used to realign the simulated detector grid at many simulated positions with respect to the beam central axis, thereby altering the low-density sampled grid. Simulations were repeated with 100 positional iterations using a 1 detector/cm{sup 2} uniform grid, a 2 detector/cm{sup 2} uniform grid, and similar random detector grids. For each simulation, %/DTA composite pass rates were calculated with various %Diff/DTA criteria and for both local and global %Diff normalization techniques. Results: For the prostate and head/neck cases studied, the pass rates obtained with gamma analysis of high density dose planes were 2%-5% higher than respective %/DTA composite analysis on average (ranging as high as 11%), depending on tolerances and normalization. Meanwhile, the pass rates obtained via local normalization were 2%-12% lower than with global maximum normalization on average (ranging as high as 27%), depending on tolerances and calculation method. Repositioning of simulated low-density sampled grids leads to a distribution of possible pass rates for each measured/calculated dose plane pair. These distributions can be predicted using a binomial distribution in order to establish confidence intervals that depend largely on the sampling density and the observed pass rate (i.e., the degree of difference between measured and calculated dose). These results can be extended to apply to 3D arrays of detectors, as well. Conclusions: Dose plane QA analysis can be greatly affected by choice of calculation metric and user-defined parameters, and so all pass rates should be reported with a complete description of calculation method. Pass rates for low-density arrays are subject to statistical uncertainty (vs. the high-density pass rate), but these sampling errors can be modeled using statistical confidence intervals derived from the sampled pass rate and detector density. Thus, pass rates for low-density array measurements should be accompanied by a confidence interval indicating the uncertainty of each pass rate.« less

  1. PySE: Python Source Extractor for radio astronomical images

    NASA Astrophysics Data System (ADS)

    Spreeuw, Hanno; Swinbank, John; Molenaar, Gijs; Staley, Tim; Rol, Evert; Sanders, John; Scheers, Bart; Kuiack, Mark

    2018-05-01

    PySE finds and measures sources in radio telescope images. It is run with several options, such as the detection threshold (a multiple of the local noise), grid size, and the forced clean beam fit, followed by a list of input image files in standard FITS or CASA format. From these, PySe provides a list of found sources; information such as the calculated background image, source list in different formats (e.g. text, region files importable in DS9), and other data may be saved. PySe can be integrated into a pipeline; it was originally written as part of the LOFAR Transient Detection Pipeline (TraP, ascl:1412.011).

  2. Extending CFD Modeling to Near-Continuum Flows Using Enhanced Thermophysical Modeling

    DTIC Science & Technology

    2008-03-01

    the fluid and cell size. Information propagates through the grid through lo - cal calculation of the convective derivative. In implicit-integration CFD...3.14) 29 and SR = max [ λm(U Roe), λm(Ur) ] (3.15) Also, URoe ≡ Q̂ (3.16) λ1 = Ũl − al → λ1(Q̂) = ŨRoe − aRoe (3.17) and λm = Ũ + a (3.18) Starred...Ũi − S ∗) + pi. (3.29) 31 The Roe-Averaged Variables are defined by R = √ ρR ρL (3.30) ρRoe = RρL (3.31) uRoe = RuR + uL 1 + R (3.32) vRoe = RvR

  3. Juncture flow improvement for wing/pylon configurations by using CFD methodology

    NASA Technical Reports Server (NTRS)

    Gea, Lie-Mine; Chyu, Wei J.; Stortz, Michael W.; Chow, Chuen-Yen

    1993-01-01

    Transonic flow field around a fighter wing/pylon configuration was simulated by using an implicit upwinding Navier-Stokes flow solver (F3D) and overset grid technology (Chimera). Flow separation and local shocks near the wing/pylon junction were observed in flight and predicted by numerical calculations. A new pylon/fairing shape was proposed to improve the flow quality. Based on numerical results, the size of separation area is significantly reduced and the onset of separation is delayed farther downstream. A smoother pressure gradient is also obtained near the junction area. This paper demonstrates that computational fluid dynamics (CFD) methodology can be used as a practical tool for aircraft design.

  4. Numerical Analysis of Projectile Impact and Deep Penetration into Earth Media

    DTIC Science & Technology

    1975-08-01

    Soil ) :ind ?.vr2 (Shai h.) SECTION III COMPII1TAT IONAL METIIOD 3.1 lAVF-L CODlP The NA \\VL code was employed for these calculations. WAV.-L is a...8217,gh the computational grid. For these calculations, the prtjectile/target interface *a- assumed to be frictionless. A lithostatic field in the soil was...generated by pre’scribing initially compressed soil states. .3.2 GRID DECOUPLiN; The sliding interface formulation in WAVE-L includes the capability

  5. Aspects and applications of patched grid calculations

    NASA Technical Reports Server (NTRS)

    Walters, R. W.; Switzer, G. F.; Thomas, J. L.

    1986-01-01

    Patched grid calculations within the framework of an implicit, flux-vector split upwind/relaxation algorithm for the Euler equations are presented. The effect of a metric-discontinuous interface on the convergence rate of the algorithm is discussed along with the spatial accuracy of the solution and the effect of curvature along an interface. Results are presented and discussed for the free-stream problem, shock reflection problem, supersonic inlet with a 5 degree ramp, aerodynamically choked inlet, and three-dimensional analytic forebody.

  6. Combined effect of pulse density and grid cell size on predicting and mapping aboveground carbon in fast‑growing Eucalyptus forest plantation using airborne LiDAR data

    Treesearch

    Carlos Alberto Silva; Andrew Thomas Hudak; Carine Klauberg; Lee Alexandre Vierling; Carlos Gonzalez‑Benecke; Samuel de Padua Chaves Carvalho; Luiz Carlos Estraviz Rodriguez; Adrian Cardil

    2017-01-01

    LiDAR measurements can be used to predict and map AGC across variable-age Eucalyptus plantations with adequate levels of precision and accuracy using 5 pulses m− 2 and a grid cell size of 5 m. The promising results for AGC modeling in this study will allow for greater confidence in comparing AGC estimates with varying LiDAR sampling densities for Eucalyptus plantations...

  7. Adaptive beamlet-based finite-size pencil beam dose calculation for independent verification of IMRT and VMAT.

    PubMed

    Park, Justin C; Li, Jonathan G; Arhjoul, Lahcen; Yan, Guanghua; Lu, Bo; Fan, Qiyong; Liu, Chihray

    2015-04-01

    The use of sophisticated dose calculation procedure in modern radiation therapy treatment planning is inevitable in order to account for complex treatment fields created by multileaf collimators (MLCs). As a consequence, independent volumetric dose verification is time consuming, which affects the efficiency of clinical workflow. In this study, the authors present an efficient adaptive beamlet-based finite-size pencil beam (AB-FSPB) dose calculation algorithm that minimizes the computational procedure while preserving the accuracy. The computational time of finite-size pencil beam (FSPB) algorithm is proportional to the number of infinitesimal and identical beamlets that constitute an arbitrary field shape. In AB-FSPB, dose distribution from each beamlet is mathematically modeled such that the sizes of beamlets to represent an arbitrary field shape no longer need to be infinitesimal nor identical. As a result, it is possible to represent an arbitrary field shape with combinations of different sized and minimal number of beamlets. In addition, the authors included the model parameters to consider MLC for its rounded edge and transmission. Root mean square error (RMSE) between treatment planning system and conventional FSPB on a 10 × 10 cm(2) square field using 10 × 10, 2.5 × 2.5, and 0.5 × 0.5 cm(2) beamlet sizes were 4.90%, 3.19%, and 2.87%, respectively, compared with RMSE of 1.10%, 1.11%, and 1.14% for AB-FSPB. This finding holds true for a larger square field size of 25 × 25 cm(2), where RMSE for 25 × 25, 2.5 × 2.5, and 0.5 × 0.5 cm(2) beamlet sizes were 5.41%, 4.76%, and 3.54% in FSPB, respectively, compared with RMSE of 0.86%, 0.83%, and 0.88% for AB-FSPB. It was found that AB-FSPB could successfully account for the MLC transmissions without major discrepancy. The algorithm was also graphical processing unit (GPU) compatible to maximize its computational speed. For an intensity modulated radiation therapy (∼12 segments) and a volumetric modulated arc therapy fields (∼90 control points) with a 3D grid size of 2.0 × 2.0 × 2.0 mm(3), dose was computed within 3-5 and 10-15 s timeframe, respectively. The authors have developed an efficient adaptive beamlet-based pencil beam dose calculation algorithm. The fast computation nature along with GPU compatibility has shown better performance than conventional FSPB. This enables the implementation of AB-FSPB in the clinical environment for independent volumetric dose verification.

  8. Implication of observed cloud variability for parameterizations of microphysical and radiative transfer processes in climate models

    NASA Astrophysics Data System (ADS)

    Huang, D.; Liu, Y.

    2014-12-01

    The effects of subgrid cloud variability on grid-average microphysical rates and radiative fluxes are examined by use of long-term retrieval products at the Tropical West Pacific (TWP), Southern Great Plains (SGP), and North Slope of Alaska (NSA) sites of the Department of Energy's Atmospheric Radiation Measurement (ARM) Program. Four commonly used distribution functions, the truncated Gaussian, Gamma, lognormal, and Weibull distributions, are constrained to have the same mean and standard deviation as observed cloud liquid water content. The PDFs are then used to upscale relevant physical processes to obtain grid-average process rates. It is found that the truncated Gaussian representation results in up to 30% mean bias in autoconversion rate whereas the mean bias for the lognormal representation is about 10%. The Gamma and Weibull distribution function performs the best for the grid-average autoconversion rate with the mean relative bias less than 5%. For radiative fluxes, the lognormal and truncated Gaussian representations perform better than the Gamma and Weibull representations. The results show that the optimal choice of subgrid cloud distribution function depends on the nonlinearity of the process of interest and thus there is no single distribution function that works best for all parameterizations. Examination of the scale (window size) dependence of the mean bias indicates that the bias in grid-average process rates monotonically increases with increasing window sizes, suggesting the increasing importance of subgrid variability with increasing grid sizes.

  9. Study on the influence of three-grid assembly thermal deformation on breakdown times and an ion extraction process

    NASA Astrophysics Data System (ADS)

    Mingming, SUN; Yanhui, JIA; Yongjie, HUANG; Juntai, YANG; Xiaodong, WEN; Meng, WANG

    2018-04-01

    In order to study the influence of three-grid assembly thermal deformation caused by heat accumulation on breakdown times and an ion extraction process, a hot gap test and a breakdown time test are carried out to obtain thermal deformation of the grids when the thruster is in 5 kW operation mode. Meanwhile, the fluid simulation method and particle-in-cell-Monte Carlo collision (PIC-MCC) method are adopted to simulate the ion extraction process according to the previous test results. The numerical calculation results are verified by the ion thruster performance test. The results show that after about 1.2 h operation, the hot gap between the screen grid and the accelerator grid reduce to 0.25–0.3 mm, while the hot gap between the accelerator grid and the decelerator grid increase from 1 mm to about 1.4 mm when the grids reach thermal equilibrium, and the hot gap is almost unchanged. In addition, the breakdown times experiment shows that 0.26 mm is the minimal safe hot gap for the grid assembly as the breakdown times improves significantly when the gap is smaller than this value. Fluid simulation results show that the plasma density of the screen grid is in the range 6 × 1017–6 × 1018 m13 and displays a parabolic characteristic, while the electron temperature gradually increases along the axial direction. The PIC-MCC results show that the current falling of an ion beam through a single aperture is significant. Meanwhile, the intercepted current of the accelerator grid and the decelerator grid both increase with the change in the hot gap. The ion beam current has optimal perveance status without thermal deformation, and the intercepted current of the accelerator grid and the decelerator grid are 3.65 mA and 6.26 mA, respectively. Furthermore, under the effect of thermal deformation, the ion beam current has over-perveance status, and the intercepted current of the accelerator grid and the decelerator grid are 10.46 mA and 18.24 mA, respectively. Performance test results indicate that the breakdown times increase obviously. The intercepted current of the accelerator grid and the decelerator grid increases to 13 mA and 16.5 mA, respectively, due to the change in the hot gap after 1.5 h operation. The numerical calculation results are well consistent with performance test results, and the error comes mainly from the test uncertainty of the hot gap.

  10. Control strategy based on SPWM switching patterns for grid connected photovoltaic inverter

    NASA Astrophysics Data System (ADS)

    Hassaine, L.; Mraoui, A.

    2017-02-01

    Generally, for lower installation of photovoltaic systems connected to the grid, pulse width modulation (PWM) is a widely used technique for controlling the voltage source inverters injects currents into the grid. The current injected must be sinusoidal with reduced harmonic distortion. In this paper, a digital implementation of a control strategy based on PWM switching patterns for an inverter for photovoltaic system connected to the grid is presented. This strategy synchronize a sinusoidal inverter output current with a grid voltage The digital implementation of the proposed PWM switching pattern when is compared with the conventional one exhibit the advantage: Simplicity, reduction of the memory requirements and power calculation for the control

  11. Investigation of Advanced Counterrotation Blade Configuration Concepts for High Speed Turboprop Systems. Task 2: Unsteady Ducted Propfan Analysis

    NASA Technical Reports Server (NTRS)

    Hall, Edward J.; Delaney, Robert A.; Bettner, James L.

    1991-01-01

    The primary objective was the development of a time dependent 3-D Euler/Navier-Stokes aerodynamic analysis to predict unsteady compressible transonic flows about ducted and unducted propfan propulsion systems at angle of attack. The resulting computer codes are referred to as Advanced Ducted Propfan Analysis Codes (ADPAC). A computer program user's manual is presented for the ADPAC. Aerodynamic calculations were based on a four stage Runge-Kutta time marching finite volume solution technique with added numerical dissipation. A time accurate implicit residual smoothing operator was used for unsteady flow predictions. For unducted propfans, a single H-type grid was used to discretize each blade passage of the complete propeller. For ducted propfans, a coupled system of five grid blocks utilizing an embedded C grid about the cowl leading edge was used to discretize each blade passage. Grid systems were generated by a combined algebraic/elliptic algorithm developed specifically for ducted propfans. Numerical calculations were compared with experimental data for both ducted and unducted flows.

  12. Power transformations improve interpolation of grids for molecular mechanics interaction energies.

    PubMed

    Minh, David D L

    2018-02-18

    A common strategy for speeding up molecular docking calculations is to precompute nonbonded interaction energies between a receptor molecule and a set of three-dimensional grids. The grids are then interpolated to compute energies for ligand atoms in many different binding poses. Here, I evaluate a smoothing strategy of taking a power transformation of grid point energies and inverse transformation of the result from trilinear interpolation. For molecular docking poses from 85 protein-ligand complexes, this smoothing procedure leads to significant accuracy improvements, including an approximately twofold reduction in the root mean square error at a grid spacing of 0.4 Å and retaining the ability to rank docking poses even at a grid spacing of 0.7 Å. © 2018 Wiley Periodicals, Inc. © 2018 Wiley Periodicals, Inc.

  13. Nonuniform grid implicit spatial finite difference method for acoustic wave modeling in tilted transversely isotropic media

    NASA Astrophysics Data System (ADS)

    Chu, Chunlei; Stoffa, Paul L.

    2012-01-01

    Discrete earth models are commonly represented by uniform structured grids. In order to ensure accurate numerical description of all wave components propagating through these uniform grids, the grid size must be determined by the slowest velocity of the entire model. Consequently, high velocity areas are always oversampled, which inevitably increases the computational cost. A practical solution to this problem is to use nonuniform grids. We propose a nonuniform grid implicit spatial finite difference method which utilizes nonuniform grids to obtain high efficiency and relies on implicit operators to achieve high accuracy. We present a simple way of deriving implicit finite difference operators of arbitrary stencil widths on general nonuniform grids for the first and second derivatives and, as a demonstration example, apply these operators to the pseudo-acoustic wave equation in tilted transversely isotropic (TTI) media. We propose an efficient gridding algorithm that can be used to convert uniformly sampled models onto vertically nonuniform grids. We use a 2D TTI salt model to demonstrate its effectiveness and show that the nonuniform grid implicit spatial finite difference method can produce highly accurate seismic modeling results with enhanced efficiency, compared to uniform grid explicit finite difference implementations.

  14. Recent Developments in OVERGRID, OVERFLOW-2 and Chimera Grid Tools Scripts

    NASA Technical Reports Server (NTRS)

    Chan, William M.

    2004-01-01

    OVERGRID and OVERFLOW-2 feature easy to use multiple-body dynamics. The new features of OVERGRID include a preliminary chemistry interface, standard atmosphere and mass properties calculators, a simple unsteady solution viewer, and a debris tracking interface. Script library development in Chimera Grid Tools has applications in turbopump grid generation. This viewgraph presentation profiles multiple component dynamics, validation test cases for a sphere, cylinder, and oscillating airfoil, and debris analysis.

  15. Current Grid Generation Strategies and Future Requirements in Hypersonic Vehicle Design, Analysis and Testing

    NASA Technical Reports Server (NTRS)

    Papadopoulos, Periklis; Venkatapathy, Ethiraj; Prabhu, Dinesh; Loomis, Mark P.; Olynick, Dave; Arnold, James O. (Technical Monitor)

    1998-01-01

    Recent advances in computational power enable computational fluid dynamic modeling of increasingly complex configurations. A review of grid generation methodologies implemented in support of the computational work performed for the X-38 and X-33 are presented. In strategizing topological constructs and blocking structures factors considered are the geometric configuration, optimal grid size, numerical algorithms, accuracy requirements, physics of the problem at hand, computational expense, and the available computer hardware. Also addressed are grid refinement strategies, the effects of wall spacing, and convergence. The significance of grid is demonstrated through a comparison of computational and experimental results of the aeroheating environment experienced by the X-38 vehicle. Special topics on grid generation strategies are also addressed to model control surface deflections, and material mapping.

  16. Polarizing Grids, their Assemblies and Beams of Radiation

    NASA Technical Reports Server (NTRS)

    Houde, Martin; Akeson, Rachel L.; Carlstrom, John E.; Lamb, James W.; Schleuning, David A.; Woody, David P.

    2001-01-01

    This article gives an analysis of the behavior of polarizing grids and reflecting polarizers by solving Maxwell's equations, for arbitrary angles of incidence and grid rotation, for cases where the excitation is provided by an incident plane wave or a beam of radiation. The scattering and impedance matrix representations are derived and used to solve more complicated configurations of grid assemblies. The results are also compared with data obtained in the calibration of reflecting polarizers at the Owens Valley Radio Observatory (OVRO). From these analysis, we propose a method for choosing the optimum grid parameters (wire radius and spacing). We also provide a study of the effects of two types of errors (in wire separation and radius size) that can be introduced in the fabrication of a grid.

  17. Regional photochemical air quality modeling in the Mexico-US border area

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mendoza, A.; Russell, A.G.; Mejia, G.M.

    1998-12-31

    The Mexico-United States border area has become an increasingly important region due to its commercial, industrial and urban growth. As a result, environmental concerns have risen. Treaties like the North American Free Trade Agreement (NAFTA) have further motivated the development of environmental impact assessment in the area. Of particular concern are air quality, and how the activities on both sides of the border contribute to its degradation. This paper presents results of applying a three-dimensional photochemical airshed model to study air pollution dynamics along the Mexico-United States border. In addition, studies were conducted to assess how size resolution impacts themore » model performance. The model performed within acceptable statistic limits using 12.5 x 12.5 km{sup 2} grid cells, and the benefits using finer grids were limited. Results were further used to assess the influence of grid-cell size on the modeling of control strategies, where coarser grids lead to significant loss of information.« less

  18. Mathematical Aspects of Finite Element Methods for Incompressible Viscous Flows.

    DTIC Science & Technology

    1986-09-01

    respectively. Here h is a parameter which is usually related to the size of the grid associated with the finite element partitioning of Q. Then one... grid and of not at least performing serious mesh refinement studies. It also points out the usefulness of rigorous results concerning the stability...overconstrained the .1% approximate velocity field. However, by employing different grids for the ’z pressure and velocity fields, the linear-constant

  19. Analysis of Franck-Condon factors for CO+ molecule using the Fourier Grid Hamiltonian method

    NASA Astrophysics Data System (ADS)

    Syiemiong, Arnestar; Swer, Shailes; Jha, Ashok Kumar; Saxena, Atul

    2018-04-01

    Franck-Condon factors (FCFs) are important parameters and it plays a very important role in determining the intensities of the vibrational bands in electronic transitions. In this paper, we illustrate the Fourier Grid Hamiltonian (FGH) method, a relatively simple method to calculate the FCFs. The FGH is a method used for calculating the vibrational eigenvalues and eigenfunctions of bound electronic states of diatomic molecules. The obtained vibrational wave functions for the ground and the excited states are used to calculate the vibrational overlap integral and then the FCFs. In this computation, we used the Morse potential and Bi-Exponential potential model for constructing and diagonalizing the molecular Hamiltonians. The effects of the change in equilibrium internuclear distance (xe), dissociation energy (De), and the nature of the excited state electronic energy curve on the FCFs have been determined. Here we present our work for the qualitative analysis of Franck-Condon Factorsusing this Fourier Grid Hamiltonian Method.

  20. An overlapped grid method for multigrid, finite volume/difference flow solvers: MaGGiE

    NASA Technical Reports Server (NTRS)

    Baysal, Oktay; Lessard, Victor R.

    1990-01-01

    The objective is to develop a domain decomposition method via overlapping/embedding the component grids, which is to be used by upwind, multi-grid, finite volume solution algorithms. A computer code, given the name MaGGiE (Multi-Geometry Grid Embedder) is developed to meet this objective. MaGGiE takes independently generated component grids as input, and automatically constructs the composite mesh and interpolation data, which can be used by the finite volume solution methods with or without multigrid convergence acceleration. Six demonstrative examples showing various aspects of the overlap technique are presented and discussed. These cases are used for developing the procedure for overlapping grids of different topologies, and to evaluate the grid connection and interpolation data for finite volume calculations on a composite mesh. Time fluxes are transferred between mesh interfaces using a trilinear interpolation procedure. Conservation losses are minimal at the interfaces using this method. The multi-grid solution algorithm, using the coaser grid connections, improves the convergence time history as compared to the solution on composite mesh without multi-gridding.

  1. Sedimentary Geothermal Feasibility Study: October 2016

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Augustine, Chad; Zerpa, Luis

    The objective of this project is to analyze the feasibility of commercial geothermal projects using numerical reservoir simulation, considering a sedimentary reservoir with low permeability that requires productivity enhancement. A commercial thermal reservoir simulator (STARS, from Computer Modeling Group, CMG) is used in this work for numerical modeling. In the first stage of this project (FY14), a hypothetical numerical reservoir model was developed, and validated against an analytical solution. The following model parameters were considered to obtain an acceptable match between the numerical and analytical solutions: grid block size, time step and reservoir areal dimensions; the latter related to boundarymore » effects on the numerical solution. Systematic model runs showed that insufficient grid sizing generates numerical dispersion that causes the numerical model to underestimate the thermal breakthrough time compared to the analytic model. As grid sizing is decreased, the model results converge on a solution. Likewise, insufficient reservoir model area introduces boundary effects in the numerical solution that cause the model results to differ from the analytical solution.« less

  2. PSO-Based Smart Grid Application for Sizing and Optimization of Hybrid Renewable Energy Systems

    PubMed Central

    Mohamed, Mohamed A.; Eltamaly, Ali M.; Alolah, Abdulrahman I.

    2016-01-01

    This paper introduces an optimal sizing algorithm for a hybrid renewable energy system using smart grid load management application based on the available generation. This algorithm aims to maximize the system energy production and meet the load demand with minimum cost and highest reliability. This system is formed by photovoltaic array, wind turbines, storage batteries, and diesel generator as a backup source of energy. Demand profile shaping as one of the smart grid applications is introduced in this paper using load shifting-based load priority. Particle swarm optimization is used in this algorithm to determine the optimum size of the system components. The results obtained from this algorithm are compared with those from the iterative optimization technique to assess the adequacy of the proposed algorithm. The study in this paper is performed in some of the remote areas in Saudi Arabia and can be expanded to any similar regions around the world. Numerous valuable results are extracted from this study that could help researchers and decision makers. PMID:27513000

  3. PSO-Based Smart Grid Application for Sizing and Optimization of Hybrid Renewable Energy Systems.

    PubMed

    Mohamed, Mohamed A; Eltamaly, Ali M; Alolah, Abdulrahman I

    2016-01-01

    This paper introduces an optimal sizing algorithm for a hybrid renewable energy system using smart grid load management application based on the available generation. This algorithm aims to maximize the system energy production and meet the load demand with minimum cost and highest reliability. This system is formed by photovoltaic array, wind turbines, storage batteries, and diesel generator as a backup source of energy. Demand profile shaping as one of the smart grid applications is introduced in this paper using load shifting-based load priority. Particle swarm optimization is used in this algorithm to determine the optimum size of the system components. The results obtained from this algorithm are compared with those from the iterative optimization technique to assess the adequacy of the proposed algorithm. The study in this paper is performed in some of the remote areas in Saudi Arabia and can be expanded to any similar regions around the world. Numerous valuable results are extracted from this study that could help researchers and decision makers.

  4. Facial recognition using simulated prosthetic pixelized vision.

    PubMed

    Thompson, Robert W; Barnett, G David; Humayun, Mark S; Dagnelie, Gislin

    2003-11-01

    To evaluate a model of simulated pixelized prosthetic vision using noncontiguous circular phosphenes, to test the effects of phosphene and grid parameters on facial recognition. A video headset was used to view a reference set of four faces, followed by a partially averted image of one of those faces viewed through a square pixelizing grid that contained 10x10 to 32x32 dots separated by gaps. The grid size, dot size, gap width, dot dropout rate, and gray-scale resolution were varied separately about a standard test condition, for a total of 16 conditions. All tests were first performed at 99% contrast and then repeated at 12.5% contrast. Discrimination speed and performance were influenced by all stimulus parameters. The subjects achieved highly significant facial recognition accuracy for all high-contrast tests except for grids with 70% random dot dropout and two gray levels. In low-contrast tests, significant facial recognition accuracy was achieved for all but the most adverse grid parameters: total grid area less than 17% of the target image, 70% dropout, four or fewer gray levels, and a gap of 40.5 arcmin. For difficult test conditions, a pronounced learning effect was noticed during high-contrast trials, and a more subtle practice effect on timing was evident during subsequent low-contrast trials. These findings suggest that reliable face recognition with crude pixelized grids can be learned and may be possible, even with a crude visual prosthesis.

  5. Post-Test Inspection of NASA's Evolutionary Xenon Thruster Long-Duration Test Hardware: Discharge Chamber

    NASA Technical Reports Server (NTRS)

    Shastry, Rohit; Soulas, George C.

    2016-01-01

    The NEXT Long-Duration Test is part of a comprehensive thruster service life assessment intended to demonstrate overall throughput capability, validate service life models, quantify wear rates as a function of time and operating condition, and identify any unknown life-limiting mechanisms. The test was voluntarily terminated in February 2014 after demonstrating 51,184 hours of high-voltage operation, 918 kg of propellant throughput, and 35.5 MN-s of total impulse. The post-test inspection of the thruster hardware began shortly afterwards with a combination of non-destructive and destructive analysis techniques, and is presently nearing completion. This paper presents relevant results of the post-test inspection for the discharge chamber as well as other miscellaneous components such as the high-voltage propellant isolators and electrical cabling. Comparison of magnetic field measurements taken during pretest and post-test inspections indicate that the field strength did not degrade, consistent with performance data obtained during the test. Inspection of discharge chamber mesh samples show a deposition coating primarily composed of grid material that is approximately 15 micrometers in thickness. This thickness is well within the retention capability of the mesh and is therefore not expected to present any issues. Approximately 3.1 grams of deposition flakes were found at the bottom of the discharge chamber, composed primarily of grid material and carbon. Calculated size histograms of these flakes indicate that 99% have a maximum dimension of 200 micrometers or smaller, which is significantly less than the ion optics grid gap. Larger flakes that are capable of causing a grid-to-grid short will be analyzed to determine if their formation will occur in flight or is a facility effect. The high-voltage propellant isolators as well as numerous other electrical insulators were inspected and no evidence of arcing or any other issues were found.

  6. Characterizing the digital radiography system in terms of effective detective quantum efficiency and CDRAD measurement

    NASA Astrophysics Data System (ADS)

    Yalcin, A.; Olgar, T.

    2018-07-01

    The aim of this study was to assess the performance of a digital radiography system in terms of effective detective quantum efficiency (eDQE) for different tube voltages, polymethyl methacrylate (PMMA) phantom thicknesses and different grid types. The image performance of the digital radiography system was also evaluated by using CDRAD measurements at the same conditions and the correlation of CDRAD results with eDQE was compared. The eDQE was calculated via measurement of effective modulation transfer function (eMTF), effective normalized noise power spectra (eNNPS), scatter fraction (SF) and transmission factors (TF). SFs and TFs were also calculated for different beam qualities by using MCNP4C Monte Carlo simulation code. The integrated eDQE (IeDQE) over the frequency range was used to find the correlation with the inverse image quality figure (IQFinv) obtained from CDRAD measurements. The highest eDQE was obtained with 60 lp/cm grid frequency and 10:1 grid ratio. No remarkable effect was observed on eDQE with different grid frequency, but eDQE decreased with increasing grid ratio. A significant correlation was found between IeDQE and IQFinv.

  7. Uncertainties in estimates of mortality attributable to ambient PM2.5 in Europe

    NASA Astrophysics Data System (ADS)

    Kushta, Jonilda; Pozzer, Andrea; Lelieveld, Jos

    2018-06-01

    The assessment of health impacts associated with airborne particulate matter smaller than 2.5 μm in diameter (PM2.5) relies on aerosol concentrations derived either from monitoring networks, satellite observations, numerical models, or a combination thereof. When global chemistry-transport models are used for estimating PM2.5, their relatively coarse resolution has been implied to lead to underestimation of health impacts in densely populated and industrialized areas. In this study the role of spatial resolution and of vertical layering of a regional air quality model, used to compute PM2.5 impacts on public health and mortality, is investigated. We utilize grid spacings of 100 km and 20 km to calculate annual mean PM2.5 concentrations over Europe, which are in turn applied to the estimation of premature mortality by cardiovascular and respiratory diseases. Using model results at a 100 km grid resolution yields about 535 000 annual premature deaths over the extended European domain (242 000 within the EU-28), while numbers approximately 2.4% higher are derived by using the 20 km resolution. Using the surface (i.e. lowest) layer of the model for PM2.5 yields about 0.6% higher mortality rates compared with PM2.5 averaged over the first 200 m above ground. Further, the calculation of relative risks (RR) from PM2.5, using 0.1 μg m‑3 size resolution bins compared to the commonly used 1 μg m‑3, is associated with ±0.8% uncertainty in estimated deaths. We conclude that model uncertainties contribute a small part of the overall uncertainty expressed by the 95% confidence intervals, which are of the order of ±30%, mostly related to the RR calculations based on epidemiological data.

  8. A Cell-Centered Multigrid Algorithm for All Grid Sizes

    NASA Technical Reports Server (NTRS)

    Gjesdal, Thor

    1996-01-01

    Multigrid methods are optimal; that is, their rate of convergence is independent of the number of grid points, because they use a nested sequence of coarse grids to represent different scales of the solution. This nesting does, however, usually lead to certain restrictions of the permissible size of the discretised problem. In cases where the modeler is free to specify the whole problem, such constraints are of little importance because they can be taken into consideration from the outset. We consider the situation in which there are other competing constraints on the resolution. These restrictions may stem from the physical problem (e.g., if the discretised operator contains experimental data measured on a fixed grid) or from the need to avoid limitations set by the hardware. In this paper we discuss a modification to the cell-centered multigrid algorithm, so that it can be used br problems with any resolution. We discuss in particular a coarsening strategy and choice of intergrid transfer operators that can handle grids with both an even or odd number of cells. The method is described and applied to linear equations obtained by discretization of two- and three-dimensional second-order elliptic PDEs.

  9. From grid cells to place cells with realistic field sizes

    PubMed Central

    2017-01-01

    While grid cells in the medial entorhinal cortex (MEC) of rodents have multiple, regularly arranged firing fields, place cells in the cornu ammonis (CA) regions of the hippocampus mostly have single spatial firing fields. Since there are extensive projections from MEC to the CA regions, many models have suggested that a feedforward network can transform grid cell firing into robust place cell firing. However, these models generate place fields that are consistently too small compared to those recorded in experiments. Here, we argue that it is implausible that grid cell activity alone can be transformed into place cells with robust place fields of realistic size in a feedforward network. We propose two solutions to this problem. Firstly, weakly spatially modulated cells, which are abundant throughout EC, provide input to downstream place cells along with grid cells. This simple model reproduces many place cell characteristics as well as results from lesion studies. Secondly, the recurrent connections between place cells in the CA3 network generate robust and realistic place fields. Both mechanisms could work in parallel in the hippocampal formation and this redundancy might account for the robustness of place cell responses to a range of disruptions of the hippocampal circuitry. PMID:28750005

  10. [Analysis on difference of richness of traditional Chinese medicine resources in Chongqing based on grid technology].

    PubMed

    Zhang, Xiao-Bo; Qu, Xian-You; Li, Meng; Wang, Hui; Jing, Zhi-Xian; Liu, Xiang; Zhang, Zhi-Wei; Guo, Lan-Ping; Huang, Lu-Qi

    2017-11-01

    After the end of the national and local medicine resources census work, a large number of Chinese medicine resources and distribution of data will be summarized. The species richness between the regions is a valid indicator for objective reflection of inter-regional resources of Chinese medicine. Due to the large difference in the size of the county area, the assessment of the intercropping of the resources of the traditional Chinese medicine by the county as a statistical unit will lead to the deviation of the regional abundance statistics. Based on the rule grid or grid statistical methods, the size of the statistical unit due to different can be reduced, the differences in the richness of traditional Chinese medicine resources are caused. Taking Chongqing as an example, based on the existing survey data, the difference of richness of traditional Chinese medicine resources under different grid scale were compared and analyzed. The results showed that the 30 km grid could be selected and the richness of Chinese medicine resources in Chongqing could reflect the objective situation of intercropping resources richness in traditional Chinese medicine better. Copyright© by the Chinese Pharmaceutical Association.

  11. An electrical betweenness approach for vulnerability assessment of power grids considering the capacity of generators and load

    NASA Astrophysics Data System (ADS)

    Wang, Kai; Zhang, Bu-han; Zhang, Zhe; Yin, Xiang-gen; Wang, Bo

    2011-11-01

    Most existing research on the vulnerability of power grids based on complex networks ignores the electrical characteristics and the capacity of generators and load. In this paper, the electrical betweenness is defined by considering the maximal demand of load and the capacity of generators in power grids. The loss of load, which reflects the ability of power grids to provide sufficient power to customers, is introduced to measure the vulnerability together with the size of the largest cluster. The simulation results of the IEEE-118 bus system and the Central China Power Grid show that the cumulative distributions of node electrical betweenness follow a power-law and that the nodes with high electrical betweenness play critical roles in both topological structure and power transmission of power grids. The results prove that the model proposed in this paper is effective for analyzing the vulnerability of power grids.

  12. A method for grounding grid corrosion rate prediction

    NASA Astrophysics Data System (ADS)

    Han, Juan; Du, Jingyi

    2017-06-01

    Involved in a variety of factors, prediction of grounding grid corrosion complex, and uncertainty in the acquisition process, we propose a combination of EAHP (extended AHP) and fuzzy nearness degree of effective grounding grid corrosion rate prediction model. EAHP is used to establish judgment matrix and calculate the weight of each factors corrosion of grounding grid; different sample classification properties have different corrosion rate of contribution, and combining the principle of close to predict corrosion rate.The application result shows, the model can better capture data variation, thus to improve the validity of the model to get higher prediction precision.

  13. Measurement of beam divergence of 30-centimeter dished grids

    NASA Technical Reports Server (NTRS)

    Danilowicz, R. L.; Rawlin, V. K.; Banks, B. A.; Wintucky, E. G.

    1973-01-01

    The beam divergence of a 30-centimeter diameter thruster with dished grids was calculated from current densities measured with a probe rake containing seventeen planar molybdenum probes. The measured data were analyzed as a function of a number of parameters. The most sensitive parameters were the amount of compensation of the accelerator grid and the ratio of net to total accelerating voltage. The thrust losses were reduced by over 5 percent with the use of compensated grids alone, and by variation of other parameters the overall thrust losses due to beam divergence were reduced to less than 2 percent.

  14. Measurement of beam divergence of 30-centimeter dished grids

    NASA Technical Reports Server (NTRS)

    Danilowicz, R. L.; Rawlin, V. K.; Banks, B. A.; Wintucky, E. G.

    1973-01-01

    The beam divergence of a 30-centimeter diameter thrustor with dished grids was calculated from current densities measured with a probe rake containing seventeen planar molybdenum probes. The measured data were analyzed as a function of a number of parameters. The most sensitive parameters were the amount of compensation of the accelerator grid and the ratio of net to total accelerating voltage. The thrust losses were reduced by over 5 percent with the use of compensated grids alone, and by variation of other parameters the overall thrust losses due to beam divergence were reduced to less than 2 percent.

  15. Closing the scale gap between land surface parameterizations and GCMs with a new scheme, SiB3-Bins: SOIL MOISTURE SCALE GAP

    DOE PAGES

    Baker, I. T.; Sellers, P. J.; Denning, A. S.; ...

    2017-03-01

    The interaction of land with the atmosphere is sensitive to soil moisture (W). Evapotranspiration (ET) reacts to soil moisture in a nonlinear way, f(W), as soils dry from saturation to wilt point. This nonlinear behavior and the fact that soil moisture varies on scales as small as 1–10 m in nature, while numerical general circulation models (GCMs) have grid cell sizes on the order of 1 to 100s of kilometers, makes the calculation of grid cell-average ET problematic. It is impractical to simulate the land in GCMs on the small scales seen in nature, so techniques have been developed tomore » represent subgrid scale heterogeneity, including: (1) statistical-dynamical representations of grid subelements of varying wetness, (2) relaxation of f(W), (3) moderating f(W) with approximations of catchment hydrology, (4) “tiling” the landscape into vegetation types, and (5) hyperresolution. Here we present an alternative method for representing subgrid variability in W, one proven in a conceptual framework where landscape-scale W is represented as a series of “Bins” of increasing wetness from dry to saturated. The grid cell-level f(W) is defined by the integral of the fractional area of the wetness bins and the value of f(W) associated with each. This approach accounts for the spatiotemporal dynamics of W. We implemented this approach in the SiB3 land surface parameterization and then evaluated its performance against a control, which assumes a horizontally uniform field of W. We demonstrate that the Bins method, with a physical basis, attenuates unrealistic jumps in model state and ET seen in the control runs.« less

  16. Closing the scale gap between land surface parameterizations and GCMs with a new scheme, SiB3-Bins: SOIL MOISTURE SCALE GAP

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Baker, I. T.; Sellers, P. J.; Denning, A. S.

    The interaction of land with the atmosphere is sensitive to soil moisture (W). Evapotranspiration (ET) reacts to soil moisture in a nonlinear way, f(W), as soils dry from saturation to wilt point. This nonlinear behavior and the fact that soil moisture varies on scales as small as 1–10 m in nature, while numerical general circulation models (GCMs) have grid cell sizes on the order of 1 to 100s of kilometers, makes the calculation of grid cell-average ET problematic. It is impractical to simulate the land in GCMs on the small scales seen in nature, so techniques have been developed tomore » represent subgrid scale heterogeneity, including: (1) statistical-dynamical representations of grid subelements of varying wetness, (2) relaxation of f(W), (3) moderating f(W) with approximations of catchment hydrology, (4) “tiling” the landscape into vegetation types, and (5) hyperresolution. Here we present an alternative method for representing subgrid variability in W, one proven in a conceptual framework where landscape-scale W is represented as a series of “Bins” of increasing wetness from dry to saturated. The grid cell-level f(W) is defined by the integral of the fractional area of the wetness bins and the value of f(W) associated with each. This approach accounts for the spatiotemporal dynamics of W. We implemented this approach in the SiB3 land surface parameterization and then evaluated its performance against a control, which assumes a horizontally uniform field of W. We demonstrate that the Bins method, with a physical basis, attenuates unrealistic jumps in model state and ET seen in the control runs.« less

  17. Evaluation of a risk-based environmental hot spot delineation algorithm.

    PubMed

    Sinha, Parikhit; Lambert, Michael B; Schew, William A

    2007-10-22

    Following remedial investigations of hazardous waste sites, remedial strategies may be developed that target the removal of "hot spots," localized areas of elevated contamination. For a given exposure area, a hot spot may be defined as a sub-area that causes risks for the whole exposure area to be unacceptable. The converse of this statement may also apply: when a hot spot is removed from within an exposure area, risks for the exposure area may drop below unacceptable thresholds. The latter is the motivation for a risk-based approach to hot spot delineation, which was evaluated using Monte Carlo simulation. Random samples taken from a virtual site ("true site") were used to create an interpolated site. The latter was gridded and concentrations from the center of each grid box were used to calculate 95% upper confidence limits on the mean site contaminant concentration and corresponding hazard quotients for a potential receptor. Grid cells with the highest concentrations were removed and hazard quotients were recalculated until the site hazard quotient dropped below the threshold of 1. The grid cells removed in this way define the spatial extent of the hot spot. For each of the 100,000 Monte Carlo iterations, the delineated hot spot was compared to the hot spot in the "true site." On average, the algorithm was able to delineate hot spots that were collocated with and equal to or greater in size than the "true hot spot." When delineated hot spots were mapped onto the "true site," setting contaminant concentrations in the mapped area to zero, the hazard quotients for these "remediated true sites" were on average within 5% of the acceptable threshold of 1.

  18. Introducing GFWED: The Global Fire Weather Database

    NASA Technical Reports Server (NTRS)

    Field, R. D.; Spessa, A. C.; Aziz, N. A.; Camia, A.; Cantin, A.; Carr, R.; de Groot, W. J.; Dowdy, A. J.; Flannigan, M. D.; Manomaiphiboon, K.; hide

    2015-01-01

    The Canadian Forest Fire Weather Index (FWI) System is the mostly widely used fire danger rating system in the world. We have developed a global database of daily FWI System calculations, beginning in 1980, called the Global Fire WEather Database (GFWED) gridded to a spatial resolution of 0.5 latitude by 2-3 longitude. Input weather data were obtained from the NASA Modern Era Retrospective-Analysis for Research and Applications (MERRA), and two different estimates of daily precipitation from rain gauges over land. FWI System Drought Code calculations from the gridded data sets were compared to calculations from individual weather station data for a representative set of 48 stations in North, Central and South America, Europe, Russia,Southeast Asia and Australia. Agreement between gridded calculations and the station-based calculations tended to be most different at low latitudes for strictly MERRA based calculations. Strong biases could be seen in either direction: MERRA DC over the Mato Grosso in Brazil reached unrealistically high values exceeding DCD1500 during the dry season but was too low over Southeast Asia during the dry season. These biases are consistent with those previously identified in MERRAs precipitation, and they reinforce the need to consider alternative sources of precipitation data. GFWED can be used for analyzing historical relationships between fire weather and fire activity at continental and global scales, in identifying large-scale atmosphereocean controls on fire weather, and calibration of FWI-based fire prediction models.

  19. Large-Eddy Simulation of Turbulent Wall-Pressure Fluctuations

    NASA Technical Reports Server (NTRS)

    Singer, Bart A.

    1996-01-01

    Large-eddy simulations of a turbulent boundary layer with Reynolds number based on displacement thickness equal to 3500 were performed with two grid resolutions. The computations were continued for sufficient time to obtain frequency spectra with resolved frequencies that correspond to the most important structural frequencies on an aircraft fuselage. The turbulent stresses were adequately resolved with both resolutions. Detailed quantitative analysis of a variety of statistical quantities associated with the wall-pressure fluctuations revealed similar behavior for both simulations. The primary differences were associated with the lack of resolution of the high-frequency data in the coarse-grid calculation and the increased jitter (due to the lack of multiple realizations for averaging purposes) in the fine-grid calculation. A new curve fit was introduced to represent the spanwise coherence of the cross-spectral density.

  20. Solution of free-boundary problems using finite-element/Newton methods and locally refined grids - Application to analysis of solidification microstructure

    NASA Technical Reports Server (NTRS)

    Tsiveriotis, K.; Brown, R. A.

    1993-01-01

    A new method is presented for the solution of free-boundary problems using Lagrangian finite element approximations defined on locally refined grids. The formulation allows for direct transition from coarse to fine grids without introducing non-conforming basis functions. The calculation of elemental stiffness matrices and residual vectors are unaffected by changes in the refinement level, which are accounted for in the loading of elemental data to the global stiffness matrix and residual vector. This technique for local mesh refinement is combined with recently developed mapping methods and Newton's method to form an efficient algorithm for the solution of free-boundary problems, as demonstrated here by sample calculations of cellular interfacial microstructure during directional solidification of a binary alloy.

  1. Solution adaptive grids applied to low Reynolds number flow

    NASA Astrophysics Data System (ADS)

    de With, G.; Holdø, A. E.; Huld, T. A.

    2003-08-01

    A numerical study has been undertaken to investigate the use of a solution adaptive grid for flow around a cylinder in the laminar flow regime. The main purpose of this work is twofold. The first aim is to investigate the suitability of a grid adaptation algorithm and the reduction in mesh size that can be obtained. Secondly, the uniform asymmetric flow structures are ideal to validate the mesh structures due to mesh refinement and consequently the selected refinement criteria. The refinement variable used in this work is a product of the rate of strain and the mesh cell size, and contains two variables Cm and Cstr which determine the order of each term. By altering the order of either one of these terms the refinement behaviour can be modified.

  2. Method and apparatus for jetting, manufacturing and attaching uniform solder balls

    DOEpatents

    Yost, F.G.; Frear, D.R.; Schmale, D.T.

    1999-01-05

    An apparatus and process are disclosed for jetting molten solder in the form of balls directly onto all the metallized interconnects lands for a ball grid array package in one step with no solder paste required. Molten solder is jetted out of a grid of holes using a piston attached to a piezoelectric crystal. When voltage is applied to the crystal it expands forcing the piston to extrude a desired volume of solder through holes in the aperture plate. When the voltage is decreased the piston reverses motion creating an instability in the molten solder at the aperture plate surface and thereby forming spherical solder balls that fall onto a metallized substrate. The molten solder balls land on the substrate and form a metallurgical bond with the metallized lands. The size of the solder balls is determined by a combination of the size of the holes in the aperture plate, the duration of the piston pulse, and the displacement of the piston. The layout of the balls is dictated by the location of the hooks in the grid. Changes in ball size and layout can be easily accomplished by changing the grid plate. This invention also allows simple preparation of uniform balls for subsequent supply to BGA users. 7 figs.

  3. Method and apparatus for jetting, manufacturing and attaching uniform solder balls

    DOEpatents

    Yost, Frederick G.; Frear, Darrel R.; Schmale, David T.

    1999-01-01

    An apparatus and process for jetting molten solder in the form of balls directly onto all the metallized interconnects lands for a ball grid array package in one step with no solder paste required. Molten solder is jetted out of a grid of holes using a piston attached to a piezoelectric crystal. When voltage is applied to the crystal it expands forcing the piston to extrude a desired volume of solder through holes in the aperture plate. When the voltage is decreased the piston reverses motion creating an instability in the molten solder at the aperture plate surface and thereby forming spherical solder balls that fall onto a metallized substrate. The molten solder balls land on the substrate and form a metallurgical bond with the metallized lands. The size of the solder balls is determined by a combination of the size of the holes in the aperture plate, the duration of the piston pulse, and the displacement of the piston. The layout of the balls is dictated by the location of the hooks in the grid. Changes in ball size and layout can be easily accomplished by changing the grid plate. This invention also allows simple preparation of uniform balls for subsequent supply to BGA users.

  4. Orientation domains: A mobile grid clustering algorithm with spherical corrections

    NASA Astrophysics Data System (ADS)

    Mencos, Joana; Gratacós, Oscar; Farré, Mercè; Escalante, Joan; Arbués, Pau; Muñoz, Josep Anton

    2012-12-01

    An algorithm has been designed and tested which was devised as a tool assisting the analysis of geological structures solely from orientation data. More specifically, the algorithm was intended for the analysis of geological structures that can be approached as planar and piecewise features, like many folded strata. Input orientation data is expressed as pairs of angles (azimuth and dip). The algorithm starts by considering the data in Cartesian coordinates. This is followed by a search for an initial clustering solution, which is achieved by comparing the results output from the systematic shift of a regular rigid grid over the data. This initial solution is optimal (achieves minimum square error) once the grid size and the shift increment are fixed. Finally, the algorithm corrects for the variable spread that is generally expected from the data type using a reshaped non-rigid grid. The algorithm is size-oriented, which implies the application of conditions over cluster size through all the process in contrast to density-oriented algorithms, also widely used when dealing with spatial data. Results are derived in few seconds and, when tested over synthetic examples, they were found to be consistent and reliable. This makes the algorithm a valuable alternative to the time-consuming traditional approaches available to geologists.

  5. Moving Computational Domain Method and Its Application to Flow Around a High-Speed Car Passing Through a Hairpin Curve

    NASA Astrophysics Data System (ADS)

    Watanabe, Koji; Matsuno, Kenichi

    This paper presents a new method for simulating flows driven by a body traveling with neither restriction on motion nor a limit of a region size. In the present method named 'Moving Computational Domain Method', the whole of the computational domain including bodies inside moves in the physical space without the limit of region size. Since the whole of the grid of the computational domain moves according to the movement of the body, a flow solver of the method has to be constructed on the moving grid system and it is important for the flow solver to satisfy physical and geometric conservation laws simultaneously on moving grid. For this issue, the Moving-Grid Finite-Volume Method is employed as the flow solver. The present Moving Computational Domain Method makes it possible to simulate flow driven by any kind of motion of the body in any size of the region with satisfying physical and geometric conservation laws simultaneously. In this paper, the method is applied to the flow around a high-speed car passing through a hairpin curve. The distinctive flow field driven by the car at the hairpin curve has been demonstrated in detail. The results show the promising feature of the method.

  6. Conceptual Design of the Everglades Depth Estimation Network (EDEN) Grid

    USGS Publications Warehouse

    Jones, John W.; Price, Susan D.

    2007-01-01

    INTRODUCTION The Everglades Depth Estimation Network (EDEN) offers a consistent and documented dataset that can be used to guide large-scale field operations, to integrate hydrologic and ecological responses, and to support biological and ecological assessments that measure ecosystem responses to the Comprehensive Everglades Restoration Plan (Telis, 2006). Ground elevation data for the greater Everglades and the digital ground elevation models derived from them form the foundation for all EDEN water depth and associated ecologic/hydrologic modeling (Jones, 2004, Jones and Price, 2007). To use EDEN water depth and duration information most effectively, it is important to be able to view and manipulate information on elevation data quality and other land cover and habitat characteristics across the Everglades region. These requirements led to the development of the geographic data layer described in this techniques and methods report. Relying on extensive experience in GIS data development, distribution, and analysis, a great deal of forethought went into the design of the geographic data layer used to index elevation and other surface characteristics for the Greater Everglades region. To allow for simplicity of design and use, the EDEN area was broken into a large number of equal-sized rectangles ('Cells') that in total are referred to here as the 'grid'. Some characteristics of this grid, such as the size of its cells, its origin, the area of Florida it is designed to represent, and individual grid cell identifiers, could not be changed once the grid database was developed. Therefore, these characteristics were selected to design as robust a grid as possible and to ensure the grid's long-term utility. It is desirable to include all pertinent information known about elevation and elevation data collection as grid attributes. Also, it is very important to allow for efficient grid post-processing, sub-setting, analysis, and distribution. This document details the conceptual design of the EDEN grid spatial parameters and cell attribute-table content.

  7. An interactive grid generation procedure for axial and radial flow turbomachinery

    NASA Technical Reports Server (NTRS)

    Beach, Timothy A.

    1989-01-01

    A combination algebraic/elliptic technique is presented for the generation of three dimensional grids about turbo-machinery blade rows for both axial and radial flow machinery. The technique is built around use of an advanced engineering workstation to construct several two dimensional grids interactively on predetermined blade-to-blade surfaces. A three dimensional grid is generated by interpolating these surface grids onto an axisymmetric grid. On each blade-to-blade surface, a grid is created using algebraic techniques near the blade to control orthogonality within the boundary layer region and elliptic techniques in the mid-passage to achieve smoothness. The interactive definition of bezier curves as internal boundaries is the key to simple construction. This procedure lends itself well to zonal grid construction, an important example being the tip clearance region. Calculations done to date include a space shuttle main engine turbopump blade, a radial inflow turbine blade, and the first stator of the United Technologies Research Center large scale rotating rig. A finite Navier-Stokes solver was used in each case.

  8. The three-dimensional Multi-Block Advanced Grid Generation System (3DMAGGS)

    NASA Technical Reports Server (NTRS)

    Alter, Stephen J.; Weilmuenster, Kenneth J.

    1993-01-01

    As the size and complexity of three dimensional volume grids increases, there is a growing need for fast and efficient 3D volumetric elliptic grid solvers. Present day solvers are limited by computational speed and do not have all the capabilities such as interior volume grid clustering control, viscous grid clustering at the wall of a configuration, truncation error limiters, and convergence optimization residing in one code. A new volume grid generator, 3DMAGGS (Three-Dimensional Multi-Block Advanced Grid Generation System), which is based on the 3DGRAPE code, has evolved to meet these needs. This is a manual for the usage of 3DMAGGS and contains five sections, including the motivations and usage, a GRIDGEN interface, a grid quality analysis tool, a sample case for verifying correct operation of the code, and a comparison to both 3DGRAPE and GRIDGEN3D. Since it was derived from 3DGRAPE, this technical memorandum should be used in conjunction with the 3DGRAPE manual (NASA TM-102224).

  9. Shearing-induced asymmetry in entorhinal grid cells.

    PubMed

    Stensola, Tor; Stensola, Hanne; Moser, May-Britt; Moser, Edvard I

    2015-02-12

    Grid cells are neurons with periodic spatial receptive fields (grids) that tile two-dimensional space in a hexagonal pattern. To provide useful information about location, grids must be stably anchored to an external reference frame. The mechanisms underlying this anchoring process have remained elusive. Here we show in differently sized familiar square enclosures that the axes of the grids are offset from the walls by an angle that minimizes symmetry with the borders of the environment. This rotational offset is invariably accompanied by an elliptic distortion of the grid pattern. Reversing the ellipticity analytically by a shearing transformation removes the angular offset. This, together with the near-absence of rotation in novel environments, suggests that the rotation emerges through non-coaxial strain as a function of experience. The systematic relationship between rotation and distortion of the grid pattern points to shear forces arising from anchoring to specific geometric reference points as key elements of the mechanism for alignment of grid patterns to the external world.

  10. Summary of the Fourth AIAA CFD Drag Prediction Workshop

    NASA Technical Reports Server (NTRS)

    Vassberg, John C.; Tinoco, Edward N.; Mani, Mori; Rider, Ben; Zickuhr, Tom; Levy, David W.; Brodersen, Olaf P.; Eisfeld, Bernhard; Crippa, Simone; Wahls, Richard A.; hide

    2010-01-01

    Results from the Fourth AIAA Drag Prediction Workshop (DPW-IV) are summarized. The workshop focused on the prediction of both absolute and differential drag levels for wing-body and wing-body-horizontal-tail configurations that are representative of transonic transport air- craft. Numerical calculations are performed using industry-relevant test cases that include lift- specific flight conditions, trimmed drag polars, downwash variations, dragrises and Reynolds- number effects. Drag, lift and pitching moment predictions from numerous Reynolds-Averaged Navier-Stokes computational fluid dynamics methods are presented. Solutions are performed on structured, unstructured and hybrid grid systems. The structured-grid sets include point- matched multi-block meshes and over-set grid systems. The unstructured and hybrid grid sets are comprised of tetrahedral, pyramid, prismatic, and hexahedral elements. Effort is made to provide a high-quality and parametrically consistent family of grids for each grid type about each configuration under study. The wing-body-horizontal families are comprised of a coarse, medium and fine grid; an optional extra-fine grid augments several of the grid families. These mesh sequences are utilized to determine asymptotic grid-convergence characteristics of the solution sets, and to estimate grid-converged absolute drag levels of the wing-body-horizontal configuration using Richardson extrapolation.

  11. Photochemical grid model performance with varying horizontal grid resolution and sub-grid plume treatment for the Martins Creek near-field SO2 study

    NASA Astrophysics Data System (ADS)

    Baker, Kirk R.; Hawkins, Andy; Kelly, James T.

    2014-12-01

    Near source modeling is needed to assess primary and secondary pollutant impacts from single sources and single source complexes. Source-receptor relationships need to be resolved from tens of meters to tens of kilometers. Dispersion models are typically applied for near-source primary pollutant impacts but lack complex photochemistry. Photochemical models provide a realistic chemical environment but are typically applied using grid cell sizes that may be larger than the distance between sources and receptors. It is important to understand the impacts of grid resolution and sub-grid plume treatments on photochemical modeling of near-source primary pollution gradients. Here, the CAMx photochemical grid model is applied using multiple grid resolutions and sub-grid plume treatment for SO2 and compared with a receptor mesonet largely impacted by nearby sources approximately 3-17 km away in a complex terrain environment. Measurements are compared with model estimates of SO2 at 4- and 1-km resolution, both with and without sub-grid plume treatment and inclusion of finer two-way grid nests. Annual average estimated SO2 mixing ratios are highest nearest the sources and decrease as distance from the sources increase. In general, CAMx estimates of SO2 do not compare well with the near-source observations when paired in space and time. Given the proximity of these sources and receptors, accuracy in wind vector estimation is critical for applications that pair pollutant predictions and observations in time and space. In typical permit applications, predictions and observations are not paired in time and space and the entire distributions of each are directly compared. Using this approach, model estimates using 1-km grid resolution best match the distribution of observations and are most comparable to similar studies that used dispersion and Lagrangian modeling systems. Model-estimated SO2 increases as grid cell size decreases from 4 km to 250 m. However, it is notable that the 1-km model estimates using 1-km meteorological model input are higher than the 1-km model simulation that used interpolated 4-km meteorology. The inclusion of sub-grid plume treatment did not improve model skill in predicting SO2 in time and space and generally acts to keep emitted mass aloft.

  12. Introducing MCgrid 2.0: Projecting cross section calculations on grids

    NASA Astrophysics Data System (ADS)

    Bothmann, Enrico; Hartland, Nathan; Schumann, Steffen

    2015-11-01

    MCgrid is a software package that provides access to interpolation tools for Monte Carlo event generator codes, allowing for the fast and flexible variation of scales, coupling parameters and PDFs in cutting edge leading- and next-to-leading-order QCD calculations. We present the upgrade to version 2.0 which has a broader scope of interfaced interpolation tools, now providing access to fastNLO, and features an approximated treatment for the projection of MC@NLO-type calculations onto interpolation grids. MCgrid 2.0 also now supports the extended information provided through the HepMC event record used in the recent SHERPA version 2.2.0. The additional information provided therein allows for the support of multi-jet merged QCD calculations in a future update of MCgrid.

  13. A Calculation Method for Convective Heat and Mass Transfer in Multiply-Slotted Film-Cooling Applications.

    DTIC Science & Technology

    1980-01-01

    Transport of Heat ..... .......... 8 3. THE SOLUTION PROCEDURE ..... .. ................. 8 3.1 The Finite-Difference Grid Network ... .......... 8 3.2...The Finite-Difference Grid Network. Figure 4: The Iterative Solution Procedure used at each Streamwise Station. Figure 5: Velocity Profiles in the...the finite-difference grid in the y-direction. I is the mixing length. L is the distance in the x-direction from the injection slot entrance to the

  14. Arctic Acoustic Workshop Proceedings, 14-15 February 1989.

    DTIC Science & Technology

    1989-06-01

    measurements. The measurements reported by Levine et al. (1987) were taken from current and temperature sensors moored in two triangular grids . The internal...requires a resampling of the data series on a uniform depth-time grid . Statistics calculated from the resampled series will be used to test numerical...from an isolated keel. Figure 2: 2-D Modeling Geometry - The model is based on a 2-D Cartesian grid with an axis of symmetry on the left. A pulsed

  15. A propagation method with adaptive mesh grid based on wave characteristics for wave optics simulation

    NASA Astrophysics Data System (ADS)

    Tang, Qiuyan; Wang, Jing; Lv, Pin; Sun, Quan

    2015-10-01

    Propagation simulation method and choosing mesh grid are both very important to get the correct propagation results in wave optics simulation. A new angular spectrum propagation method with alterable mesh grid based on the traditional angular spectrum method and the direct FFT method is introduced. With this method, the sampling space after propagation is not limited to propagation methods no more, but freely alterable. However, choosing mesh grid on target board influences the validity of simulation results directly. So an adaptive mesh choosing method based on wave characteristics is proposed with the introduced propagation method. We can calculate appropriate mesh grids on target board to get satisfying results. And for complex initial wave field or propagation through inhomogeneous media, we can also calculate and set the mesh grid rationally according to above method. Finally, though comparing with theoretical results, it's shown that the simulation result with the proposed method coinciding with theory. And by comparing with the traditional angular spectrum method and the direct FFT method, it's known that the proposed method is able to adapt to a wider range of Fresnel number conditions. That is to say, the method can simulate propagation results efficiently and correctly with propagation distance of almost zero to infinity. So it can provide better support for more wave propagation applications such as atmospheric optics, laser propagation and so on.

  16. Tensor numerical methods in quantum chemistry: from Hartree-Fock to excitation energies.

    PubMed

    Khoromskaia, Venera; Khoromskij, Boris N

    2015-12-21

    We resume the recent successes of the grid-based tensor numerical methods and discuss their prospects in real-space electronic structure calculations. These methods, based on the low-rank representation of the multidimensional functions and integral operators, first appeared as an accurate tensor calculus for the 3D Hartree potential using 1D complexity operations, and have evolved to entirely grid-based tensor-structured 3D Hartree-Fock eigenvalue solver. It benefits from tensor calculation of the core Hamiltonian and two-electron integrals (TEI) in O(n log n) complexity using the rank-structured approximation of basis functions, electron densities and convolution integral operators all represented on 3D n × n × n Cartesian grids. The algorithm for calculating TEI tensor in a form of the Cholesky decomposition is based on multiple factorizations using algebraic 1D "density fitting" scheme, which yield an almost irreducible number of product basis functions involved in the 3D convolution integrals, depending on a threshold ε > 0. The basis functions are not restricted to separable Gaussians, since the analytical integration is substituted by high-precision tensor-structured numerical quadratures. The tensor approaches to post-Hartree-Fock calculations for the MP2 energy correction and for the Bethe-Salpeter excitation energies, based on using low-rank factorizations and the reduced basis method, were recently introduced. Another direction is towards the tensor-based Hartree-Fock numerical scheme for finite lattices, where one of the numerical challenges is the summation of electrostatic potentials of a large number of nuclei. The 3D grid-based tensor method for calculation of a potential sum on a L × L × L lattice manifests the linear in L computational work, O(L), instead of the usual O(L(3) log L) scaling by the Ewald-type approaches.

  17. Visual Analytics for Power Grid Contingency Analysis

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wong, Pak C.; Huang, Zhenyu; Chen, Yousu

    2014-01-20

    Contingency analysis is the process of employing different measures to model scenarios, analyze them, and then derive the best response to remove the threats. This application paper focuses on a class of contingency analysis problems found in the power grid management system. A power grid is a geographically distributed interconnected transmission network that transmits and delivers electricity from generators to end users. The power grid contingency analysis problem is increasingly important because of both the growing size of the underlying raw data that need to be analyzed and the urgency to deliver working solutions in an aggressive timeframe. Failure tomore » do so may bring significant financial, economic, and security impacts to all parties involved and the society at large. The paper presents a scalable visual analytics pipeline that transforms about 100 million contingency scenarios to a manageable size and form for grid operators to examine different scenarios and come up with preventive or mitigation strategies to address the problems in a predictive and timely manner. Great attention is given to the computational scalability, information scalability, visual scalability, and display scalability issues surrounding the data analytics pipeline. Most of the large-scale computation requirements of our work are conducted on a Cray XMT multi-threaded parallel computer. The paper demonstrates a number of examples using western North American power grid models and data.« less

  18. Changing from computing grid to knowledge grid in life-science grid.

    PubMed

    Talukdar, Veera; Konar, Amit; Datta, Ayan; Choudhury, Anamika Roy

    2009-09-01

    Grid computing has a great potential to become a standard cyber infrastructure for life sciences that often require high-performance computing and large data handling, which exceeds the computing capacity of a single institution. Grid computer applies the resources of many computers in a network to a single problem at the same time. It is useful to scientific problems that require a great number of computer processing cycles or access to a large amount of data.As biologists,we are constantly discovering millions of genes and genome features, which are assembled in a library and distributed on computers around the world.This means that new, innovative methods must be developed that exploit the re-sources available for extensive calculations - for example grid computing.This survey reviews the latest grid technologies from the viewpoints of computing grid, data grid and knowledge grid. Computing grid technologies have been matured enough to solve high-throughput real-world life scientific problems. Data grid technologies are strong candidates for realizing a "resourceome" for bioinformatics. Knowledge grids should be designed not only from sharing explicit knowledge on computers but also from community formulation for sharing tacit knowledge among a community. By extending the concept of grid from computing grid to knowledge grid, it is possible to make use of a grid as not only sharable computing resources, but also as time and place in which people work together, create knowledge, and share knowledge and experiences in a community.

  19. Finding the proper methodology for geodiversity assessment: a recent approach in Brazil and Portugal

    NASA Astrophysics Data System (ADS)

    Pereira, D.; Santos, L.; Silva, J.; Pereira, P.; Brilha, J.; França, J.; Rodrigues, C.

    2012-04-01

    Quantification of geodiversity is a quite new topic. A first set of assessment methodologies was developed during the last years, although with no fully satisfactory results. This is mainly because the whole concept of geodiversity does not tend to be considered, but also because the results are difficult to apply practically. Several major key-points remain unsolved, including the criteria to be used, the scale-factor to be dealt with, the influence of the size of the area under analysis in the type of criteria and indicators, and the graphic presentation of the results. A methodology for the quantitative assessment of geodiversity was defined and tested at various scales. It was applied to the Xingu River Basin, Amazon, Brazil (about 510,000 km2), Paraná state, Brazil (about 200,000 km2), and Portugal mainland (about 89,000 km2). This method is intended to assess all geodiversity components and to avoid overrating any particular component, such as lithology or relief, a common weakness of other methods. The method is based on the overlay of a grid over different maps at scales that range according to the areas under analysis, with the final Geodiversity Index being the sum of five partial indexes calculated on the grid. The partial indexes represent the main components of geodiversity, namely geology (stratigraphy and lithology), geomorphology, palaeontology and soils. Another partial index covers singular occurrences of geodiversity, such precious stones and metals, energy and industrial minerals, mineral waters and springs. Partial indexes were calculated using GIS software by counting all the occurrences present in the selected maps for each grid square. The Geodiversity Index can take the form of a GIS automatically generated isoline map, allowing an easy interpretation by those without or with little geological background. The map can be used as a tool in land-use planning, particularly in identifying priority areas for conservation, management and use of natural resources.

  20. Modelling noise propagation using Grid Resources. Progress within GDI-Grid

    NASA Astrophysics Data System (ADS)

    Kiehle, Christian; Mayer, Christian; Padberg, Alexander; Stapelfeld, Hartmut

    2010-05-01

    Modelling noise propagation using Grid Resources. Progress within GDI-Grid. GDI-Grid (english: SDI-Grid) is a research project funded by the German Ministry for Science and Education (BMBF). It aims at bridging the gaps between OGC Web Services (OWS) and Grid infrastructures and identifying the potential of utilizing the superior storage capacities and computational power of grid infrastructures for geospatial applications while keeping the well-known service interfaces specified by the OGC. The project considers all major OGC webservice interfaces for Web Mapping (WMS), Feature access (Web Feature Service), Coverage access (Web Coverage Service) and processing (Web Processing Service). The major challenge within GDI-Grid is the harmonization of diverging standards as defined by standardization bodies for Grid computing and spatial information exchange. The project started in 2007 and will continue until June 2010. The concept for the gridification of OWS developed by lat/lon GmbH and the Department of Geography of the University of Bonn is applied to three real-world scenarios in order to check its practicability: a flood simulation, a scenario for emergency routing and a noise propagation simulation. The latter scenario is addressed by the Stapelfeldt Ingenieurgesellschaft mbH located in Dortmund adapting their LimA software to utilize grid resources. Noise mapping of e.g. traffic noise in urban agglomerates and along major trunk roads is a reoccurring demand of the EU Noise Directive. Input data requires road net and traffic, terrain, buildings and noise protection screens as well as population distribution. Noise impact levels are generally calculated in 10 m grid and along relevant building facades. For each receiver position sources within a typical range of 2000 m are split down into small segments, depending on local geometry. For each of the segments propagation analysis includes diffraction effects caused by all obstacles on the path of sound propagation. This immense intensive calculation needs to be performed for a major part of European landscape. A LINUX version of the commercial LimA software for noise mapping analysis has been implemented on a test cluster within the German D-GRID computer network. Results and performance indicators will be presented. The presentation is an extension to last-years presentation "Spatial Data Infrastructures and Grid Computing: the GDI-Grid project" that described the gridification concept developed in the GDI-Grid project and provided an overview of the conceptual gaps between Grid Computing and Spatial Data Infrastructures. Results from the GDI-Grid project are incorporated in the OGC-OGF (Open Grid Forum) collaboration efforts as well as the OGC WPS 2.0 standards working group developing the next major version of the WPS specification.

  1. Panoramic Night Vision Goggle Testing For Diagnosis and Repair

    DTIC Science & Technology

    2000-01-01

    Visual Acuity Visual Acuity [ Marasco & Task, 1999] measures how well a human observer can see high contrast targets at specified light levels through...grid through the PNVG in-board and out-board channels simultaneously and comparing the defects to the size of grid features ( Marasco & Task, 1999). The

  2. SU-D-210-07: The Dependence On Acoustic Velocity of Medium On the Needle Template and Electronic Grid Alignment in Ultrasound QA for Prostate Brachytherapy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kapoor, P; Kapoor, R; Curran, B

    Purpose: To analyze the impact on acoustic velocity (AV) of two different media (water and milk) using the needle template/electronic grid alignment test. Water, easily available, makes a good material to test the alignment of the template and grid although water’s AV (1498 m/s at 25°C) is significantly different from tissue (1540 m/s). Milk, with an AV much closer (1548 m/s) to prostate tissue, may be a good substitute for water in ultrasound quality assurance testing. Methods: Tests were performed using a Hitachi ultrasound unit with a mechanical arrangement designed to position needles parallel to the transducer. In this work,more » two materials – distilled water and homogenized whole milk (AVs of 1498 and 1548 m/s at 25°C) were used in a phantom to test ultrasound needle/grid alignment. The images were obtained with both materials and analyzed for their placement accuracy. Results: The needle template/electronic grid alignment tests showed displacement errors between measured and calculated values. The measurements showed displacements of 2.3mm (water) and 0.4mm (milk), and 1.6mm (water) and 0.3mm (milk) at depths of 7cm and 5cm respectively from true needle positions. The calculated results showed a displacement of 2.36 mm (water); 0.435mm (milk), and 1.66mm (water) and 0.31mm (milk) at a depth of 7cm and 5cm respectively. The displacements in the X and Y directions were also calculated. At depths of 7cm and 5cm, the (ΔX,ΔY) displacements in water were (0.829mm, 2.21mm) and (0.273mm, 1.634mm) and for milk were (0.15mm, 0.44mm) and (0.05mm, 0.302mm) respectively. Conclusion: The measured and calculated values were in good agreement for all tests. They show that milk provides superior results when performing needle template and electronic grid alignment tests for ultrasound units used in prostate brachytherapy.« less

  3. Hole-ness of point clouds

    NASA Astrophysics Data System (ADS)

    Gronz, Oliver; Seeger, Manuel; Klaes, Björn; Casper, Markus C.; Ries, Johannes B.

    2015-04-01

    Accurate and dense 3D models of soil surfaces can be used in various ways: They can be used as initial shapes for erosion models. They can be used as benchmark shapes for erosion model outputs. They can be used to derive metrics, such as random roughness... One easy and low-cost method to produce these models is structure from motion (SfM). Using this method, two questions arise: Does the soil moisture, which changes the colour, albedo and reflectivity of the soil, influence the model quality? How can the model quality be evaluated? To answer these questions, a suitable data set has been produced: soil has been placed on a tray and areas with different roughness structures have been formed. For different moisture states - dry, medium, saturated - and two different lighting conditions - direct and indirect - sets of high-resolution images at the same camera positions have been taken. From the six image sets, 3D point clouds have been produced using VisualSfM. The visual inspection of the 3D models showed that all models have different areas, where holes of different sizes occur. But it is obviously a subjective task to determine the model's quality by visual inspection. One typical approach to evaluate model quality objectively is to estimate the point density on a regular, two-dimensional grid: the number of 3D points in each grid cell projected on a plane is calculated. This works well for surfaces that do not show vertical structures. Along vertical structures, many points will be projected on the same grid cell and thus the point density rather depends on the shape of the surface but less on the quality of the model. Another approach has been applied by using the points resulting from Poisson Surface Reconstructions. One of this algorithm's properties is the filling of holes: new points are interpolated inside the holes. Using the original 3D point cloud and the interpolated Poisson point set, two analyses have been performed: For all Poisson points, the distance to the closest original point cloud member has been calculated. For the resulting set of distances, histograms have been produced that show the distribution of point distances. As the Poisson points also make up a connected mesh, the size and distribution of single holes can also be estimated by labeling Poisson points that belong to the same hole: each hole gets a specific number. Afterwards, the area of the mesh formed by each set of Poisson hole points can be calculated. The result is a set of distinctive holes and their sizes. The two approaches showed that the hole-ness of the point cloud depends on the soil moisture respectively the reflectivity: the distance distribution of the model of the saturated soil shows the smallest number of large distances. The histogram of the medium state shows more large distances and the dry model shows the largest distances. Models resulting from indirect lighting are better than the models resulting from direct light for all moisture states.

  4. An Upwind Multigrid Algorithm for Calculating Flows on Unstructured Grids

    NASA Technical Reports Server (NTRS)

    Bonhaus, Daryl L.

    1993-01-01

    An algorithm is described that calculates inviscid, laminar, and turbulent flows on triangular meshes with an upwind discretization. A brief description of the base solver and the multigrid implementation is given, followed by results that consist mainly of convergence rates for inviscid and viscous flows over a NACA four-digit airfoil section. The results show that multigrid does accelerate convergence when the same relaxation parameters that yield good single-grid performance are used; however, larger gains in performance can be realized by doing less work in the relaxation scheme.

  5. Construction of the Fock Matrix on a Grid-Based Molecular Orbital Basis Using GPGPUs.

    PubMed

    Losilla, Sergio A; Watson, Mark A; Aspuru-Guzik, Alán; Sundholm, Dage

    2015-05-12

    We present a GPGPU implementation of the construction of the Fock matrix in the molecular orbital basis using the fully numerical, grid-based bubbles representation. For a test set of molecules containing up to 90 electrons, the total Hartree-Fock energies obtained from reference GTO-based calculations are reproduced within 10(-4) Eh to 10(-8) Eh for most of the molecules studied. Despite the very large number of arithmetic operations involved, the high performance obtained made the calculations possible on a single Nvidia Tesla K40 GPGPU card.

  6. Predicting grid-size-dependent fracture strains of DP980 with a microstructure-based post-necking model

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cheng, G.; Hu, X. H.; Choi, K. S.

    Ductile fracture is a local phenomenon, and it is well established that fracture strain levels depend on both stress triaxiality and the resolution (grid size) of strain measurements. Two-dimensional plane strain post-necking models with different representative volume element (RVE) sizes are used to predict the size-dependent fracture strain of a commercial dual-phase steel, DP980. The models are generated from the actual microstructures, and the individual phase flow properties and literature-based individual phase damage parameters for the Johnson-Cook model are used for ferrite and martensite. A monotonic relationship is predicted: the smaller the model size, the higher the fracture strain. Thus,more » a general framework is developed to quantify the size-dependent fracture strains for multiphase materials. In addition to the RVE sizes, the influences of intrinsic microstructure features, i.e., the flow curve and fracture strains of the two constituent phases, on the predicted fracture strains also are examined. Application of the derived fracture strain versus RVE size relationship is demonstrated with large clearance trimming simulations with different element sizes.« less

  7. Calculation of flow about two-dimensional bodies by means of the velocity-vorticity formulation on a staggered grid

    NASA Technical Reports Server (NTRS)

    Stremel, Paul M.

    1991-01-01

    A method for calculating the incompressible viscous flow about two-dimensional bodies, utilizing the velocity-vorticity form of the Navier-Stokes equations using a staggered-grid formulation is presented. The solution is obtained by employing an alternative-direction implicit method for the solution of the block tridiagonal matrix resulting from the finite-difference representation of the governing equations. The boundary vorticity and the conservation of mass are calculated implicitly as a part of the solution. The mass conservation is calculated to machine zero for the duration of the computation. Calculations for the flow about a circular cylinder, a 2-pct thick flat plate at 90-deg incidence, an elliptic cylinder at 45-deg incidence, and a NACA 0012, with and without a deflected flap, at - 90-deg incidence are performed and compared with the results of other numerical investigations.

  8. Grid Sensitivity Study for Slat Noise Simulations

    NASA Technical Reports Server (NTRS)

    Lockard, David P.; Choudhari, Meelan M.; Buning, Pieter G.

    2014-01-01

    The slat noise from the 30P/30N high-lift system is being investigated through computational fluid dynamics simulations in conjunction with a Ffowcs Williams-Hawkings acoustics solver. Many previous simulations have been performed for the configuration, and the case was introduced as a new category for the Second AIAA workshop on Benchmark problems for Airframe Noise Configurations (BANC-II). However, the cost of the simulations has restricted the study of grid resolution effects to a baseline grid and coarser meshes. In the present study, two different approaches are being used to investigate the effect of finer resolution of near-field unsteady structures. First, a standard grid refinement by a factor of two is used, and the calculations are performed by using the same CFL3D solver employed in the majority of the previous simulations. Second, the OVERFLOW code is applied to the baseline grid, but with a 5th-order upwind spatial discretization as compared with the second-order discretization used in the CFL3D simulations. In general, the fine grid CFL3D simulation and OVERFLOW calculation are in very good agreement and exhibit the lowest levels of both surface pressure fluctuations and radiated noise. Although the smaller scales resolved by these simulations increase the velocity fluctuation levels, they appear to mitigate the influence of the larger scales on the surface pressure. These new simulations are used to investigate the influence of the grid on unsteady high-lift simulations and to gain a better understanding of the physics responsible for the noise generation and radiation.

  9. Installed F/A-18 inlet flow calculations at 30 degrees angle-of-attack: A comparative study

    NASA Technical Reports Server (NTRS)

    Smith, C. Frederic; Podleski, Steve D.

    1994-01-01

    NASA Lewis is currently engaged in a research effort as a team member of the High Alpha Technology Program (HATP) within NASA. This program utilizes a specially equipped F/A-18, the High Alpha Research Vehicle (HARV), in an ambitious effort to improve the maneuverability of high-performance military aircraft at low subsonic speed, high angle of attack conditions. The overall objective of the Lewis effort is to develop inlet technology that will ensure efficient airflow delivery to the engine during these maneuvers. One part of the Lewis approach utilizes computational fluid dynamics codes to predict the installed performance of inlets for these highly maneuverable aircraft. Full Navier-Stokes (FNS) calculations on the installed F/A-18 inlet at 30 degrees angle of attack, 0 degrees yaw, and a freestream Mach number of 0.2 have been obtained in this study using an algebraic turbulence model with two grids (original and revised). Results obtained with the original grid were used to determine where further grid refinements and additional geometry were needed. In order to account properly for the external effects, the forebody, leading edge extension (LEX), ramp, and wing were included with inlet geometry. In the original grid, the diverter, LEX slot, and leading edge flap were not included due to insufficient geometry definition, but were included in a revised grid. In addition, a thin-layer Navier-Stokes (TLNS) code is used with the revised grid and the numerical results are compared to those obtained with the FNS code. The TLNS code was used to evaluate the effects on the solution using a code with more recent CFD developments such as upwinding with TVD schemes versus central differencing with artificial dissipation. The calculations are compared to a limited amount of available experimental data. The predicted forebody/fuselage surface static pressures compared well with data of all solutions. The predicted trajectory of the vortex generated under the LEX was different for each solution. These discrepancies are attributed to differences in the grid resolution and turbulence modeling. All solutions predict that this vortex is ingested by the inlet. The predicted inlet total pressure recoveries are lower than data and the distortions are higher than data. The results obtained with the revised grid were significantly improved from the original grid results. The original grid results indicated the ingested vortex migrated to the engine face and caused additional distortions to those already present due to secondary flow development. The revised grid results indicate that the ingested vortex is dissipated along the inlet duct inboard wall. The TLNS results indicate the flow at the engine face was much more distorted than the FNS results and is attributed to the pole boundary condition introducing numerical distortions into the flow field.

  10. An open source software for fast grid-based data-mining in spatial epidemiology (FGBASE).

    PubMed

    Baker, David M; Valleron, Alain-Jacques

    2014-10-30

    Examining whether disease cases are clustered in space is an important part of epidemiological research. Another important part of spatial epidemiology is testing whether patients suffering from a disease are more, or less, exposed to environmental factors of interest than adequately defined controls. Both approaches involve determining the number of cases and controls (or population at risk) in specific zones. For cluster searches, this often must be done for millions of different zones. Doing this by calculating distances can lead to very lengthy computations. In this work we discuss the computational advantages of geographical grid-based methods, and introduce an open source software (FGBASE) which we have created for this purpose. Geographical grids based on the Lambert Azimuthal Equal Area projection are well suited for spatial epidemiology because they preserve area: each cell of the grid has the same area. We describe how data is projected onto such a grid, as well as grid-based algorithms for spatial epidemiological data-mining. The software program (FGBASE), that we have developed, implements these grid-based methods. The grid based algorithms perform extremely fast. This is particularly the case for cluster searches. When applied to a cohort of French Type 1 Diabetes (T1D) patients, as an example, the grid based algorithms detected potential clusters in a few seconds on a modern laptop. This compares very favorably to an equivalent cluster search using distance calculations instead of a grid, which took over 4 hours on the same computer. In the case study we discovered 4 potential clusters of T1D cases near the cities of Le Havre, Dunkerque, Toulouse and Nantes. One example of environmental analysis with our software was to study whether a significant association could be found between distance to vineyards with heavy pesticide. None was found. In both examples, the software facilitates the rapid testing of hypotheses. Grid-based algorithms for mining spatial epidemiological data provide advantages in terms of computational complexity thus improving the speed of computations. We believe that these methods and this software tool (FGBASE) will lower the computational barriers to entry for those performing epidemiological research.

  11. SU-C-BRC-07: Parametrized GPU Accelerated Electron Monte Carlo Second Check

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Haywood, J

    Purpose: I am presenting a parameterized 3D GPU accelerated electron Monte Carlo second check program. Method: I wrote the 3D grid dose calculation algorithm in CUDA and utilized an NVIDIA GeForce GTX 780 Ti to run all of the calculations. The electron path beyond the distal end of the cone is governed by four parameters: the amplitude of scattering (AMP), the mean and width of a Gaussian energy distribution (E and α), and the percentage of photons. In my code, I adjusted all parameters until the calculated PDD and profile fit the measured 10×10 open beam data within 1%/1mm. Imore » then wrote a user interface for reading the DICOM treatment plan and images in Python. In order to verify the algorithm, I calculated 3D dose distributions on a variety of phantoms and geometries, and compared them with the Eclipse eMC calculations. I also calculated several patient specific dose distributions, including a nose and an ear. Finally, I compared my algorithm’s computation times to Eclipse’s. Results: The calculated MU for all of the investigated geometries agree with the TPS within the TG-114 action level of 5%. The MU for the nose was < 0.5 % different while the MU for the ear at 105 SSD was ∼2 %. Calculation times for a 12MeV 10×10 open beam ranged from 1 second for a 2.5 mm grid resolution with ∼15 million particles to 33 seconds on a 1 mm grid with ∼460 million particles. Eclipse calculation runtimes distributed over 10 FAS workers were 9 seconds to 15 minutes respectively. Conclusion: The GPU accelerated second check allows quick MU verification while accounting for patient specific geometry and heterogeneity.« less

  12. Hurricane Risk Variability along the Gulf of Mexico Coastline

    PubMed Central

    Trepanier, Jill C.; Ellis, Kelsey N.; Tucker, Clay S.

    2015-01-01

    Hurricane risk characteristics are examined across the U. S. Gulf of Mexico coastline using a hexagonal tessellation. Using an extreme value model, parameters are collected representing the rate or λ (frequency), the scale or σ (range), and the shape or ξ (intensity) of the extreme wind distribution. These latent parameters and the 30-year return level are visualized across the grid. The greatest 30-year return levels are located toward the center of the Gulf of Mexico, and for inland locations, along the borders of Louisiana, Mississippi, and Alabama. Using a geographically weighted regression model, the relationship of these parameters to sea surface temperature (SST) is found to assess sensitivity to change. It is shown that as SSTs increase near the coast, the frequency of hurricanes in these grids decrease significantly. This reinforces the importance of SST in areas of likely tropical cyclogenesis in determining the number of hurricanes near the coast, along with SSTs along the lifespan of the storm, rather than simply local SST. The range of hurricane wind speeds experienced near Florida is shown to increase with increasing SSTs (insignificant), suggesting that increased temperatures may allow hurricanes to maintain their strength as they pass over the Florida peninsula. The modifiable areal unit problem is assessed using multiple grid sizes. Moran’s I and the local statistic G are calculated to examine spatial autocorrelation in the parameters. This research opens up future questions regarding rapid intensification and decay close to the coast and the relationship to changing SSTs. PMID:25767885

  13. A new NASA/MSFC mission analysis global cloud cover data base

    NASA Technical Reports Server (NTRS)

    Brown, S. C.; Jeffries, W. R., III

    1985-01-01

    A global cloud cover data set, derived from the USAF 3D NEPH Analysis, was developed for use in climate studies and for Earth viewing applications. This data set contains a single parameter - total sky cover - separated in time by 3 or 6 hr intervals and in space by approximately 50 n.mi. Cloud cover amount is recorded for each grid point (of a square grid) by a single alphanumeric character representing each 5 percent increment of sky cover. The data are arranged in both quarterly and monthly formats. The data base currently provides daily, 3-hr observed total sky cover for the Northern Hemisphere from 1972 through 1977 less 1976. For the Southern Hemisphere, there are data at 6-hr intervals for 1976 through 1978 and at 3-hr intervals for 1979 and 1980. More years of data are being added. To validate the data base, the percent frequency of or = 0.3 and or = 0.8 cloud cover was compared with ground observed cloud amounts at several locations with generally good agreement. Mean or other desired cloud amounts can be calculated for any time period and any size area from a single grid point to a hemisphere. The data base is especially useful in evaluating the consequence of cloud cover on Earth viewing space missions. The temporal and spatial frequency of the data allow simulations that closely approximate any projected viewing mission. No adjustments are required to account for cloud continuity.

  14. Hurricane risk variability along the Gulf of Mexico coastline.

    PubMed

    Trepanier, Jill C; Ellis, Kelsey N; Tucker, Clay S

    2015-01-01

    Hurricane risk characteristics are examined across the U. S. Gulf of Mexico coastline using a hexagonal tessellation. Using an extreme value model, parameters are collected representing the rate or λ (frequency), the scale or σ (range), and the shape or ξ (intensity) of the extreme wind distribution. These latent parameters and the 30-year return level are visualized across the grid. The greatest 30-year return levels are located toward the center of the Gulf of Mexico, and for inland locations, along the borders of Louisiana, Mississippi, and Alabama. Using a geographically weighted regression model, the relationship of these parameters to sea surface temperature (SST) is found to assess sensitivity to change. It is shown that as SSTs increase near the coast, the frequency of hurricanes in these grids decrease significantly. This reinforces the importance of SST in areas of likely tropical cyclogenesis in determining the number of hurricanes near the coast, along with SSTs along the lifespan of the storm, rather than simply local SST. The range of hurricane wind speeds experienced near Florida is shown to increase with increasing SSTs (insignificant), suggesting that increased temperatures may allow hurricanes to maintain their strength as they pass over the Florida peninsula. The modifiable areal unit problem is assessed using multiple grid sizes. Moran's I and the local statistic G are calculated to examine spatial autocorrelation in the parameters. This research opens up future questions regarding rapid intensification and decay close to the coast and the relationship to changing SSTs.

  15. A Simple Algebraic Grid Adaptation Scheme with Applications to Two- and Three-dimensional Flow Problems

    NASA Technical Reports Server (NTRS)

    Hsu, Andrew T.; Lytle, John K.

    1989-01-01

    An algebraic adaptive grid scheme based on the concept of arc equidistribution is presented. The scheme locally adjusts the grid density based on gradients of selected flow variables from either finite difference or finite volume calculations. A user-prescribed grid stretching can be specified such that control of the grid spacing can be maintained in areas of known flowfield behavior. For example, the grid can be clustered near a wall for boundary layer resolution and made coarse near the outer boundary of an external flow. A grid smoothing technique is incorporated into the adaptive grid routine, which is found to be more robust and efficient than the weight function filtering technique employed by other researchers. Since the present algebraic scheme requires no iteration or solution of differential equations, the computer time needed for grid adaptation is trivial, making the scheme useful for three-dimensional flow problems. Applications to two- and three-dimensional flow problems show that a considerable improvement in flowfield resolution can be achieved by using the proposed adaptive grid scheme. Although the scheme was developed with steady flow in mind, it is a good candidate for unsteady flow computations because of its efficiency.

  16. Simulations of the DARPA Suboff Submarine Including Self-Propulsion with the E1619 Propeller

    DTIC Science & Technology

    2012-01-01

    and experiments are remarkable, including the maximum velocity in the wake of the 37 blades , the velocity deficit induced by the tip vortices...added to the wake matches the grid size of the fine grids used for the tips of the blades , thus providing a grid of consistent refinement for the...geometry or larger number of blades for the same advance coefficient. These two mechanisms in a marine propeller lead to larger induced wake

  17. Evaluation of subgrid-scale turbulence models using a fully simulated turbulent flow

    NASA Technical Reports Server (NTRS)

    Clark, R. A.; Ferziger, J. H.; Reynolds, W. C.

    1977-01-01

    An exact turbulent flow field was calculated on a three-dimensional grid with 64 points on a side. The flow simulates grid-generated turbulence from wind tunnel experiments. In this simulation, the grid spacing is small enough to include essentially all of the viscous energy dissipation, and the box is large enough to contain the largest eddy in the flow. The method is limited to low-turbulence Reynolds numbers, in our case R sub lambda = 36.6. To complete the calculation using a reasonable amount of computer time with reasonable accuracy, a third-order time-integration scheme was developed which runs at about the same speed as a simple first-order scheme. It obtains this accuracy by saving the velocity field and its first-time derivative at each time step. Fourth-order accurate space-differencing is used.

  18. Information Theoretically Secure, Enhanced Johnson Noise Based Key Distribution over the Smart Grid with Switched Filters

    PubMed Central

    2013-01-01

    We introduce a protocol with a reconfigurable filter system to create non-overlapping single loops in the smart power grid for the realization of the Kirchhoff-Law-Johnson-(like)-Noise secure key distribution system. The protocol is valid for one-dimensional radial networks (chain-like power line) which are typical of the electricity distribution network between the utility and the customer. The speed of the protocol (the number of steps needed) versus grid size is analyzed. When properly generalized, such a system has the potential to achieve unconditionally secure key distribution over the smart power grid of arbitrary geometrical dimensions. PMID:23936164

  19. Information theoretically secure, enhanced Johnson noise based key distribution over the smart grid with switched filters.

    PubMed

    Gonzalez, Elias; Kish, Laszlo B; Balog, Robert S; Enjeti, Prasad

    2013-01-01

    We introduce a protocol with a reconfigurable filter system to create non-overlapping single loops in the smart power grid for the realization of the Kirchhoff-Law-Johnson-(like)-Noise secure key distribution system. The protocol is valid for one-dimensional radial networks (chain-like power line) which are typical of the electricity distribution network between the utility and the customer. The speed of the protocol (the number of steps needed) versus grid size is analyzed. When properly generalized, such a system has the potential to achieve unconditionally secure key distribution over the smart power grid of arbitrary geometrical dimensions.

  20. A Solution Adaptive Structured/Unstructured Overset Grid Flow Solver with Applications to Helicopter Rotor Flows

    NASA Technical Reports Server (NTRS)

    Duque, Earl P. N.; Biswas, Rupak; Strawn, Roger C.

    1995-01-01

    This paper summarizes a method that solves both the three dimensional thin-layer Navier-Stokes equations and the Euler equations using overset structured and solution adaptive unstructured grids with applications to helicopter rotor flowfields. The overset structured grids use an implicit finite-difference method to solve the thin-layer Navier-Stokes/Euler equations while the unstructured grid uses an explicit finite-volume method to solve the Euler equations. Solutions on a helicopter rotor in hover show the ability to accurately convect the rotor wake. However, isotropic subdivision of the tetrahedral mesh rapidly increases the overall problem size.

  1. Arithmetic Data Cube as a Data Intensive Benchmark

    NASA Technical Reports Server (NTRS)

    Frumkin, Michael A.; Shabano, Leonid

    2003-01-01

    Data movement across computational grids and across memory hierarchy of individual grid machines is known to be a limiting factor for application involving large data sets. In this paper we introduce the Data Cube Operator on an Arithmetic Data Set which we call Arithmetic Data Cube (ADC). We propose to use the ADC to benchmark grid capabilities to handle large distributed data sets. The ADC stresses all levels of grid memory by producing 2d views of an Arithmetic Data Set of d-tuples described by a small number of parameters. We control data intensity of the ADC by controlling the sizes of the views through choice of the tuple parameters.

  2. Simulation of wave propagation in three-dimensional random media

    NASA Technical Reports Server (NTRS)

    Coles, William A.; Filice, J. P.; Frehlich, R. G.; Yadlowsky, M.

    1993-01-01

    Quantitative error analysis for simulation of wave propagation in three dimensional random media assuming narrow angular scattering are presented for the plane wave and spherical wave geometry. This includes the errors resulting from finite grid size, finite simulation dimensions, and the separation of the two-dimensional screens along the propagation direction. Simple error scalings are determined for power-law spectra of the random refractive index of the media. The effects of a finite inner scale are also considered. The spatial spectra of the intensity errors are calculated and compared to the spatial spectra of intensity. The numerical requirements for a simulation of given accuracy are determined for realizations of the field. The numerical requirements for accurate estimation of higher moments of the field are less stringent.

  3. Linear scaling computation of the Fock matrix. VI. Data parallel computation of the exchange-correlation matrix

    NASA Astrophysics Data System (ADS)

    Gan, Chee Kwan; Challacombe, Matt

    2003-05-01

    Recently, early onset linear scaling computation of the exchange-correlation matrix has been achieved using hierarchical cubature [J. Chem. Phys. 113, 10037 (2000)]. Hierarchical cubature differs from other methods in that the integration grid is adaptive and purely Cartesian, which allows for a straightforward domain decomposition in parallel computations; the volume enclosing the entire grid may be simply divided into a number of nonoverlapping boxes. In our data parallel approach, each box requires only a fraction of the total density to perform the necessary numerical integrations due to the finite extent of Gaussian-orbital basis sets. This inherent data locality may be exploited to reduce communications between processors as well as to avoid memory and copy overheads associated with data replication. Although the hierarchical cubature grid is Cartesian, naive boxing leads to irregular work loads due to strong spatial variations of the grid and the electron density. In this paper we describe equal time partitioning, which employs time measurement of the smallest sub-volumes (corresponding to the primitive cubature rule) to load balance grid-work for the next self-consistent-field iteration. After start-up from a heuristic center of mass partitioning, equal time partitioning exploits smooth variation of the density and grid between iterations to achieve load balance. With the 3-21G basis set and a medium quality grid, equal time partitioning applied to taxol (62 heavy atoms) attained a speedup of 61 out of 64 processors, while for a 110 molecule water cluster at standard density it achieved a speedup of 113 out of 128. The efficiency of equal time partitioning applied to hierarchical cubature improves as the grid work per processor increases. With a fine grid and the 6-311G(df,p) basis set, calculations on the 26 atom molecule α-pinene achieved a parallel efficiency better than 99% with 64 processors. For more coarse grained calculations, superlinear speedups are found to result from reduced computational complexity associated with data parallelism.

  4. Three-phase short circuit calculation method based on pre-computed surface for doubly fed induction generator

    NASA Astrophysics Data System (ADS)

    Ma, J.; Liu, Q.

    2018-02-01

    This paper presents an improved short circuit calculation method, based on pre-computed surface to determine the short circuit current of a distribution system with multiple doubly fed induction generators (DFIGs). The short circuit current, injected into power grid by DFIG, is determined by low voltage ride through (LVRT) control and protection under grid fault. However, the existing methods are difficult to calculate the short circuit current of DFIG in engineering practice due to its complexity. A short circuit calculation method, based on pre-computed surface, was proposed by developing the surface of short circuit current changing with the calculating impedance and the open circuit voltage. And the short circuit currents were derived by taking into account the rotor excitation and crowbar activation time. Finally, the pre-computed surfaces of short circuit current at different time were established, and the procedure of DFIG short circuit calculation considering its LVRT was designed. The correctness of proposed method was verified by simulation.

  5. A comparative analysis of dynamic grids vs. virtual grids using the A3pviGrid framework.

    PubMed

    Shankaranarayanan, Avinas; Amaldas, Christine

    2010-11-01

    With the proliferation of Quad/Multi-core micro-processors in mainstream platforms such as desktops and workstations; a large number of unused CPU cycles can be utilized for running virtual machines (VMs) as dynamic nodes in distributed environments. Grid services and its service oriented business broker now termed cloud computing could deploy image based virtualization platforms enabling agent based resource management and dynamic fault management. In this paper we present an efficient way of utilizing heterogeneous virtual machines on idle desktops as an environment for consumption of high performance grid services. Spurious and exponential increases in the size of the datasets are constant concerns in medical and pharmaceutical industries due to the constant discovery and publication of large sequence databases. Traditional algorithms are not modeled at handing large data sizes under sudden and dynamic changes in the execution environment as previously discussed. This research was undertaken to compare our previous results with running the same test dataset with that of a virtual Grid platform using virtual machines (Virtualization). The implemented architecture, A3pviGrid utilizes game theoretic optimization and agent based team formation (Coalition) algorithms to improve upon scalability with respect to team formation. Due to the dynamic nature of distributed systems (as discussed in our previous work) all interactions were made local within a team transparently. This paper is a proof of concept of an experimental mini-Grid test-bed compared to running the platform on local virtual machines on a local test cluster. This was done to give every agent its own execution platform enabling anonymity and better control of the dynamic environmental parameters. We also analyze performance and scalability of Blast in a multiple virtual node setup and present our findings. This paper is an extension of our previous research on improving the BLAST application framework using dynamic Grids on virtualization platforms such as the virtual box.

  6. Development of a Global Fire Weather Database

    NASA Technical Reports Server (NTRS)

    Field, R. D.; Spessa, A. C.; Aziz, N. A.; Camia, A.; Cantin, A.; Carr, R.; de Groot, W. J.; Dowdy, A. J.; Flannigan, M. D.; Manomaiphiboon, K.; hide

    2015-01-01

    The Canadian Forest Fire Weather Index (FWI) System is the mostly widely used fire danger rating system in the world. We have developed a global database of daily FWI System calculations, beginning in 1980, called the Global Fire WEather Database (GFWED) gridded to a spatial resolution of 0.5 latitude by 2/3 longitude. Input weather data were obtained from the NASA Modern Era Retrospective- Analysis for Research and Applications (MERRA), and two different estimates of daily precipitation from rain gauges over land. FWI System Drought Code calculations from the gridded data sets were compared to calculations from individual weather station data for a representative set of 48 stations in North, Central and South America, Europe, Russia, Southeast Asia and Australia. Agreement between gridded calculations and the station-based calculations tended to be most different at low latitudes for strictly MERRA based calculations. Strong biases could be seen in either direction: MERRA DC over the Mato Grosso in Brazil reached unrealistically high values exceeding DCD1500 during the dry season but was too low over Southeast Asia during the dry season. These biases are consistent with those previously identified in MERRA's precipitation, and they reinforce the need to consider alternative sources of precipitation data. GFWED can be used for analyzing historical relationships between fire weather and fire activity at continental and global scales, in identifying large-scale atmosphere-ocean controls on fire weather, and calibration of FWI-based fire prediction models.

  7. Normal modes of the shallow water system on the cubed sphere

    NASA Astrophysics Data System (ADS)

    Kang, H. G.; Cheong, H. B.; Lee, C. H.

    2017-12-01

    Spherical harmonics expressed as the Rossby-Haurwitz waves are the normal modes of non-divergent barotropic model. Among the normal modes in the numerical models, the most unstable mode will contaminate the numerical results, and therefore the investigation of normal mode for a given grid system and a discretiztaion method is important. The cubed-sphere grid which consists of six identical faces has been widely adopted in many atmospheric models. This grid system is non-orthogonal grid so that calculation of the normal mode is quiet challenge problem. In the present study, the normal modes of the shallow water system on the cubed sphere discretized by the spectral element method employing the Gauss-Lobatto Lagrange interpolating polynomials as orthogonal basis functions is investigated. The algebraic equations for the shallow water equation on the cubed sphere are derived, and the huge global matrix is constructed. The linear system representing the eigenvalue-eigenvector relations is solved by numerical libraries. The normal mode calculated for the several horizontal resolution and lamb parameters will be discussed and compared to the normal mode from the spherical harmonics spectral method.

  8. Dose calculation for photon-emitting brachytherapy sources with average energy higher than 50 keV: report of the AAPM and ESTRO.

    PubMed

    Perez-Calatayud, Jose; Ballester, Facundo; Das, Rupak K; Dewerd, Larry A; Ibbott, Geoffrey S; Meigooni, Ali S; Ouhib, Zoubir; Rivard, Mark J; Sloboda, Ron S; Williamson, Jeffrey F

    2012-05-01

    Recommendations of the American Association of Physicists in Medicine (AAPM) and the European Society for Radiotherapy and Oncology (ESTRO) on dose calculations for high-energy (average energy higher than 50 keV) photon-emitting brachytherapy sources are presented, including the physical characteristics of specific (192)Ir, (137)Cs, and (60)Co source models. This report has been prepared by the High Energy Brachytherapy Source Dosimetry (HEBD) Working Group. This report includes considerations in the application of the TG-43U1 formalism to high-energy photon-emitting sources with particular attention to phantom size effects, interpolation accuracy dependence on dose calculation grid size, and dosimetry parameter dependence on source active length. Consensus datasets for commercially available high-energy photon sources are provided, along with recommended methods for evaluating these datasets. Recommendations on dosimetry characterization methods, mainly using experimental procedures and Monte Carlo, are established and discussed. Also included are methodological recommendations on detector choice, detector energy response characterization and phantom materials, and measurement specification methodology. Uncertainty analyses are discussed and recommendations for high-energy sources without consensus datasets are given. Recommended consensus datasets for high-energy sources have been derived for sources that were commercially available as of January 2010. Data are presented according to the AAPM TG-43U1 formalism, with modified interpolation and extrapolation techniques of the AAPM TG-43U1S1 report for the 2D anisotropy function and radial dose function.

  9. Lattice dynamics and electron-phonon coupling calculations using nondiagonal supercells

    NASA Astrophysics Data System (ADS)

    Lloyd-Williams, Jonathan; Monserrat, Bartomeu

    Quantities derived from electron-phonon coupling matrix elements require a fine sampling of the vibrational Brillouin zone. Converged results are typically not obtainable using the direct method, in which a perturbation is frozen into the system and the total energy derivatives are calculated using a finite difference approach, because the size of simulation cell needed is prohibitively large. We show that it is possible to determine the response of a periodic system to a perturbation characterized by a wave vector with reduced fractional coordinates (m1 /n1 ,m2 /n2 ,m3 /n3) using a supercell containing a number of primitive cells equal to the least common multiple of n1, n2, and n3. This is accomplished by utilizing supercell matrices containing nonzero off-diagonal elements. We present the results of electron-phonon coupling calculations using the direct method to sample the vibrational Brillouin zone with grids of unprecedented size for a range of systems, including the canonical example of diamond. We also demonstrate that the use of nondiagonal supercells reduces by over an order of magnitude the computational cost of obtaining converged vibrational densities of states and phonon dispersion curves. J.L.-W. is supported by the Engineering and Physical Sciences Research Council (EPSRC). B.M. is supported by Robinson College, Cambridge, and the Cambridge Philosophical Society. This work was supported by EPSRC Grants EP/J017639/1 and EP/K013564/1.

  10. Aerodynamic and heat transfer analysis of the low aspect ratio turbine using a 3D Navier-Stokes code

    NASA Astrophysics Data System (ADS)

    Choi, D.; Knight, C. J.

    1991-06-01

    The single-stage, high-pressure ratio Garrett Low Aspect Ratio Turbine (LART) test data obtained in a shock tunnel are employed as a basis for evaluating a new three-dimensional Navier Stokes code based on the O-H grid system. It uses Coakley's two-equation turbulence modeling with viscous sublayer resolution. For the nozzle guide vanes, calculations were made based on two grid zones: an O-grid zone wrapping around airfoil and an H-grid zone outside of the O-grid zone, including the regions upstream of the leadig edge and downstream of the trailing edge. For the rotor blade row, a third O-grid zone was added for the tip-gap region leakage flow. The computational results compare well with experiment. These comparisons include heat transfer distributions on the airfoils and end-walls. The leakage flow through the tip-gap clearance is well resolved.

  11. A generic efficient adaptive grid scheme for rocket propulsion modeling

    NASA Technical Reports Server (NTRS)

    Mo, J. D.; Chow, Alan S.

    1993-01-01

    The objective of this research is to develop an efficient, time-accurate numerical algorithm to discretize the Navier-Stokes equations for the predictions of internal one-, two-dimensional and axisymmetric flows. A generic, efficient, elliptic adaptive grid generator is implicitly coupled with the Lower-Upper factorization scheme in the development of ALUNS computer code. The calculations of one-dimensional shock tube wave propagation and two-dimensional shock wave capture, wave-wave interactions, shock wave-boundary interactions show that the developed scheme is stable, accurate and extremely robust. The adaptive grid generator produced a very favorable grid network by a grid speed technique. This generic adaptive grid generator is also applied in the PARC and FDNS codes and the computational results for solid rocket nozzle flowfield and crystal growth modeling by those codes will be presented in the conference, too. This research work is being supported by NASA/MSFC.

  12. Efficient Cache use for Stencil Operations on Structured Discretization Grids

    NASA Technical Reports Server (NTRS)

    Frumkin, Michael; VanderWijngaart, Rob F.

    2001-01-01

    We derive tight bounds on the cache misses for evaluation of explicit stencil operators on structured grids. Our lower bound is based on the isoperimetrical property of the discrete octahedron. Our upper bound is based on a good surface to volume ratio of a parallelepiped spanned by a reduced basis of the interference lattice of a grid. Measurements show that our algorithm typically reduces the number of cache misses by a factor of three, relative to a compiler optimized code. We show that stencil calculations on grids whose interference lattice have a short vector feature abnormally high numbers of cache misses. We call such grids unfavorable and suggest to avoid these in computations by appropriate padding. By direct measurements on a MIPS R10000 processor we show a good correlation between abnormally high numbers of cache misses and unfavorable three-dimensional grids.

  13. Grid generation methodology and CFD simulations in sliding vane compressors and expanders

    NASA Astrophysics Data System (ADS)

    Bianchi, Giuseppe; Rane, Sham; Kovacevic, Ahmed; Cipollone, Roberto; Murgia, Stefano; Contaldi, Giulio

    2017-08-01

    The limiting factor for the employment of advanced 3D CFD tools in the analysis and design of rotary vane machines is the unavailability of methods for generation of computational grids suitable for fast and reliable numerical analysis. The paper addresses this challenge presenting the development of an analytical grid generation for vane machines that is based on the user defined nodal displacement. In particular, mesh boundaries are defined as parametric curves generated using trigonometrical modelling of the axial cross section of the machine while the distribution of computational nodes is performed using algebraic algorithms with transfinite interpolation, post orthogonalisation and smoothing. Algebraic control functions are introduced for distribution of nodes on the rotor and casing boundaries in order to achieve good grid quality in terms of cell size and expansion. In this way, the moving and deforming fluid domain of the sliding vane machine is discretized and the conservation of intrinsic quantities in ensured by maintaining the cell connectivity and structure. For validation of generated grids, a mid-size air compressor and a small-scale expander for Organic Rankine Cycle applications have been investigated in this paper. Remarks on implementation of the mesh motion algorithm, stability and robustness experienced with the ANSYS CFX solver as well as the obtained flow results are presented.

  14. Towards the Development of a More Accurate Monitoring Procedure for Invertebrate Populations, in the Presence of an Unknown Spatial Pattern of Population Distribution in the Field

    PubMed Central

    Petrovskaya, Natalia B.; Forbes, Emily; Petrovskii, Sergei V.; Walters, Keith F. A.

    2018-01-01

    Studies addressing many ecological problems require accurate evaluation of the total population size. In this paper, we revisit a sampling procedure used for the evaluation of the abundance of an invertebrate population from assessment data collected on a spatial grid of sampling locations. We first discuss how insufficient information about the spatial population density obtained on a coarse sampling grid may affect the accuracy of an evaluation of total population size. Such information deficit in field data can arise because of inadequate spatial resolution of the population distribution (spatially variable population density) when coarse grids are used, which is especially true when a strongly heterogeneous spatial population density is sampled. We then argue that the average trap count (the quantity routinely used to quantify abundance), if obtained from a sampling grid that is too coarse, is a random variable because of the uncertainty in sampling spatial data. Finally, we show that a probabilistic approach similar to bootstrapping techniques can be an efficient tool to quantify the uncertainty in the evaluation procedure in the presence of a spatial pattern reflecting a patchy distribution of invertebrates within the sampling grid. PMID:29495513

  15. Fabrication and characterization of self-folding thermoplastic sheets using unbalanced thermal shrinkage.

    PubMed

    Danielson, Christian; Mehrnezhad, Ali; YekrangSafakar, Ashkan; Park, Kidong

    2017-06-14

    Self-folding or micro-origami technologies are actively investigated as a novel manufacturing process to fabricate three-dimensional macro/micro-structures. In this paper, we present a simple process to produce a self-folding structure with a biaxially oriented polystyrene sheet (BOPS) or Shrinky Dinks. A BOPS sheet is known to shrink to one-third of its original size in plane, when it is heated above 160 °C. A grid pattern is engraved on one side of the BOPS film with a laser engraver to decrease the thermal shrinkage of the engraved side. The thermal shrinkage of the non-engraved side remains the same and this unbalanced thermal shrinkage causes folding of the structure as the structure shrinks at high temperature. We investigated the self-folding mechanism and characterized how the grid geometry, the grid size, and the power of the laser engraver affect the bending curvature. The developed fabrication process to locally modulate thermomechanical properties of the material by engraving the grid pattern and the demonstrated design methodology to harness the unbalanced thermal shrinkage can be applied to develop complicated self-folding macro/micro structures.

  16. Multi-Resolution Unstructured Grid-Generation for Geophysical Applications on the Sphere

    NASA Technical Reports Server (NTRS)

    Engwirda, Darren

    2015-01-01

    An algorithm for the generation of non-uniform unstructured grids on ellipsoidal geometries is described. This technique is designed to generate high quality triangular and polygonal meshes appropriate for general circulation modelling on the sphere, including applications to atmospheric and ocean simulation, and numerical weather predication. Using a recently developed Frontal-Delaunay-refinement technique, a method for the construction of high-quality unstructured ellipsoidal Delaunay triangulations is introduced. A dual polygonal grid, derived from the associated Voronoi diagram, is also optionally generated as a by-product. Compared to existing techniques, it is shown that the Frontal-Delaunay approach typically produces grids with near-optimal element quality and smooth grading characteristics, while imposing relatively low computational expense. Initial results are presented for a selection of uniform and non-uniform ellipsoidal grids appropriate for large-scale geophysical applications. The use of user-defined mesh-sizing functions to generate smoothly graded, non-uniform grids is discussed.

  17. An incompressible two-dimensional multiphase particle-in-cell model for dense particle flows

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Snider, D.M.; O`Rourke, P.J.; Andrews, M.J.

    1997-06-01

    A two-dimensional, incompressible, multiphase particle-in-cell (MP-PIC) method is presented for dense particle flows. The numerical technique solves the governing equations of the fluid phase using a continuum model and those of the particle phase using a Lagrangian model. Difficulties associated with calculating interparticle interactions for dense particle flows with volume fractions above 5% have been eliminated by mapping particle properties to a Eulerian grid and then mapping back computed stress tensors to particle positions. This approach utilizes the best of Eulerian/Eulerian continuum models and Eulerian/Lagrangian discrete models. The solution scheme allows for distributions of types, sizes, and density of particles,more » with no numerical diffusion from the Lagrangian particle calculations. The computational method is implicit with respect to pressure, velocity, and volume fraction in the continuum solution thus avoiding courant limits on computational time advancement. MP-PIC simulations are compared with one-dimensional problems that have analytical solutions and with two-dimensional problems for which there are experimental data.« less

  18. Fast and accurate grid representations for atom-based docking with partner flexibility.

    PubMed

    de Vries, Sjoerd J; Zacharias, Martin

    2017-06-30

    Macromolecular docking methods can broadly be divided into geometric and atom-based methods. Geometric methods use fast algorithms that operate on simplified, grid-like molecular representations, while atom-based methods are more realistic and flexible, but far less efficient. Here, a hybrid approach of grid-based and atom-based docking is presented, combining precalculated grid potentials with neighbor lists for fast and accurate calculation of atom-based intermolecular energies and forces. The grid representation is compatible with simultaneous multibody docking and can tolerate considerable protein flexibility. When implemented in our docking method ATTRACT, grid-based docking was found to be ∼35x faster. With the OPLSX forcefield instead of the ATTRACT coarse-grained forcefield, the average speed improvement was >100x. Grid-based representations may allow atom-based docking methods to explore large conformational spaces with many degrees of freedom, such as multiple macromolecules including flexibility. This increases the domain of biological problems to which docking methods can be applied. © 2017 Wiley Periodicals, Inc. © 2017 Wiley Periodicals, Inc.

  19. AVQS: attack route-based vulnerability quantification scheme for smart grid.

    PubMed

    Ko, Jongbin; Lim, Hyunwoo; Lee, Seokjun; Shon, Taeshik

    2014-01-01

    A smart grid is a large, consolidated electrical grid system that includes heterogeneous networks and systems. Based on the data, a smart grid system has a potential security threat in its network connectivity. To solve this problem, we develop and apply a novel scheme to measure the vulnerability in a smart grid domain. Vulnerability quantification can be the first step in security analysis because it can help prioritize the security problems. However, existing vulnerability quantification schemes are not suitable for smart grid because they do not consider network vulnerabilities. We propose a novel attack route-based vulnerability quantification scheme using a network vulnerability score and an end-to-end security score, depending on the specific smart grid network environment to calculate the vulnerability score for a particular attack route. To evaluate the proposed approach, we derive several attack scenarios from the advanced metering infrastructure domain. The experimental results of the proposed approach and the existing common vulnerability scoring system clearly show that we need to consider network connectivity for more optimized vulnerability quantification.

  20. Streamline integration as a method for two-dimensional elliptic grid generation

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Wiesenberger, M., E-mail: Matthias.Wiesenberger@uibk.ac.at; Held, M.; Einkemmer, L.

    We propose a new numerical algorithm to construct a structured numerical elliptic grid of a doubly connected domain. Our method is applicable to domains with boundaries defined by two contour lines of a two-dimensional function. Furthermore, we can adapt any analytically given boundary aligned structured grid, which specifically includes polar and Cartesian grids. The resulting coordinate lines are orthogonal to the boundary. Grid points as well as the elements of the Jacobian matrix can be computed efficiently and up to machine precision. In the simplest case we construct conformal grids, yet with the help of weight functions and monitor metricsmore » we can control the distribution of cells across the domain. Our algorithm is parallelizable and easy to implement with elementary numerical methods. We assess the quality of grids by considering both the distribution of cell sizes and the accuracy of the solution to elliptic problems. Among the tested grids these key properties are best fulfilled by the grid constructed with the monitor metric approach. - Graphical abstract: - Highlights: • Construct structured, elliptic numerical grids with elementary numerical methods. • Align coordinate lines with or make them orthogonal to the domain boundary. • Compute grid points and metric elements up to machine precision. • Control cell distribution by adaption functions or monitor metrics.« less

  1. Grid-connected photovoltaic (PV) systems with batteries storage as solution to electrical grid outages in Burkina Faso

    NASA Astrophysics Data System (ADS)

    Abdoulaye, D.; Koalaga, Z.; Zougmore, F.

    2012-02-01

    This paper deals with a key solution for power outages problem experienced by many African countries and this through grid-connected photovoltaic (PV) systems with batteries storage. African grids are characterized by an insufficient power supply and frequent interruptions. Due to this fact, users who especially use classical grid-connected photovoltaic systems are unable to profit from their installation even if there is sun. In this study, we suggest the using of a grid-connected photovoltaic system with batteries storage as a solution to these problems. This photovoltaic system works by injecting the surplus of electricity production into grid and can also deliver electricity as a stand-alone system with all security needed. To achieve our study objectives, firstly we conducted a survey of a real situation of one African electrical grid, the case of Burkina Faso (SONABEL: National Electricity Company of Burkina). Secondly, as study case, we undertake a sizing, a modeling and a simulation of a grid-connected PV system with batteries storage for the LAME laboratory at the University of Ouagadougou. The simulation shows that the proposed grid-connected system allows users to profit from their photovoltaic installation at any time even if the public electrical grid has some failures either during the day or at night.

  2. Power Grid Construction Project Portfolio Optimization Based on Bi-level programming model

    NASA Astrophysics Data System (ADS)

    Zhao, Erdong; Li, Shangqi

    2017-08-01

    As the main body of power grid operation, county-level power supply enterprises undertake an important emission to guarantee the security of power grid operation and safeguard social power using order. The optimization of grid construction projects has been a key issue of power supply capacity and service level of grid enterprises. According to the actual situation of power grid construction project optimization of county-level power enterprises, on the basis of qualitative analysis of the projects, this paper builds a Bi-level programming model based on quantitative analysis. The upper layer of the model is the target restriction of the optimal portfolio; the lower layer of the model is enterprises’ financial restrictions on the size of the enterprise project portfolio. Finally, using a real example to illustrate operation proceeding and the optimization result of the model. Through qualitative analysis and quantitative analysis, the bi-level programming model improves the accuracy and normative standardization of power grid enterprises projects.

  3. Turbulent premixed flames on fractal-grid-generated turbulence

    NASA Astrophysics Data System (ADS)

    Soulopoulos, N.; Kerl, J.; Sponfeldner, T.; Beyrau, F.; Hardalupas, Y.; Taylor, A. M. K. P.; Vassilicos, J. C.

    2013-12-01

    A space-filling, low blockage fractal grid is used as a novel turbulence generator in a premixed turbulent flame stabilized by a rod. The study compares the flame behaviour with a fractal grid to the behaviour when a standard square mesh grid with the same effective mesh size and solidity as the fractal grid is used. The isothermal gas flow turbulence characteristics, including mean flow velocity and rms of velocity fluctuations and Taylor length, were evaluated from hot-wire measurements. The behaviour of the flames was assessed with direct chemiluminescence emission from the flame and high-speed OH-laser-induced fluorescence. The characteristics of the two flames are considered in terms of turbulent flame thickness, local flame curvature and turbulent flame speed. It is found that, for the same flow rate and stoichiometry and at the same distance downstream of the location of the grid, fractal-grid-generated turbulence leads to a more turbulent flame with enhanced burning rate and increased flame surface area.

  4. Efficient algorithm for locating and sizing series compensation devices in large power transmission grids: I. Model implementation

    NASA Astrophysics Data System (ADS)

    Frolov, Vladimir; Backhaus, Scott; Chertkov, Misha

    2014-10-01

    We explore optimization methods for planning the placement, sizing and operations of flexible alternating current transmission system (FACTS) devices installed to relieve transmission grid congestion. We limit our selection of FACTS devices to series compensation (SC) devices that can be represented by modification of the inductance of transmission lines. Our master optimization problem minimizes the l1 norm of the inductance modification subject to the usual line thermal-limit constraints. We develop heuristics that reduce this non-convex optimization to a succession of linear programs (LP) that are accelerated further using cutting plane methods. The algorithm solves an instance of the MatPower Polish Grid model (3299 lines and 2746 nodes) in 40 seconds per iteration on a standard laptop—a speed that allows the sizing and placement of a family of SC devices to correct a large set of anticipated congestions. We observe that our algorithm finds feasible solutions that are always sparse, i.e., SC devices are placed on only a few lines. In a companion manuscript, we demonstrate our approach on realistically sized networks that suffer congestion from a range of causes, including generator retirement. In this manuscript, we focus on the development of our approach, investigate its structure on a small test system subject to congestion from uniform load growth, and demonstrate computational efficiency on a realistically sized network.

  5. Efficient algorithm for locating and sizing series compensation devices in large power transmission grids: I. Model implementation

    DOE PAGES

    Frolov, Vladimir; Backhaus, Scott; Chertkov, Misha

    2014-10-24

    We explore optimization methods for planning the placement, sizing and operations of Flexible Alternating Current Transmission System (FACTS) devices installed to relieve transmission grid congestion. We limit our selection of FACTS devices to Series Compensation (SC) devices that can be represented by modification of the inductance of transmission lines. Our master optimization problem minimizes the l 1 norm of the inductance modification subject to the usual line thermal-limit constraints. We develop heuristics that reduce this non-convex optimization to a succession of Linear Programs (LP) which are accelerated further using cutting plane methods. The algorithm solves an instance of the MatPowermore » Polish Grid model (3299 lines and 2746 nodes) in 40 seconds per iteration on a standard laptop—a speed up that allows the sizing and placement of a family of SC devices to correct a large set of anticipated congestions. We observe that our algorithm finds feasible solutions that are always sparse, i.e., SC devices are placed on only a few lines. In a companion manuscript, we demonstrate our approach on realistically-sized networks that suffer congestion from a range of causes including generator retirement. In this manuscript, we focus on the development of our approach, investigate its structure on a small test system subject to congestion from uniform load growth, and demonstrate computational efficiency on a realistically-sized network.« less

  6. A scale-invariant cellular-automata model for distributed seismicity

    NASA Technical Reports Server (NTRS)

    Barriere, Benoit; Turcotte, Donald L.

    1991-01-01

    In the standard cellular-automata model for a fault an element of stress is randomly added to a grid of boxes until a box has four elements, these are then redistributed to the adjacent boxes on the grid. The redistribution can result in one or more of these boxes having four or more elements in which case further redistributions are required. On the average added elements are lost from the edges of the grid. The model is modified so that the boxes have a scale-invariant distribution of sizes. The objective is to model a scale-invariant distribution of fault sizes. When a redistribution from a box occurs it is equivalent to a characteristic earthquake on the fault. A redistribution from a small box (a foreshock) can trigger an instability in a large box (the main shock). A redistribution from a large box always triggers many instabilities in the smaller boxes (aftershocks). The frequency-size statistics for both main shocks and aftershocks satisfy the Gutenberg-Richter relation with b = 0.835 for main shocks and b = 0.635 for aftershocks. Model foreshocks occur 28 percent of the time.

  7. Plasma particle simulation of electrostatic ion thrusters

    NASA Technical Reports Server (NTRS)

    Peng, Xiaohang; Keefer, Dennis; Ruyten, Wilhelmus

    1990-01-01

    Charge exchange collisons between beam ions and neutral propellant gas can result in erosion of the accelerator grid surfaces of an ion engine. A particle in cell (PIC) is developed along with a Monte Carlo method to simulate the ion dynamics and charge exchange processes in the grid region of an ion thruster. The simulation is two-dimensional axisymmetric and uses three velocity components (2d3v) to investigate the influence of charge exchange collisions on the ion sputtering of the accelerator grid surfaces. An example calculation has been performed for an ion thruster operated on xenon propellant. The simulation shows that the greatest sputtering occurs on the downstream surface of the grid, but some sputtering can also occur on the upstream surface as well as on the interior of the grid aperture.

  8. Conservative zonal schemes for patched grids in 2 and 3 dimensions

    NASA Technical Reports Server (NTRS)

    Hessenius, Kristin A.

    1987-01-01

    The computation of flow over complex geometries, such as realistic aircraft configurations, poses difficult grid generation problems for computational aerodynamicists. The creation of a traditional, single-module grid of acceptable quality about an entire configuration may be impossible even with the most sophisticated of grid generation techniques. A zonal approach, wherein the flow field is partitioned into several regions within which grids are independently generated, is a practical alternative for treating complicated geometries. This technique not only alleviates the problems of discretizing a complex region, but also facilitates a block processing approach to computation thereby circumventing computer memory limitations. The use of such a zonal scheme, however, requires the development of an interfacing procedure that ensures a stable, accurate, and conservative calculation for the transfer of information across the zonal borders.

  9. Solution of the neutronics code dynamic benchmark by finite element method

    NASA Astrophysics Data System (ADS)

    Avvakumov, A. V.; Vabishchevich, P. N.; Vasilev, A. O.; Strizhov, V. F.

    2016-10-01

    The objective is to analyze the dynamic benchmark developed by Atomic Energy Research for the verification of best-estimate neutronics codes. The benchmark scenario includes asymmetrical ejection of a control rod in a water-type hexagonal reactor at hot zero power. A simple Doppler feedback mechanism assuming adiabatic fuel temperature heating is proposed. The finite element method on triangular calculation grids is used to solve the three-dimensional neutron kinetics problem. The software has been developed using the engineering and scientific calculation library FEniCS. The matrix spectral problem is solved using the scalable and flexible toolkit SLEPc. The solution accuracy of the dynamic benchmark is analyzed by condensing calculation grid and varying degree of finite elements.

  10. Twelve Principles for Green Energy Storage in Grid Applications.

    PubMed

    Arbabzadeh, Maryam; Johnson, Jeremiah X; Keoleian, Gregory A; Rasmussen, Paul G; Thompson, Levi T

    2016-01-19

    The introduction of energy storage technologies to the grid could enable greater integration of renewables, improve system resilience and reliability, and offer cost effective alternatives to transmission and distribution upgrades. The integration of energy storage systems into the electrical grid can lead to different environmental outcomes based on the grid application, the existing generation mix, and the demand. Given this complexity, a framework is needed to systematically inform design and technology selection about the environmental impacts that emerge when considering energy storage options to improve sustainability performance of the grid. To achieve this, 12 fundamental principles specific to the design and grid application of energy storage systems are developed to inform policy makers, designers, and operators. The principles are grouped into three categories: (1) system integration for grid applications, (2) the maintenance and operation of energy storage, and (3) the design of energy storage systems. We illustrate the application of each principle through examples published in the academic literature, illustrative calculations, and a case study with an off-grid application of vanadium redox flow batteries (VRFBs). In addition, trade-offs that can emerge between principles are highlighted.

  11. ITERATIVE SCATTER CORRECTION FOR GRID-LESS BEDSIDE CHEST RADIOGRAPHY: PERFORMANCE FOR A CHEST PHANTOM.

    PubMed

    Mentrup, Detlef; Jockel, Sascha; Menser, Bernd; Neitzel, Ulrich

    2016-06-01

    The aim of this work was to experimentally compare the contrast improvement factors (CIFs) of a newly developed software-based scatter correction to the CIFs achieved by an antiscatter grid. To this end, three aluminium discs were placed in the lung, the retrocardial and the abdominal areas of a thorax phantom, and digital radiographs of the phantom were acquired both with and without a stationary grid. The contrast generated by the discs was measured in both images, and the CIFs achieved by grid usage were determined for each disc. Additionally, the non-grid images were processed with a scatter correction software. The contrasts generated by the discs were determined in the scatter-corrected images, and the corresponding CIFs were calculated. The CIFs obtained with the grid and with the software were in good agreement. In conclusion, the experiment demonstrates quantitatively that software-based scatter correction allows restoring the image contrast of a non-grid image in a manner comparable with an antiscatter grid. © The Author 2015. Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com.

  12. Gridded sunshine duration climate data record for Germany based on combined satellite and in situ observations

    NASA Astrophysics Data System (ADS)

    Walawender, Jakub; Kothe, Steffen; Trentmann, Jörg; Pfeifroth, Uwe; Cremer, Roswitha

    2017-04-01

    The purpose of this study is to create a 1 km2 gridded daily sunshine duration data record for Germany covering the period from 1983 to 2015 (33 years) based on satellite estimates of direct normalised surface solar radiation and in situ sunshine duration observations using a geostatistical approach. The CM SAF SARAH direct normalized irradiance (DNI) satellite climate data record and in situ observations of sunshine duration from 121 weather stations operated by DWD are used as input datasets. The selected period of 33 years is associated with the availability of satellite data. The number of ground stations is limited to 121 as there are only time series with less than 10% of missing observations over the selected period included to keep the long-term consistency of the output sunshine duration data record. In the first step, DNI data record is used to derive sunshine hours by applying WMO threshold of 120 W/m2 (SDU = DNI ≥ 120 W/m2) and weighting of sunny slots to correct the sunshine length between two instantaneous image data due to cloud movement. In the second step, linear regression between SDU and in situ sunshine duration is calculated to adjust the satellite product to the ground observations and the output regression coefficients are applied to create a regression grid. In the last step regression residuals are interpolated with ordinary kriging and added to the regression grid. A comprehensive accuracy assessment of the gridded sunshine duration data record is performed by calculating prediction errors (cross-validation routine). "R" is used for data processing. A short analysis of the spatial distribution and temporal variability of sunshine duration over Germany based on the created dataset will be presented. The gridded sunshine duration data are useful for applications in various climate-related studies, agriculture and solar energy potential calculations.

  13. Schnek: A C++ library for the development of parallel simulation codes on regular grids

    NASA Astrophysics Data System (ADS)

    Schmitz, Holger

    2018-05-01

    A large number of algorithms across the field of computational physics are formulated on grids with a regular topology. We present Schnek, a library that enables fast development of parallel simulations on regular grids. Schnek contains a number of easy-to-use modules that greatly reduce the amount of administrative code for large-scale simulation codes. The library provides an interface for reading simulation setup files with a hierarchical structure. The structure of the setup file is translated into a hierarchy of simulation modules that the developer can specify. The reader parses and evaluates mathematical expressions and initialises variables or grid data. This enables developers to write modular and flexible simulation codes with minimal effort. Regular grids of arbitrary dimension are defined as well as mechanisms for defining physical domain sizes, grid staggering, and ghost cells on these grids. Ghost cells can be exchanged between neighbouring processes using MPI with a simple interface. The grid data can easily be written into HDF5 files using serial or parallel I/O.

  14. Scheduling multicore workload on shared multipurpose clusters

    NASA Astrophysics Data System (ADS)

    Templon, J. A.; Acosta-Silva, C.; Flix Molina, J.; Forti, A. C.; Pérez-Calero Yzquierdo, A.; Starink, R.

    2015-12-01

    With the advent of workloads containing explicit requests for multiple cores in a single grid job, grid sites faced a new set of challenges in workload scheduling. The most common batch schedulers deployed at HEP computing sites do a poor job at multicore scheduling when using only the native capabilities of those schedulers. This paper describes how efficient multicore scheduling was achieved at the sites the authors represent, by implementing dynamically-sized multicore partitions via a minimalistic addition to the Torque/Maui batch system already in use at those sites. The paper further includes example results from use of the system in production, as well as measurements on the dependence of performance (especially the ramp-up in throughput for multicore jobs) on node size and job size.

  15. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Martínez-Rovira, I., E-mail: immamartinez@gmail.com; Prezado, Y.; Fois, G.

    Purpose: Spatial fractionation of the dose has proven to be a promising approach to increase the tolerance of healthy tissue, which is the main limitation of radiotherapy. A good example of that is GRID therapy, which has been successfully used in the management of large tumors with low toxicity. The aim of this work is to explore new avenues using nonconventional sources: GRID therapy by using kilovoltage (synchrotron) x-rays, the use of very high-energy electrons, and proton GRID therapy. They share in common the use of the smallest possible grid sizes in order to exploit the dose–volume effects. Methods: Montemore » Carlo simulations (PENELOPE/PENEASY and GEANT4/GATE codes) were used as a method to study dose distributions resulting from irradiations in different configurations of the three proposed techniques. As figure of merit, percentage (peak and valley) depth dose curves, penumbras, and central peak-to-valley dose ratios (PVDR) were evaluated. As shown in previous biological experiments, high PVDR values are requested for healthy tissue sparing. A superior tumor control may benefit from a lower PVDR. Results: High PVDR values were obtained in the healthy tissue for the three cases studied. When low energy photons are used, the treatment of deep-seated tumors can still be performed with submillimetric grid sizes. Superior PVDR values were reached with the other two approaches in the first centimeters along the beam path. The use of protons has the advantage of delivering a uniform dose distribution in the tumor, while healthy tissue benefits from the spatial fractionation of the dose. In the three evaluated techniques, there is a net reduction in penumbra with respect to radiosurgery. Conclusions: The high PVDR values in the healthy tissue and the use of small grid sizes in the three presented approaches might constitute a promising alternative to treat tumors with such spatially fractionated radiotherapy techniques. The dosimetric results presented here support the interest of performing radiobiology experiments in order to evaluate these new avenues.« less

  16. Validation of OSLD and a treatment planning system for surface dose determination in IMRT treatments

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhuang, Audrey H., E-mail: hzhuang@usc.edu; Olch, Arthur J.

    2014-08-15

    Purpose: To evaluate the accuracy of skin dose determination for composite multibeam 3D conformal radiation therapy (3DCRT) and intensity modulated radiation therapy (IMRT) treatments using optically stimulated luminescent dosimeters (OSLDs) and Eclipse treatment planning system. Methods: Surface doses measured by OSLDs in the buildup region for open field 6 MV beams, either perpendicular or oblique to the surface, were evaluated by comparing against dose measured by Markus Parallel Plate (PP) chamber, surface diodes, and calculated by Monte Carlo simulations. The accuracy of percent depth dose (PDD) calculation in the buildup region from the authors’ Eclipse system (Version 10), which wasmore » precisely commissioned in the buildup region and was used with 1 mm calculation grid, was also evaluated by comparing to PP chamber measurements and Monte Carlo simulations. Finally, an anthropomorphic pelvic phantom was CT scanned with OSLDs in place at three locations. A planning target volume (PTV) was defined that extended close to the surface. Both an 8 beam 3DCRT and IMRT plan were generated in Eclipse. OSLDs were placed at the CT scanned reference locations to measure the skin doses and were compared to diode measurements and Eclipse calculations. Efforts were made to ensure that the dose comparison was done at the effective measurement points of each detector and corresponding locations in CT images. Results: The depth of the effective measurement point is 0.8 mm for OSLD when used in the buildup region in a 6 MV beam and is 0.7 mm for the authors’ surface diode. OSLDs and Eclipse system both agree well with Monte Carlo and/or Markus PP ion chamber and/or diode in buildup regions in 6 MV beams with normal or oblique incidence and across different field sizes. For the multiple beam 3DCRT plan and IMRT plans, the differences between OSLDs and Eclipse calculations on the surface of the anthropomorphic phantom were within 3% and distance-to-agreement less than 0.3 mm. Conclusions: The authors’ experiment showed that OSLD is an accurate dosimeter for skin dose measurements in complex 3DCRT or IMRT plans. It also showed that an Eclipse system with accurate commissioning of the data in the buildup region and 1 mm calculation grid can calculate surface doses with high accuracy and has a potential to replacein vivo measurements.« less

  17. SU-E-T-278: Realization of Dose Verification Tool for IMRT Plan Based On DPM

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cai, Jinfeng; Cao, Ruifen; Dai, Yumei

    Purpose: To build a Monte Carlo dose verification tool for IMRT Plan by implementing a irradiation source model into DPM code. Extend the ability of DPM to calculate any incident angles and irregular-inhomogeneous fields. Methods: With the virtual source and the energy spectrum which unfolded from the accelerator measurement data,combined with optimized intensity maps to calculate the dose distribution of the irradiation irregular-inhomogeneous field. The irradiation source model of accelerator was substituted by a grid-based surface source. The contour and the intensity distribution of the surface source were optimized by ARTS (Accurate/Advanced Radiotherapy System) optimization module based on the tumormore » configuration. The weight of the emitter was decided by the grid intensity. The direction of the emitter was decided by the combination of the virtual source and the emitter emitting position. The photon energy spectrum unfolded from the accelerator measurement data was adjusted by compensating the contaminated electron source. For verification, measured data and realistic clinical IMRT plan were compared with DPM dose calculation. Results: The regular field was verified by comparing with the measured data. It was illustrated that the differences were acceptable (<2% inside the field, 2–3mm in the penumbra). The dose calculation of irregular field by DPM simulation was also compared with that of FSPB (Finite Size Pencil Beam) and the passing rate of gamma analysis was 95.1% for peripheral lung cancer. The regular field and the irregular rotational field were all within the range of permitting error. The computing time of regular fields were less than 2h, and the test of peripheral lung cancer was 160min. Through parallel processing, the adapted DPM could complete the calculation of IMRT plan within half an hour. Conclusion: The adapted parallelized DPM code with irradiation source model is faster than classic Monte Carlo codes. Its computational accuracy and speed satisfy the clinical requirement, and it is expectable to be a Monte Carlo dose verification tool for IMRT Plan. Strategic Priority Research Program of the China Academy of Science(XDA03040000); National Natural Science Foundation of China (81101132)« less

  18. Glaucophane chloritoid-bearing assemblages from NE Oman: petrologic significance and a petrogenetic grid for high P metapelites

    NASA Astrophysics Data System (ADS)

    El-Shazly, A. K.; Liou, J. G.

    1991-04-01

    Pelitic layers and lenses interbedded with blueschists and eclogites in Saih Hatat, NE Oman contain chloritoid- and sodic amphibole-bearing mineral assemblages that are useful for reconstructing the P-T history of the area. Textural and mineral chemical relations suggest that coexisting glaucophane (Gln) and chloritoid (Ctd) formed at the expense of chlorite (Chl)+paragonite (Pg) and later broke down to garnet (Gt)+Pg during prograde metamorphism according to the reaction: Gln+Ctd+Qz=Gt+Pg+H2O. During retrogression, Gln and Chl first formed at the expense of Gt and Pg, followed by the breakdown of Ctd and Gt to Chl. The final stages of retrogression are marked by the breakdown of Gln to an aggregate of Chl+albite (Ab). A projection from quartz (Qz), H2O and phengite (Ph) on the (Al2O3+Fe2O3)-(FeO+MgO)-Na2O plane in the system NFMASH is best suited for the representation of the phase relations in high P metapelites. Petrogenetic grids for the model systems NMASH and NFASH were calculated using program GEO-CALC (Berman et al. 1987) and its database (Berman 1988) after the retrieval of S{i/o}and ΔH{f/o}for Gln and Ctd by mathematical programming and calculating all possible reactions among Gln, Ctd, Chl, jadeite (Jd), Ab, Gt, Pg, talc (Tc), pyrophyllite (Prl) and kyanite (Ky). The calculated petrogenetic grid for the system NFASH shows that Fe-Ctd and ferroglaucophane coexist at P>6.5 kbar and T<525°C, whereas the assemblage: Ct-Gln-Pg is stable between 435 and 630°C and P>6 kbar. This grid is consistent with the P-T estimates for high P metapelites from Oman, New Caledonia, Seward Peninsula, Ile de Groix, Sifnos and Peloponnese, where Gln+Ctd bearing units are interbedded with cofacial mafic blueschists and eclogites. The grid also explains the observed textural relations in the metapelites of Oman, and is consistent with the “clockwise” P-T path proposed for this area, but differs from the grids of Guiraud et al. (1990) in showing a larger stability field for Gln+Ctd that overlaps with the stability fields of Jd+Qz as well as Ab. The petrogenetic grid calculated for the system NMASH shows that Mg-Ctd+Gln-bearing assemblages require unusually low geothermal gradients to form in metapelites. It also shows that Mg-rich Ctd+Tc coexist at high pressures over a wider P-T range than predicted by Guiraud et al. (1990). This grid can therefore be applied to high P metamorphic assemblages from the eastern and western Alps.

  19. 2.5D complex resistivity modeling and inversion using unstructured grids

    NASA Astrophysics Data System (ADS)

    Xu, Kaijun; Sun, Jie

    2016-04-01

    The characteristic of complex resistivity on rock and ore has been recognized by people for a long time. Generally we have used the Cole-Cole Model(CCM) to describe complex resistivity. It has been proved that the electrical anomaly of geologic body can be quantitative estimated by CCM parameters such as direct resistivity(ρ0), chargeability(m), time constant(τ) and frequency dependence(c). Thus it is very important to obtain the complex parameters of geologic body. It is difficult to approximate complex structures and terrain using traditional rectangular grid. In order to enhance the numerical accuracy and rationality of modeling and inversion, we use an adaptive finite-element algorithm for forward modeling of the frequency-domain 2.5D complex resistivity and implement the conjugate gradient algorithm in the inversion of 2.5D complex resistivity. An adaptive finite element method is applied for solving the 2.5D complex resistivity forward modeling of horizontal electric dipole source. First of all, the CCM is introduced into the Maxwell's equations to calculate the complex resistivity electromagnetic fields. Next, the pseudo delta function is used to distribute electric dipole source. Then the electromagnetic fields can be expressed in terms of the primary fields caused by layered structure and the secondary fields caused by inhomogeneities anomalous conductivity. At last, we calculated the electromagnetic fields response of complex geoelectric structures such as anticline, syncline, fault. The modeling results show that adaptive finite-element methods can automatically improve mesh generation and simulate complex geoelectric models using unstructured grids. The 2.5D complex resistivity invertion is implemented based the conjugate gradient algorithm.The conjugate gradient algorithm doesn't need to compute the sensitivity matrix but directly computes the sensitivity matrix or its transpose multiplying vector. In addition, the inversion target zones are segmented with fine grids and the background zones are segmented with big grid, the method can reduce the grid amounts of inversion, it is very helpful to improve the computational efficiency. The inversion results verify the validity and stability of conjugate gradient inversion algorithm. The results of theoretical calculation indicate that the modeling and inversion of 2.5D complex resistivity using unstructured grids are feasible. Using unstructured grids can improve the accuracy of modeling, but the large number of grids inversion is extremely time-consuming, so the parallel computation for the inversion is necessary. Acknowledgments: We thank to the support of the National Natural Science Foundation of China(41304094).

  20. Accurate pressure gradient calculations in hydrostatic atmospheric models

    NASA Technical Reports Server (NTRS)

    Carroll, John J.; Mendez-Nunez, Luis R.; Tanrikulu, Saffet

    1987-01-01

    A method for the accurate calculation of the horizontal pressure gradient acceleration in hydrostatic atmospheric models is presented which is especially useful in situations where the isothermal surfaces are not parallel to the vertical coordinate surfaces. The present method is shown to be exact if the potential temperature lapse rate is constant between the vertical pressure integration limits. The technique is applied to both the integration of the hydrostatic equation and the computation of the slope correction term in the horizontal pressure gradient. A fixed vertical grid and a dynamic grid defined by the significant levels in the vertical temperature distribution are employed.

  1. 3D data processing with advanced computer graphics tools

    NASA Astrophysics Data System (ADS)

    Zhang, Song; Ekstrand, Laura; Grieve, Taylor; Eisenmann, David J.; Chumbley, L. Scott

    2012-09-01

    Often, the 3-D raw data coming from an optical profilometer contains spiky noises and irregular grid, which make it difficult to analyze and difficult to store because of the enormously large size. This paper is to address these two issues for an optical profilometer by substantially reducing the spiky noise of the 3-D raw data from an optical profilometer, and by rapidly re-sampling the raw data into regular grids at any pixel size and any orientation with advanced computer graphics tools. Experimental results will be presented to demonstrate the effectiveness of the proposed approach.

  2. Uncertainty modelling and analysis of volume calculations based on a regular grid digital elevation model (DEM)

    NASA Astrophysics Data System (ADS)

    Li, Chang; Wang, Qing; Shi, Wenzhong; Zhao, Sisi

    2018-05-01

    The accuracy of earthwork calculations that compute terrain volume is critical to digital terrain analysis (DTA). The uncertainties in volume calculations (VCs) based on a DEM are primarily related to three factors: 1) model error (ME), which is caused by an adopted algorithm for a VC model, 2) discrete error (DE), which is usually caused by DEM resolution and terrain complexity, and 3) propagation error (PE), which is caused by the variables' error. Based on these factors, the uncertainty modelling and analysis of VCs based on a regular grid DEM are investigated in this paper. Especially, how to quantify the uncertainty of VCs is proposed by a confidence interval based on truncation error (TE). In the experiments, the trapezoidal double rule (TDR) and Simpson's double rule (SDR) were used to calculate volume, where the TE is the major ME, and six simulated regular grid DEMs with different terrain complexity and resolution (i.e. DE) were generated by a Gauss synthetic surface to easily obtain the theoretical true value and eliminate the interference of data errors. For PE, Monte-Carlo simulation techniques and spatial autocorrelation were used to represent DEM uncertainty. This study can enrich uncertainty modelling and analysis-related theories of geographic information science.

  3. Progress in high-lift aerodynamic calculations

    NASA Technical Reports Server (NTRS)

    Rogers, Stuart E.

    1993-01-01

    The current work presents progress in the effort to numerically simulate the flow over high-lift aerodynamic components, namely, multi-element airfoils and wings in either a take-off or a landing configuration. The computational approach utilizes an incompressible flow solver and an overlaid chimera grid approach. A detailed grid resolution study is presented for flow over a three-element airfoil. Two turbulence models, a one-equation Baldwin-Barth model and a two equation k-omega model are compared. Excellent agreement with experiment is obtained for the lift coefficient at all angles of attack, including the prediction of maximum lift when using the two-equation model. Results for two other flap riggings are shown. Three-dimensional results are presented for a wing with a square wing-tip as a validation case. Grid generation and topology is discussed for computing the flow over a T-39 Sabreliner wing with flap deployed and the initial calculations for this geometry are presented.

  4. Assessing species distribution using Google Street View: a pilot study with the Pine Processionary Moth.

    PubMed

    Rousselet, Jérôme; Imbert, Charles-Edouard; Dekri, Anissa; Garcia, Jacques; Goussard, Francis; Vincent, Bruno; Denux, Olivier; Robinet, Christelle; Dorkeld, Franck; Roques, Alain; Rossi, Jean-Pierre

    2013-01-01

    Mapping species spatial distribution using spatial inference and prediction requires a lot of data. Occurrence data are generally not easily available from the literature and are very time-consuming to collect in the field. For that reason, we designed a survey to explore to which extent large-scale databases such as Google maps and Google Street View could be used to derive valid occurrence data. We worked with the Pine Processionary Moth (PPM) Thaumetopoea pityocampa because the larvae of that moth build silk nests that are easily visible. The presence of the species at one location can therefore be inferred from visual records derived from the panoramic views available from Google Street View. We designed a standardized procedure allowing evaluating the presence of the PPM on a sampling grid covering the landscape under study. The outputs were compared to field data. We investigated two landscapes using grids of different extent and mesh size. Data derived from Google Street View were highly similar to field data in the large-scale analysis based on a square grid with a mesh of 16 km (96% of matching records). Using a 2 km mesh size led to a strong divergence between field and Google-derived data (46% of matching records). We conclude that Google database might provide useful occurrence data for mapping the distribution of species which presence can be visually evaluated such as the PPM. However, the accuracy of the output strongly depends on the spatial scales considered and on the sampling grid used. Other factors such as the coverage of Google Street View network with regards to sampling grid size and the spatial distribution of host trees with regards to road network may also be determinant.

  5. Assessing Species Distribution Using Google Street View: A Pilot Study with the Pine Processionary Moth

    PubMed Central

    Dekri, Anissa; Garcia, Jacques; Goussard, Francis; Vincent, Bruno; Denux, Olivier; Robinet, Christelle; Dorkeld, Franck; Roques, Alain; Rossi, Jean-Pierre

    2013-01-01

    Mapping species spatial distribution using spatial inference and prediction requires a lot of data. Occurrence data are generally not easily available from the literature and are very time-consuming to collect in the field. For that reason, we designed a survey to explore to which extent large-scale databases such as Google maps and Google street view could be used to derive valid occurrence data. We worked with the Pine Processionary Moth (PPM) Thaumetopoea pityocampa because the larvae of that moth build silk nests that are easily visible. The presence of the species at one location can therefore be inferred from visual records derived from the panoramic views available from Google street view. We designed a standardized procedure allowing evaluating the presence of the PPM on a sampling grid covering the landscape under study. The outputs were compared to field data. We investigated two landscapes using grids of different extent and mesh size. Data derived from Google street view were highly similar to field data in the large-scale analysis based on a square grid with a mesh of 16 km (96% of matching records). Using a 2 km mesh size led to a strong divergence between field and Google-derived data (46% of matching records). We conclude that Google database might provide useful occurrence data for mapping the distribution of species which presence can be visually evaluated such as the PPM. However, the accuracy of the output strongly depends on the spatial scales considered and on the sampling grid used. Other factors such as the coverage of Google street view network with regards to sampling grid size and the spatial distribution of host trees with regards to road network may also be determinant. PMID:24130675

  6. Solving Graph Laplacian Systems Through Recursive Bisections and Two-Grid Preconditioning

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ponce, Colin; Vassilevski, Panayot S.

    2016-02-18

    We present a parallelizable direct method for computing the solution to graph Laplacian-based linear systems derived from graphs that can be hierarchically bipartitioned with small edge cuts. For a graph of size n with constant-size edge cuts, our method decomposes a graph Laplacian in time O(n log n), and then uses that decomposition to perform a linear solve in time O(n log n). We then use the developed technique to design a preconditioner for graph Laplacians that do not have this property. Finally, we augment this preconditioner with a two-grid method that accounts for much of the preconditioner's weaknesses. Wemore » present an analysis of this method, as well as a general theorem for the condition number of a general class of two-grid support graph-based preconditioners. Numerical experiments illustrate the performance of the studied methods.« less

  7. Internal Passage Heat Transfer Prediction Using Multiblock Grids and a Kappa-Omega Turbulence Model

    NASA Technical Reports Server (NTRS)

    Rigby, David L.; Ameri, Ali A.; Steinthorsson, Erlendur

    1996-01-01

    Numerical simulations of the three-dimensional flow and heat transfer in a rectangular duct with a 180 C bend were performed. Results are presented for Reynolds numbers of 17,000 and 37,000 and for aspect ratios of 0.5 and I.O. A kappa-omega turbulence model with no reference to distance to a wall is used. Direct comparison between single block and multiblock grid calculations are made. Heat transfer and velocity distributions are compared to available literature with good agreement. The multi-block grid system is seen to produce more accurate results compared to a single-block grid with the same number of cells.

  8. Vacuum insulation of the high energy negative ion source for fusion application.

    PubMed

    Kojima, A; Hanada, M; Hilmi, A; Inoue, T; Watanabe, K; Taniguchi, M; Kashiwagi, M; Umeda, N; Tobari, H; Kobayashi, S; Yamano, Y; Grisham, L R

    2012-02-01

    Vacuum insulation on a large size negative ion accelerator with multiple extraction apertures and acceleration grids for fusion application was experimentally examined and designed. In the experiment, vacuum insulation characteristics were investigated in the JT-60 negative ion source with >1000 apertures on the grid with the surface area of ∼2 m(2). The sustainable voltages varied with a square root of the gap lengths between the grids, and decreased with number of the apertures and with the surface area of the grids. Based on the obtained results, the JT-60SA (super advanced) negative ion source is designed to produce 22 A, 500 keV D(-) ion beams for 100 s.

  9. An Advanced User Interface Approach for Complex Parameter Study Process Specification in the Information Power Grid

    NASA Technical Reports Server (NTRS)

    Yarrow, Maurice; McCann, Karen M.; Biswas, Rupak; VanderWijngaart, Rob; Yan, Jerry C. (Technical Monitor)

    2000-01-01

    The creation of parameter study suites has recently become a more challenging problem as the parameter studies have now become multi-tiered and the computational environment has become a supercomputer grid. The parameter spaces are vast, the individual problem sizes are getting larger, and researchers are now seeking to combine several successive stages of parameterization and computation. Simultaneously, grid-based computing offers great resource opportunity but at the expense of great difficulty of use. We present an approach to this problem which stresses intuitive visual design tools for parameter study creation and complex process specification, and also offers programming-free access to grid-based supercomputer resources and process automation.

  10. Franck-Condon Factors for Diatomics: Insights and Analysis Using the Fourier Grid Hamiltonian Method

    ERIC Educational Resources Information Center

    Ghosh, Supriya; Dixit, Mayank Kumar; Bhattacharyya, S. P.; Tembe, B. L.

    2013-01-01

    Franck-Condon factors (FCFs) play a crucial role in determining the intensities of the vibrational bands in electronic transitions. In this article, a relatively simple method to calculate the FCFs is illustrated. An algorithm for the Fourier Grid Hamiltonian (FGH) method for computing the vibrational wave functions and the corresponding energy…

  11. Grid-scale Indirect Radiative Forcing of Climate due to aerosols over the northern hemisphere simulated by the integrated WRF-CMAQ model: Preliminary results

    EPA Science Inventory

    In this study, indirect aerosol effects on grid-scale clouds were implemented in the integrated WRF3.3-CMAQ5.0 modeling system by including parameterizations for both cloud droplet and ice number concentrations calculated from the CMAQ-predicted aerosol particles. The resulting c...

  12. REopt Lite Web Tool

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    NREL developed a free, publicly available web version of the REopt (TM) renewable energy integration and optimization platform called REopt Lite. REopt Lite recommends the optimal size and dispatch strategy for grid-connected photovoltaics (PV) and battery storage at a site. It also allows users to explore how PV and storage can increase a site's resiliency during a grid outage.

  13. High Maneuverability Airframe: Investigation of Fin and Canard Sizing for Optimum Maneuverability

    DTIC Science & Technology

    2014-09-01

    overset grids (unified- grid); 5) total variation diminishing discretization based on a new multidimensional interpolation framework; 6) Riemann solvers to...Aerodynamics .........................................................................................3 3.1.1 Solver ...describes the methodology used for the simulations. 3.1.1 Solver The double-precision solver of a commercially available code, CFD ++ v12.1.1, 9

  14. Navigating aerial transects with a laptop computer

    USGS Publications Warehouse

    Anthony, R. Michael; Stehn, R.A.

    1994-01-01

    SUMMARY: A comparison is made of different methods of determining size of home range from grid trapping data. Studies of artificial populations show that a boundary strip method of measuring area and an adjusted range length give sizes closer to the true range than do minimum area or observed range length methods. In simulated trapping of artificial populations, the known range size increases with successive captures until a level is reached that approximates the true range. The same general pattern is followed whether traps are visited at random or traps nearer the center of the range are favored; but when central traps are favored the curve levels more slowly. Range size is revealed with fewer captures when traps are far apart than when they are close together. The curve levels more slowly for oblong ranges than for circular ranges of the same area. Fewer captures are required to determine range length than to determine range area. Other examples of simulated trapping in artificial populations are used to provide measurements of distances from the center of activity and distances between successive captures. These are compared with similar measurements taken from Peromyscus trapping data. The similarity of range sizes found in certain field comparisons of area trapping, colored scat collections, and trailing is cited. A comparison of home range data obtained by area trapping and nest box studies is discussed. It is shown that when traps are set too far apart to include two or more in the range of each animal, calculation of average range size gives biased results. The smaller ranges are not expressed and cannot be included in the averages. The result is that range estimates are smaller at closer spacings and greater at wider spacings, purely as a result of these erroneous calculations and not reflecting any varying behavior of the animals. The problem of variation in apparent home range with variation in trap spacing is considered further by trapping in an artificial population. It is found that trap spacing can alter the apparent size of range even when biological factors are excluded and trap visiting is random. The desirability of excluding travels outside the normal range from home range calculations is discussed. Effects of varying the trapping plan by setting alternate rows of traps, or setting two traps per site, are discussed briefly.

  15. A comparison of certain methods of measuring ranges of small mammals

    USGS Publications Warehouse

    Stickel, L.F.

    1954-01-01

    SUMMARY: A comparison is made of different methods of determining size of home range from grid trapping data. Studies of artificial populations show that a boundary strip method of measuring area and an adjusted range length give sizes closer to the true range than do minimum area or observed range length methods. In simulated trapping of artificial populations, the known range size increases with successive captures until a level is reached that approximates the true range. The same general pattern is followed whether traps are visited at random or traps nearer the center of the range are favored; but when central traps are favored the curve levels more slowly. Range size is revealed with fewer captures when traps are far apart than when they are close together. The curve levels more slowly for oblong ranges than for circular ranges of the same area. Fewer captures are required to determine range length than to determine range area. Other examples of simulated trapping in artificial populations are used to provide measurements of distances from the center of activity and distances between successive captures. These are compared with similar measurements taken from Peromyscus trapping data. The similarity of range sizes found in certain field comparisons of area trapping, colored scat collections, and trailing is cited. A comparison of home range data obtained by area trapping and nest box studies is discussed. It is shown that when traps are set too far apart to include two or more in the range of each animal, calculation of average range size gives biased results. The smaller ranges are not expressed and cannot be included in the averages. The result is that range estimates are smaller at closer spacings and greater at wider spacings, purely as a result of these erroneous calculations and not reflecting any varying behavior of the animals. The problem of variation in apparent home range with variation in trap spacing is considered further by trapping in an artificial population. It is found that trap spacing can alter the apparent size of range even when biological factors are excluded and trap visiting is random. The desirability of excluding travels outside the normal range from home range calculations is discussed. Effects of varying the trapping plan by setting alternate rows of traps, or setting two traps per site, are discussed briefly.

  16. Homogeneity and EPR metrics for assessment of regular grids used in CW EPR powder simulations.

    PubMed

    Crăciun, Cora

    2014-08-01

    CW EPR powder spectra may be approximated numerically using a spherical grid and a Voronoi tessellation-based cubature. For a given spin system, the quality of simulated EPR spectra depends on the grid type, size, and orientation in the molecular frame. In previous work, the grids used in CW EPR powder simulations have been compared mainly from geometric perspective. However, some grids with similar homogeneity degree generate different quality simulated spectra. This paper evaluates the grids from EPR perspective, by defining two metrics depending on the spin system characteristics and the grid Voronoi tessellation. The first metric determines if the grid points are EPR-centred in their Voronoi cells, based on the resonance magnetic field variations inside these cells. The second metric verifies if the adjacent Voronoi cells of the tessellation are EPR-overlapping, by computing the common range of their resonance magnetic field intervals. Beside a series of well known regular grids, the paper investigates a modified ZCW grid and a Fibonacci spherical code, which are new in the context of EPR simulations. For the investigated grids, the EPR metrics bring more information than the homogeneity quantities and are better related to the grids' EPR behaviour, for different spin system symmetries. The metrics' efficiency and limits are finally verified for grids generated from the initial ones, by using the original or magnetic field-constraint variants of the Spherical Centroidal Voronoi Tessellation method. Copyright © 2014 Elsevier Inc. All rights reserved.

  17. Aerodynamics and Aeroacoustics of Rotorcraft (l’ Aerodynamique et l’ aeroacoustique des aeronefs a voilure tournante).

    DTIC Science & Technology

    1995-08-01

    R.T.N. Chen: A survey of nonuniform 22) R.Houwink, A.E.P.Veldman: steady and inflow models for rotorcraft flight unsteady separated flow computations for...grid with con- see [17]). Because of the cylindrical nature of the stant grid sizes. If an arbitrary nonuniform grid is flow of a hovering rotor an O-H...research distributed around the blade section (figure 4) within a lairing at DRA Bedford on the DRA’s Aeromechanics Lynx Control which extends from 80

  18. Marine Physics: Internal-Surface Wave Interaction and Microstructure Measurement Program

    DTIC Science & Technology

    1974-12-31

    Stabilized Free-Fall Vehicles" 2. "On the Decay of Grid Generated Turbulence in Stratified Salt Water" Figure 1 Figure 2 Figure 3 Figure 4 Page 6 6...Ju.IJic.l modelling shows this vehicle to be stable ^h^iting tilts of less than 10-2 radians under fall into VTra e u Lving at 20 cm/sec. For...Fim^vm scaled according to an overall Froude number U/LN, scaling the vertical wake width, where U is the grid speed, L the mesh size of the grid

  19. Development of a Hybrid RANS/LES Method for Compressible Mixing Layer Simulations

    NASA Technical Reports Server (NTRS)

    Georgiadis, Nicholas J.; Alexander, J. Iwan D.; Reshotko, Eli

    2001-01-01

    A hybrid method has been developed for simulations of compressible turbulent mixing layers. Such mixing layers dominate the flows in exhaust systems of modem day aircraft and also those of hypersonic vehicles currently under development. The hybrid method uses a Reynolds-averaged Navier-Stokes (RANS) procedure to calculate wall bounded regions entering a mixing section, and a Large Eddy Simulation (LES) procedure to calculate the mixing dominated regions. A numerical technique was developed to enable the use of the hybrid RANS/LES method on stretched, non-Cartesian grids. The hybrid RANS/LES method is applied to a benchmark compressible mixing layer experiment. Preliminary two-dimensional calculations are used to investigate the effects of axial grid density and boundary conditions. Actual LES calculations, performed in three spatial directions, indicated an initial vortex shedding followed by rapid transition to turbulence, which is in agreement with experimental observations.

  20. An image-based method to measure all-terrain vehicle dimensions for engineering safety purposes.

    PubMed

    Jennissen, Charles A; Miller, Nathan S; Tang, Kaiyang; Denning, Gerene M

    2014-04-01

    All-terrain vehicle (ATV) crashes are a serious public health and safety concern. Engineering approaches that address ATV injury prevention are critically needed. Avenues to pursue include evidence-based seat design that decreases risky behaviours, such as carrying passengers and operation of adult-size vehicles by children. The goal of this study was to create and validate an image-based method to measure ATV seat length and placement. Publicly available ATV images were downloaded. Adobe Photoshop was then used to generate a vertical grid through the centre of the vehicle, to define the grid scale using the manufacturer's reported wheelbase, and to determine seat length and placement relative to the front and rear axles using this scale. Images that yielded a difference greater than 5% between the calculated and the manufacturer's reported ATV lengths were excluded from further analysis. For the 77 images that met inclusion criteria, the mean±SD for the difference in calculated versus reported vehicle length was 1.8%±1.2%. The Pearson correlation coefficient for comparing image-based seat lengths determined by two independent measurers (20 models) and image-based lengths versus lengths measured at dealerships (12 models) were 0.95 and 0.96, respectively. The image-based method provides accurate and reproducible results for determining ATV measurements, including seat length and placement. This method greatly expands the number of ATV models that can be studied, and may be generalisable to other motor vehicle types. These measurements can be used to guide engineering approaches that improve ATV safety design.

  1. The eGo grid model: An open-source and open-data based synthetic medium-voltage grid model for distribution power supply systems

    NASA Astrophysics Data System (ADS)

    Amme, J.; Pleßmann, G.; Bühler, J.; Hülk, L.; Kötter, E.; Schwaegerl, P.

    2018-02-01

    The increasing integration of renewable energy into the electricity supply system creates new challenges for distribution grids. The planning and operation of distribution systems requires appropriate grid models that consider the heterogeneity of existing grids. In this paper, we describe a novel method to generate synthetic medium-voltage (MV) grids, which we applied in our DIstribution Network GeneratOr (DINGO). DINGO is open-source software and uses freely available data. Medium-voltage grid topologies are synthesized based on location and electricity demand in defined demand areas. For this purpose, we use GIS data containing demand areas with high-resolution spatial data on physical properties, land use, energy, and demography. The grid topology is treated as a capacitated vehicle routing problem (CVRP) combined with a local search metaheuristics. We also consider the current planning principles for MV distribution networks, paying special attention to line congestion and voltage limit violations. In the modelling process, we included power flow calculations for validation. The resulting grid model datasets contain 3608 synthetic MV grids in high resolution, covering all of Germany and taking local characteristics into account. We compared the modelled networks with real network data. In terms of number of transformers and total cable length, we conclude that the method presented in this paper generates realistic grids that could be used to implement a cost-optimised electrical energy system.

  2. Wing Shape as an Indicator of Larval Rearing Conditions for Aedes albopictus and Ae. aegypti (Diptera: Culicidae)

    PubMed Central

    Stephens, C. R.; Juliano, S. A.

    2012-01-01

    Estimating a mosquito’s vector competence, or likelihood of transmitting disease, if it takes an infectious blood meal, is an important aspect of predicting when and where outbreaks of infectious diseases will occur. Vector competence can be affected by rearing temperature and inter- and intraspecific competition experienced by the individual mosquito during its larval development. This research investigates whether a new morphological indicator of larval rearing conditions, wing shape, can be used to distinguish reliably temperature and competitive conditions experienced during larval stages. Aedes albopictus and Aedes aegypti larvae were reared in low intra-specific, high intra-specific, or high inter-specific competition treatments at either 22°C or 32°C. The right wing of each dried female was removed and photographed. Nineteen landmarks and twenty semilandmarks were digitized on each wing. Shape variables were calculated using geometric morphometric software. Canonical variate analysis, randomization multivariate analysis of variance, and visualization of landmark movement using deformation grids provided evidence that although semilandmark position was significantly affected by larval competition and temperature for both species, the differences in position did not translate into differences in wing shape, as shown in deformation grids. Two classification procedures yielded success rates of 26–49%. Accounting for wing size produced no increase in classification success. There appeared to be a significant relationship between shape and size. These results, particularly the low success rate of classification based on wing shape, show that shape is unlikely to be a reliable indicator of larval rearing competition and temperature conditions for Aedes albopictus and Aedes aegypti. PMID:22897054

  3. Introducing the Global Fire WEather Database (GFWED)

    NASA Astrophysics Data System (ADS)

    Field, R. D.

    2015-12-01

    The Canadian Fire Weather Index (FWI) System is the mostly widely used fire danger rating system in the world. We have developed a global database of daily FWI System calculations beginning in 1980 called the Global Fire WEather Database (GFWED) gridded to a spatial resolution of 0.5° latitude by 2/3° longitude. Input weather data were obtained from the NASA Modern Era Retrospective-Analysis for Research (MERRA), and two different estimates of daily precipitation from rain gauges over land. FWI System Drought Code calculations from the gridded datasets were compared to calculations from individual weather station data for a representative set of 48 stations in North, Central and South America, Europe, Russia, Southeast Asia and Australia. Agreement between gridded calculations and the station-based calculations tended to be most different at low latitudes for strictly MERRA-based calculations. Strong biases could be seen in either direction: MERRA DC over the Mato Grosso in Brazil reached unrealistically high values exceeding DC=1500 during the dry season but was too low over Southeast Asia during the dry season. These biases are consistent with those previously-identified in MERRA's precipitation and reinforce the need to consider alternative sources of precipitation data. GFWED is being used by researchers around the world for analyzing historical relationships between fire weather and fire activity at large scales, in identifying large-scale atmosphere-ocean controls on fire weather, and calibration of FWI-based fire prediction models. These applications will be discussed. More information on GFWED can be found at http://data.giss.nasa.gov/impacts/gfwed/

  4. Scalability and performance of data-parallel pressure-based multigrid methods for viscous flows

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Blosch, E.L.; Shyy, W.

    1996-05-01

    A full-approximation storage multigrid method for solving the steady-state 2-d incompressible Navier-Stokes equations on staggered grids has been implemented in Fortran on the CM-5, using the array aliasing feature in CM-Fortran to avoid declaring fine-grid-sized arrays on all levels while still allowing a variable number of grid levels. Thus, the storage cost scales with the number of unknowns, allowing us to consider significantly larger problems than would otherwise be possible. Timings over a range of problem sizes and numbers of processors, up to 4096 x 4096 on 512 nodes, show that the smoothing procedure, a pressure-correction technique, is scalable andmore » that the restriction and prolongation steps are nearly so. The performance obtained for the multigrid method is 333 Mflops out of the theoretical peak 4 Gflops on a 32-node CM-5. In comparison, a single-grid computation obtained 420 Mflops. The decrease is due to the inefficiency of the smoothing iterations on the coarse grid levels. W cycles cost much more and are much less efficient than V cycles, due to the increased contribution from the coarse grids. The convergence rate characteristics of the pressure-correction multigrid method are investigated in a Re = 5000 lid-driven cavity flow and a Re = 300 symmetric backward-facing step flow, using either a defect-correction scheme or a second-order upwind scheme. A heuristic technique relating the convergence tolerances for the course grids to the truncation error of the discretization has been found effective and robust. With second-order upwinding on all grid levels, a 5-level 320 x 80 step flow solution was obtained in 20 V cycles, which corresponds to a smoothing rate of 0.7, and required 25 s on a 32-node CM-5. Overall, the convergence rates obtained in the present work are comparable to the most competitive findings reported in the literature. 62 refs., 13 figs.« less

  5. Scalability and Performance of Data-Parallel Pressure-Based Multigrid Methods for Viscous Flows

    NASA Astrophysics Data System (ADS)

    Blosch, Edwin L.; Shyy, Wei

    1996-05-01

    A full-approximation storage multigrid method for solving the steady-state 2-dincompressible Navier-Stokes equations on staggered grids has been implemented in Fortran on the CM-5,using the array aliasing feature in CM-Fortran to avoid declaring fine-grid-sized arrays on all levels while still allowing a variable number of grid levels. Thus, the storage cost scales with the number of unknowns,allowing us to consider significantly larger problems than would otherwise be possible. Timings over a range of problem sizes and numbers of processors, up to 4096 × 4096 on 512 nodes, show that the smoothing procedure, a pressure-correction technique, is scalable and that the restriction and prolongation steps are nearly so. The performance obtained for the multigrid method is 333 Mflops out of the theoretical peak 4 Gflops on a 32-node CM-5. In comparison, a single-grid computation obtained 420 Mflops. The decrease is due to the inefficiency of the smoothing iterations on the coarse grid levels. W cycles cost much more and are much less efficient than V cycles, due to the increased contribution from the coarse grids. The convergence rate characteristics of the pressure-correction multigrid method are investigated in a Re = 5000 lid-driven cavity flow and a Re = 300 symmetric backward-facing step flow, using either a defect-correction scheme or a second-order upwind scheme. A heuristic technique relating the convergence tolerances for the coarse grids to the truncation error of the discretization has been found effective and robust. With second-order upwinding on all grid levels, a 5-level 320× 80 step flow solution was obtained in 20 V cycles, which corresponds to a smoothing rate of 0.7, and required 25 s on a 32-node CM-5. Overall, the convergence rates obtained in the present work are comparable to the most competitive findings reported in the literature.

  6. Regularization techniques on least squares non-uniform fast Fourier transform.

    PubMed

    Gibiino, Fabio; Positano, Vincenzo; Landini, Luigi; Santarelli, Maria Filomena

    2013-05-01

    Non-Cartesian acquisition strategies are widely used in MRI to dramatically reduce the acquisition time while at the same time preserving the image quality. Among non-Cartesian reconstruction methods, the least squares non-uniform fast Fourier transform (LS_NUFFT) is a gridding method based on a local data interpolation kernel that minimizes the worst-case approximation error. The interpolator is chosen using a pseudoinverse matrix. As the size of the interpolation kernel increases, the inversion problem may become ill-conditioned. Regularization methods can be adopted to solve this issue. In this study, we compared three regularization methods applied to LS_NUFFT. We used truncated singular value decomposition (TSVD), Tikhonov regularization and L₁-regularization. Reconstruction performance was evaluated using the direct summation method as reference on both simulated and experimental data. We also evaluated the processing time required to calculate the interpolator. First, we defined the value of the interpolator size after which regularization is needed. Above this value, TSVD obtained the best reconstruction. However, for large interpolator size, the processing time becomes an important constraint, so an appropriate compromise between processing time and reconstruction quality should be adopted. Copyright © 2013 John Wiley & Sons, Ltd.

  7. Calculation of radiation therapy dose using all particle Monte Carlo transport

    DOEpatents

    Chandler, William P.; Hartmann-Siantar, Christine L.; Rathkopf, James A.

    1999-01-01

    The actual radiation dose absorbed in the body is calculated using three-dimensional Monte Carlo transport. Neutrons, protons, deuterons, tritons, helium-3, alpha particles, photons, electrons, and positrons are transported in a completely coupled manner, using this Monte Carlo All-Particle Method (MCAPM). The major elements of the invention include: computer hardware, user description of the patient, description of the radiation source, physical databases, Monte Carlo transport, and output of dose distributions. This facilitated the estimation of dose distributions on a Cartesian grid for neutrons, photons, electrons, positrons, and heavy charged-particles incident on any biological target, with resolutions ranging from microns to centimeters. Calculations can be extended to estimate dose distributions on general-geometry (non-Cartesian) grids for biological and/or non-biological media.

  8. Progress in unstructured-grid methods development for unsteady aerodynamic applications

    NASA Technical Reports Server (NTRS)

    Batina, John T.

    1992-01-01

    The development of unstructured-grid methods for the solution of the equations of fluid flow and what was learned over the course of the research are summarized. The focus of the discussion is on the solution of the time-dependent Euler equations including spatial discretizations, temporal discretizations, and boundary conditions. An example calculation with an implicit upwind method using a CFL number of infinity is presented for the Boeing 747 aircraft. The results were obtained in less than one hour CPU time on a Cray-2 computer, thus, demonstrating the speed and robustness of the capability. Additional calculations for the ONERA M6 wing demonstrate the accuracy of the method through the good agreement between calculated results and experimental data for a standard transonic flow case.

  9. Calculation of radiation therapy dose using all particle Monte Carlo transport

    DOEpatents

    Chandler, W.P.; Hartmann-Siantar, C.L.; Rathkopf, J.A.

    1999-02-09

    The actual radiation dose absorbed in the body is calculated using three-dimensional Monte Carlo transport. Neutrons, protons, deuterons, tritons, helium-3, alpha particles, photons, electrons, and positrons are transported in a completely coupled manner, using this Monte Carlo All-Particle Method (MCAPM). The major elements of the invention include: computer hardware, user description of the patient, description of the radiation source, physical databases, Monte Carlo transport, and output of dose distributions. This facilitated the estimation of dose distributions on a Cartesian grid for neutrons, photons, electrons, positrons, and heavy charged-particles incident on any biological target, with resolutions ranging from microns to centimeters. Calculations can be extended to estimate dose distributions on general-geometry (non-Cartesian) grids for biological and/or non-biological media. 57 figs.

  10. Hot wire in low Reynolds number flow

    NASA Technical Reports Server (NTRS)

    Kolb, M. A.; Covert, E. E.

    1984-01-01

    Progress reports were issued on the following experiments: (1) low Reynolds number flow phenomenon of periodic vortex shedding in the wake behind a cylinder as studied by applying the hot wire anemometer technique of flow measurement. The downstream diffusion of these shed vortices was of prime concern. An evaluation of the performance of the hot wire at low Reynolds number is also considered. (2) A brief examination of the back sections of the Wright Brothers wind tunnel circuits were conducted to establish whether or not gross flow deviations were present at corners, or turning vane regions. A calibration of the test sections was done. (3) The attractiveness of using rembedded grids for airfoil calculations modeled by the Euler equations was explored. These calculations were extended to C-type grids and then to Navier-Stokes calculations,

  11. Calculation of three-dimensional (3-D) internal flow by means of the velocity-vorticity formulation on a staggered grid

    NASA Technical Reports Server (NTRS)

    Stremel, Paul M.

    1995-01-01

    A method has been developed to accurately compute the viscous flow in three-dimensional (3-D) enclosures. This method is the 3-D extension of a two-dimensional (2-D) method developed for the calculation of flow over airfoils. The 2-D method has been tested extensively and has been shown to accurately reproduce experimental results. As in the 2-D method, the 3-D method provides for the non-iterative solution of the incompressible Navier-Stokes equations by means of a fully coupled implicit technique. The solution is calculated on a body fitted computational mesh incorporating a staggered grid methodology. In the staggered grid method, the three components of vorticity are defined at the centers of the computational cell sides, while the velocity components are defined as normal vectors at the centers of the computational cell faces. The staggered grid orientation provides for the accurate definition of the vorticity components at the vorticity locations, the divergence of vorticity at the mesh cell nodes and the conservation of mass at the mesh cell centers. The solution is obtained by utilizing a fractional step solution technique in the three coordinate directions. The boundary conditions for the vorticity and velocity are calculated implicitly as part of the solution. The method provides for the non-iterative solution of the flow field and satisfies the conservation of mass and divergence of vorticity to machine zero at each time step. To test the method, the calculation of simple driven cavity flows have been computed. The driven cavity flow is defined as the flow in an enclosure driven by a moving upper plate at the top of the enclosure. To demonstrate the ability of the method to predict the flow in arbitrary cavities, results will he shown for both cubic and curved cavities.

  12. Calculating realistic voltages across the US power grid utilizing measured impedances and magnetic fields

    NASA Astrophysics Data System (ADS)

    Lucas, G.; Love, J. J.; Kelbert, A.; Bedrosian, P.; Rigler, E. J.

    2017-12-01

    Space weather induces significant geoelectric fields within Earth's subsurface that can adversely affect electric power grids. The complex interaction between space weather and the solid Earth has traditionally been approached with the use of simple 1-D impedance functions relating the inducing magnetic field to the induced geoelectric field. Ongoing data collection through the NSF EarthScope program has produced measured impedance data across much of the continental US. In this work, impedance data are convolved with magnetic field variations, obtained from USGS magnetic observatories, during a geomagnetic storm. This convolution produces geoelectric fields within the earth. These geoelectric fields are then integrated across power transmission lines to determine the voltage generated within each power line as a function of time during a geomagnetic storm. The voltages generated within the electric power grid will be shown for several historic geomagnetic storms. The estimated voltages calculated from 1-D and 3-D impedances differ by more than 100 V across some transmission lines. In combination with grounding resistance data and network topology, these voltage estimates can be utilized by power companies to estimate geomagnetically-induced currents throughout the network. These voltage estimates can provide information on which power lines are most vulnerable to geomagnetic storms, and assist power grid companies investigating where to install additional protections within their grid.

  13. A Variable Resolution Atmospheric General Circulation Model for a Megasite at the North Slope of Alaska

    NASA Astrophysics Data System (ADS)

    Dennis, L.; Roesler, E. L.; Guba, O.; Hillman, B. R.; McChesney, M.

    2016-12-01

    The Atmospheric Radiation Measurement (ARM) climate research facility has three siteslocated on the North Slope of Alaska (NSA): Barrrow, Oliktok, and Atqasuk. These sites, incombination with one other at Toolik Lake, have the potential to become a "megasite" whichwould combine observational data and high resolution modeling to produce high resolutiondata products for the climate community. Such a data product requires high resolutionmodeling over the area of the megasite. We present three variable resolution atmosphericgeneral circulation model (AGCM) configurations as potential alternatives to stand-alonehigh-resolution regional models. Each configuration is based on a global cubed-sphere gridwith effective resolution of 1 degree, with a refinement in resolution down to 1/8 degree overan area surrounding the ARM megasite. The three grids vary in the size of the refined areawith 13k, 9k, and 7k elements. SquadGen, NCL, and GIMP are used to create the grids.Grids vary based upon the selection of areas of refinement which capture climate andweather processes that may affect a proposed NSA megasite. A smaller area of highresolution may not fully resolve climate and weather processes before they reach the NSA,however grids with smaller areas of refinement have a significantly reduced computationalcost compared with grids with larger areas of refinement. Optimal size and shape of thearea of refinement for a variable resolution model at the NSA is investigated.

  14. High performance GPU processing for inversion using uniform grid searches

    NASA Astrophysics Data System (ADS)

    Venetis, Ioannis E.; Saltogianni, Vasso; Stiros, Stathis; Gallopoulos, Efstratios

    2017-04-01

    Many geophysical problems are described by systems of redundant, highly non-linear systems of ordinary equations with constant terms deriving from measurements and hence representing stochastic variables. Solution (inversion) of such problems is based on numerical, optimization methods, based on Monte Carlo sampling or on exhaustive searches in cases of two or even three "free" unknown variables. Recently the TOPological INVersion (TOPINV) algorithm, a grid search-based technique in the Rn space, has been proposed. TOPINV is not based on the minimization of a certain cost function and involves only forward computations, hence avoiding computational errors. The basic concept is to transform observation equations into inequalities on the basis of an optimization parameter k and of their standard errors, and through repeated "scans" of n-dimensional search grids for decreasing values of k to identify the optimal clusters of gridpoints which satisfy observation inequalities and by definition contain the "true" solution. Stochastic optimal solutions and their variance-covariance matrices are then computed as first and second statistical moments. Such exhaustive uniform searches produce an excessive computational load and are extremely time consuming for common computers based on a CPU. An alternative is to use a computing platform based on a GPU, which nowadays is affordable to the research community, which provides a much higher computing performance. Using the CUDA programming language to implement TOPINV allows the investigation of the attained speedup in execution time on such a high performance platform. Based on synthetic data we compared the execution time required for two typical geophysical problems, modeling magma sources and seismic faults, described with up to 18 unknown variables, on both CPU/FORTRAN and GPU/CUDA platforms. The same problems for several different sizes of search grids (up to 1012 gridpoints) and numbers of unknown variables were solved on both platforms, and execution time as a function of the grid dimension for each problem was recorded. Results indicate an average speedup in calculations by a factor of 100 on the GPU platform; for example problems with 1012 grid-points require less than two hours instead of several days on conventional desktop computers. Such a speedup encourages the application of TOPINV on high performance platforms, as a GPU, in cases where nearly real time decisions are necessary, for example finite fault modeling to identify possible tsunami sources.

  15. Rapid inundation estimates at harbor scale using tsunami wave heights offshore simulation and coastal amplification laws

    NASA Astrophysics Data System (ADS)

    Gailler, A.; Loevenbruck, A.; Hebert, H.

    2013-12-01

    Numerical tsunami propagation and inundation models are well developed and have now reached an impressive level of accuracy, especially in locations such as harbors where the tsunami waves are mostly amplified. In the framework of tsunami warning under real-time operational conditions, the main obstacle for the routine use of such numerical simulations remains the slowness of the numerical computation, which is strengthened when detailed grids are required for the precise modeling of the coastline response of an individual harbor. Thus only tsunami offshore propagation modeling tools using a single sparse bathymetric computation grid are presently included within the French Tsunami Warning Center (CENALT), providing rapid estimation of tsunami warning at western Mediterranean and NE Atlantic basins scale. We present here a preliminary work that performs quick estimates of the inundation at individual harbors from these high sea forecasting tsunami simulations. The method involves an empirical correction based on theoretical amplification laws (either Green's or Synolakis laws). The main limitation is that its application to a given coastal area would require a large database of previous observations, in order to define the empirical parameters of the correction equation. As no such data (i.e., historical tide gage records of significant tsunamis) are available for the western Mediterranean and NE Atlantic basins, we use a set of synthetic mareograms, calculated for both fake and well-known historical tsunamigenic earthquakes in the area. This synthetic dataset is obtained through accurate numerical tsunami propagation and inundation modeling by using several nested bathymetric grids of increasingly fine resolution close to the shores (down to a grid cell size of 3m in some Mediterranean harbors). Non linear shallow water tsunami modeling performed on a single 2' coarse bathymetric grid are compared to the values given by time-consuming nested grids simulations (and observation when available), in order to check to which extent the simple approach based on the amplification laws can explain the data. The idea is to fit tsunami data with numerical modeling carried out without any refined coastal bathymetry/topography. To this end several parameters are discussed, namely the bathymetric depth to which model results must be extrapolated (using the Green's law), or the mean bathymetric slope to consider near the studied coast (when using the Synolakis law).

  16. Rapid inundation estimates using coastal amplification laws in the western Mediterranean basin

    NASA Astrophysics Data System (ADS)

    Gailler, Audrey; Loevenbruck, Anne; Hébert, Hélène

    2014-05-01

    Numerical tsunami propagation and inundation models are well developed and have now reached an impressive level of accuracy, especially in locations such as harbors where the tsunami waves are mostly amplified. In the framework of tsunami warning under real-time operational conditions, the main obstacle for the routine use of such numerical simulations remains the slowness of the numerical computation, which is strengthened when detailed grids are required for the precise modeling of the coastline response of an individual harbor. Thus only tsunami offshore propagation modeling tools using a single sparse bathymetric computation grid are presently included within the French Tsunami Warning Center (CENALT), providing rapid estimation of tsunami warning at western Mediterranean and NE Atlantic basins scale. We present here a preliminary work that performs quick estimates of the inundation at individual harbors from these high sea forecasting tsunami simulations. The method involves an empirical correction based on theoretical amplification laws (either Green's or Synolakis laws). The main limitation is that its application to a given coastal area would require a large database of previous observations, in order to define the empirical parameters of the correction equation. As no such data (i.e., historical tide gage records of significant tsunamis) are available for the western Mediterranean and NE Atlantic basins, we use a set of synthetic mareograms, calculated for both fake events and well-known historical tsunamigenic earthquakes in the area. This synthetic dataset is obtained through accurate numerical tsunami propagation and inundation modeling by using several nested bathymetric grids of increasingly fine resolution close to the shores (down to a grid cell size of 3m in some Mediterranean harbors). Non linear shallow water tsunami modeling performed on a single 2' coarse bathymetric grid are compared to the values given by time-consuming nested grids simulations (and observation when available), in order to check to which extent the simple approach based on the amplification laws can explain the data. The idea is to fit tsunami data with numerical modeling carried out without any refined coastal bathymetry/topography. To this end several parameters are discussed, namely the bathymetric depth to which model results must be extrapolated (using the Green's law), or the mean bathymetric slope to consider near the studied coast (when using the Synolakis law).

  17. Fast generation of three-dimensional computational boundary-conforming periodic grids of C-type. [for turbine blades and propellers

    NASA Technical Reports Server (NTRS)

    Dulikravich, D. S.

    1982-01-01

    A fast computer program, GRID3C, was developed to generate multilevel three dimensional, C type, periodic, boundary conforming grids for the calculation of realistic turbomachinery and propeller flow fields. The technique is based on two analytic functions that conformally map a cascade of semi-infinite slits to a cascade of doubly infinite strips on different Riemann sheets. Up to four consecutively refined three dimensional grids are automatically generated and permanently stored on four different computer tapes. Grid nonorthogonality is introduced by a separate coordinate shearing and stretching performed in each of three coordinate directions. The grids are easily clustered closer to the blade surface, the trailing and leading edges and the hub or shroud regions by changing appropriate input parameters. Hub and duct (or outer free boundary) have different axisymmetric shapes. A vortex sheet of arbitrary thickness emanating smoothly from the blade trailing edge is generated automatically by GRID3C. Blade cross sectional shape, chord length, twist angle, sweep angle, and dihedral angle can vary in an arbitrary smooth fashion in the spanwise direction.

  18. AVQS: Attack Route-Based Vulnerability Quantification Scheme for Smart Grid

    PubMed Central

    Lim, Hyunwoo; Lee, Seokjun; Shon, Taeshik

    2014-01-01

    A smart grid is a large, consolidated electrical grid system that includes heterogeneous networks and systems. Based on the data, a smart grid system has a potential security threat in its network connectivity. To solve this problem, we develop and apply a novel scheme to measure the vulnerability in a smart grid domain. Vulnerability quantification can be the first step in security analysis because it can help prioritize the security problems. However, existing vulnerability quantification schemes are not suitable for smart grid because they do not consider network vulnerabilities. We propose a novel attack route-based vulnerability quantification scheme using a network vulnerability score and an end-to-end security score, depending on the specific smart grid network environment to calculate the vulnerability score for a particular attack route. To evaluate the proposed approach, we derive several attack scenarios from the advanced metering infrastructure domain. The experimental results of the proposed approach and the existing common vulnerability scoring system clearly show that we need to consider network connectivity for more optimized vulnerability quantification. PMID:25152923

  19. Adapting the iSNOBAL model for improved visualization in a GIS environment

    NASA Astrophysics Data System (ADS)

    Johansen, W. J.; Delparte, D.

    2014-12-01

    Snowmelt is a primary means of crucial water resources in much of the western United States. Researchers are developing models that estimate snowmelt to aid in water resource management. One such model is the image snowcover energy and mass balance (iSNOBAL) model. It uses input climate grids to simulate the development and melting of snowpack in mountainous regions. This study looks at applying this model to the Reynolds Creek Experimental Watershed in southwestern Idaho, utilizing novel approaches incorporating geographic information systems (GIS). To improve visualization of the iSNOBAL model, we have adapted it to run in a GIS environment. This type of environment is suited to both the input grid creation and the visualization of results. The data used for input grid creation can be stored locally or on a web-server. Kriging interpolation embedded within Python scripts are used to create air temperature, soil temperature, humidity, and precipitation grids, while built-in GIS and existing tools are used to create solar radiation and wind grids. Additional Python scripting is then used to perform model calculations. The final product is a user-friendly and accessible version of the iSNOBAL model, including the ability to easily visualize and interact with model results, all within a web- or desktop-based GIS environment. This environment allows for interactive manipulation of model parameters and visualization of the resulting input grids for the model calculations. Future work is moving towards adapting the model further for use in a 3D gaming engine for improved visualization and interaction.

  20. C library for topological study of the electronic charge density.

    PubMed

    Vega, David; Aray, Yosslen; Rodríguez, Jesús

    2012-12-05

    The topological study of the electronic charge density is useful to obtain information about the kinds of bonds (ionic or covalent) and the atom charges on a molecule or crystal. For this study, it is necessary to calculate, at every space point, the electronic density and its electronic density derivatives values up to second order. In this work, a grid-based method for these calculations is described. The library, implemented for three dimensions, is based on a multidimensional Lagrange interpolation in a regular grid; by differentiating the resulting polynomial, the gradient vector, the Hessian matrix and the Laplacian formulas were obtained for every space point. More complex functions such as the Newton-Raphson method (to find the critical points, where the gradient is null) and the Cash-Karp Runge-Kutta method (used to make the gradient paths) were programmed. As in some crystals, the unit cell has angles different from 90°, the described library includes linear transformations to correct the gradient and Hessian when the grid is distorted (inclined). Functions were also developed to handle grid containing files (grd from DMol® program, CUBE from Gaussian® program and CHGCAR from VASP® program). Each one of these files contains the data for a molecular or crystal electronic property (such as charge density, spin density, electrostatic potential, and others) in a three-dimensional (3D) grid. The library can be adapted to make the topological study in any regular 3D grid by modifying the code of these functions. Copyright © 2012 Wiley Periodicals, Inc.

  1. Multigrid finite element method in stress analysis of three-dimensional elastic bodies of heterogeneous structure

    NASA Astrophysics Data System (ADS)

    Matveev, A. D.

    2016-11-01

    To calculate the three-dimensional elastic body of heterogeneous structure under static loading, a method of multigrid finite element is provided, when implemented on the basis of algorithms of finite element method (FEM), using homogeneous and composite threedimensional multigrid finite elements (MFE). Peculiarities and differences of MFE from the currently available finite elements (FE) are to develop composite MFE (without increasing their dimensions), arbitrarily small basic partition of composite solids consisting of single-grid homogeneous FE of the first order can be used, i.e. in fact, to use micro approach in finite element form. These small partitions allow one to take into account in MFE, i.e. in the basic discrete models of composite solids, complex heterogeneous and microscopically inhomogeneous structure, shape, the complex nature of the loading and fixation and describe arbitrarily closely the stress and stain state by the equations of three-dimensional elastic theory without any additional simplifying hypotheses. When building the m grid FE, m of nested grids is used. The fine grid is generated by a basic partition of MFE, the other m —1 large grids are applied to reduce MFE dimensionality, when m is increased, MFE dimensionality becomes smaller. The procedures of developing MFE of rectangular parallelepiped, irregular shape, plate and beam types are given. MFE generate the small dimensional discrete models and numerical solutions with a high accuracy. An example of calculating the laminated plate, using three-dimensional 3-grid FE and the reference discrete model is given, with that having 2.2 milliards of FEM nodal unknowns.

  2. An Intelligent Approach to Strengthening of the Rural Electrical Power Supply Using Renewable Energy Resources

    NASA Astrophysics Data System (ADS)

    Robert, F. C.; Sisodia, G. S.; Gopalan, S.

    2017-08-01

    The healthy growth of economy lies in the balance between rural and urban development. Several developing countries have achieved a successful growth of urban areas, yet rural infrastructure has been neglected until recently. The rural electrical grids are weak with heavy losses and low capacity. Renewable energy represents an efficient way to generate electricity locally. However, the renewable energy generation may be limited by the low grid capacity. The current solutions focus on grid reinforcement only. This article presents a model for improving renewable energy integration in rural grids with the intelligent combination of three strategies: 1) grid reinforcement, 2) use of storage and 3) renewable energy curtailments. Such approach provides a solution to integrate a maximum of renewable energy generation on low capacity grids while minimising project cost and increasing the percentage of utilisation of assets. The test cases show that a grid connection agreement and a main inverter sized at 60 kW (resp. 80 kW) can accommodate a 100 kWp solar park (resp. 100 kW wind turbine) with minimal storage.

  3. Effect of device design on the aerosolization of a carrier-based dry powder inhaler--a case study on Aerolizer(®) Foradile (®).

    PubMed

    Zhou, Qi Tony; Tong, Zhenbo; Tang, Patricia; Citterio, Mauro; Yang, Runyu; Chan, Hak-Kim

    2013-04-01

    The objective of this study is to investigate the effect of device design of the Aerolizer(®) on the aerosolization of a carrier-based dry powder inhaler formulation (Foradile(®)). The Aerolizer was modified by reducing the air inlet size and mouthpiece length to 1/3 of the original dimensions, or by increasing the grid voidage. Aerosolization of the powder formulation was assessed on a multi-stage liquid impinger at air flow rates of 30, 60, and 100 L/min. Coupled CFD-DEM simulations were performed to investigate the air flow pattern and particle impaction. There was no significant difference in the aerosolization behavior between the original and 1/3 mouthpiece length devices. Significant increases in FPF total and FPF emitted were demonstrated when the inlet size was reduced, and the results were explained by the increases in air velocity and turbulence from the CFD analysis. No significant differences were shown in FPF total and FPF emitted when the grid voidage was increased, but more drugs were found to deposit in induction port and to a lesser extent, the mouthpiece. This was supported by the CFD-DEM analysis which showed the particle-device collisions mainly occurred in the inhaler chamber, and the cross-grid design increased the particle-device collisions on both mouthpiece and induction port. The air inlet size and grid structure of the Aerolizer(®) were found to impact significantly on the aerosolization of the carrier-based powder.

  4. grid-model: Semi-numerical reionization code

    NASA Astrophysics Data System (ADS)

    Hutter, Anne

    2018-05-01

    grid-model computes the time and spatially dependent ionization of neutral hydrogen (HI), neutral (HeI) and singly ionized helium (HeII) in the intergalactic medium (IGM). It accounts for recombinations and provides different descriptions for the photoionization rate that are used to calculate the residual HI fraction in ionized regions. The ionizing emissivity is directly derived from the RT simulation spectra.

  5. Recent Progress on the Parallel Implementation of Moving-Body Overset Grid Schemes

    NASA Technical Reports Server (NTRS)

    Wissink, Andrew; Allen, Edwin (Technical Monitor)

    1998-01-01

    Viscous calculations about geometrically complex bodies in which there is relative motion between component parts is one of the most computationally demanding problems facing CFD researchers today. This presentation documents results from the first two years of a CHSSI-funded effort within the U.S. Army AFDD to develop scalable dynamic overset grid methods for unsteady viscous calculations with moving-body problems. The first pan of the presentation will focus on results from OVERFLOW-D1, a parallelized moving-body overset grid scheme that employs traditional Chimera methodology. The two processes that dominate the cost of such problems are the flow solution on each component and the intergrid connectivity solution. Parallel implementations of the OVERFLOW flow solver and DCF3D connectivity software are coupled with a proposed two-part static-dynamic load balancing scheme and tested on the IBM SP and Cray T3E multi-processors. The second part of the presentation will cover some recent results from OVERFLOW-D2, a new flow solver that employs Cartesian grids with various levels of refinement, facilitating solution adaption. A study of the parallel performance of the scheme on large distributed- memory multiprocessor computer architectures will be reported.

  6. The study on the control strategy of micro grid considering the economy of energy storage operation

    NASA Astrophysics Data System (ADS)

    Ma, Zhiwei; Liu, Yiqun; Wang, Xin; Li, Bei; Zeng, Ming

    2017-08-01

    To optimize the running of micro grid to guarantee the supply and demand balance of electricity, and to promote the utilization of renewable energy. The control strategy of micro grid energy storage system is studied. Firstly, the mixed integer linear programming model is established based on the receding horizon control. Secondly, the modified cuckoo search algorithm is proposed to calculate the model. Finally, a case study is carried out to study the signal characteristic of micro grid and batteries under the optimal control strategy, and the convergence of the modified cuckoo search algorithm is compared with others to verify the validity of the proposed model and method. The results show that, different micro grid running targets can affect the control strategy of energy storage system, which further affect the signal characteristics of the micro grid. Meanwhile, the convergent speed, computing time and the economy of the modified cuckoo search algorithm are improved compared with the traditional cuckoo search algorithm and differential evolution algorithm.

  7. Research and application of thermal power unit’s load dynamic adjustment based on extraction steam

    NASA Astrophysics Data System (ADS)

    Li, Jun; Li, Huicong; Li, Weiwei

    2018-02-01

    The rapid development of heat and power generation in large power plant has caused tremendous constraints on the load adjustment of power grids and power plants. By introducing the thermodynamic system of thermal power unit, the relationship between thermal power extraction steam and unit’s load has analyzed and calculated. The practical application results show that power capability of the unit affected by extraction and it is not conducive to adjust the grid frequency. By monitoring the load adjustment capacity of thermal power units, especially the combined heat and power generating units, the upper and lower limits of the unit load can be dynamically adjusted by the operator on the grid side. The grid regulation and control departments can effectively control the load adjustable intervals of the operating units and provide reliable for the cooperative action of the power grid and power plants, to ensure the safety and stability of the power grid.

  8. A package for 3-D unstructured grid generation, finite-element flow solution and flow field visualization

    NASA Technical Reports Server (NTRS)

    Parikh, Paresh; Pirzadeh, Shahyar; Loehner, Rainald

    1990-01-01

    A set of computer programs for 3-D unstructured grid generation, fluid flow calculations, and flow field visualization was developed. The grid generation program, called VGRID3D, generates grids over complex configurations using the advancing front method. In this method, the point and element generation is accomplished simultaneously, VPLOT3D is an interactive, menudriven pre- and post-processor graphics program for interpolation and display of unstructured grid data. The flow solver, VFLOW3D, is an Euler equation solver based on an explicit, two-step, Taylor-Galerkin algorithm which uses the Flux Corrected Transport (FCT) concept for a wriggle-free solution. Using these programs, increasingly complex 3-D configurations of interest to aerospace community were gridded including a complete Space Transportation System comprised of the space-shuttle orbitor, the solid-rocket boosters, and the external tank. Flow solutions were obtained on various configurations in subsonic, transonic, and supersonic flow regimes.

  9. Brief analysis of Jiangsu grid security and stability based on multi-infeed DC index in power system

    NASA Astrophysics Data System (ADS)

    Zhang, Wenjia; Wang, Quanquan; Ge, Yi; Huang, Junhui; Chen, Zhengfang

    2018-02-01

    The impact of Multi-infeed HVDC has gradually increased to security and stability operating in Jiangsu power grid. In this paper, an appraisal method of Multi-infeed HVDC power grid security and stability is raised with Multi-Infeed Effective Short Circuit Ratio, Multi-Infeed Interaction Factor and Commutation Failure Immunity Index. These indices are adopted in security and stability simulating calculation of Jiangsu Multi-infeed HVDC system. The simulation results indicate that Jiangsu power grid is operating with a strong DC system. It has high level of power grid security and stability, and meet the safety running requirements. Jinpin-Suzhou DC system is located in the receiving end with huge capacity, which is easily leading to commutation failure of the transmission line. In order to resolve this problem, dynamic reactive power compensation can be applied in power grid near Jinpin-Suzhou DC system. Simulation result shows this method is feasible to commutation failure.

  10. A Development of Lightweight Grid Interface

    NASA Astrophysics Data System (ADS)

    Iwai, G.; Kawai, Y.; Sasaki, T.; Watase, Y.

    2011-12-01

    In order to help a rapid development of Grid/Cloud aware applications, we have developed API to abstract the distributed computing infrastructures based on SAGA (A Simple API for Grid Applications). SAGA, which is standardized in the OGF (Open Grid Forum), defines API specifications to access distributed computing infrastructures, such as Grid, Cloud and local computing resources. The Universal Grid API (UGAPI), which is a set of command line interfaces (CLI) and APIs, aims to offer simpler API to combine several SAGA interfaces with richer functionalities. These CLIs of the UGAPI offer typical functionalities required by end users for job management and file access to the different distributed computing infrastructures as well as local computing resources. We have also built a web interface for the particle therapy simulation and demonstrated the large scale calculation using the different infrastructures at the same time. In this paper, we would like to present how the web interface based on UGAPI and SAGA achieve more efficient utilization of computing resources over the different infrastructures with technical details and practical experiences.

  11. Development of comprehensive numerical schemes for predicting evaporating gas-droplets flow processes of a liquid-fueled combustor

    NASA Technical Reports Server (NTRS)

    Chen, C. P.

    1990-01-01

    An existing Computational Fluid Dynamics code for simulating complex turbulent flows inside a liquid rocket combustion chamber was validated and further developed. The Advanced Rocket Injector/Combustor Code (ARICC) is simplified and validated against benchmark flow situations for laminar and turbulent flows. The numerical method used in ARICC Code is re-examined for incompressible flow calculations. For turbulent flows, both the subgrid and the two equation k-epsilon turbulence models are studied. Cases tested include idealized Burger's equation in complex geometries and boundaries, a laminar pipe flow, a high Reynolds number turbulent flow, and a confined coaxial jet with recirculations. The accuracy of the algorithm is examined by comparing the numerical results with the analytical solutions as well as experimented data with different grid sizes.

  12. a Metadata Based Approach for Analyzing Uav Datasets for Photogrammetric Applications

    NASA Astrophysics Data System (ADS)

    Dhanda, A.; Remondino, F.; Santana Quintero, M.

    2018-05-01

    This paper proposes a methodology for pre-processing and analysing Unmanned Aerial Vehicle (UAV) datasets before photogrammetric processing. In cases where images are gathered without a detailed flight plan and at regular acquisition intervals the datasets can be quite large and be time consuming to process. This paper proposes a method to calculate the image overlap and filter out images to reduce large block sizes and speed up photogrammetric processing. The python-based algorithm that implements this methodology leverages the metadata in each image to determine the end and side overlap of grid-based UAV flights. Utilizing user input, the algorithm filters out images that are unneeded for photogrammetric processing. The result is an algorithm that can speed up photogrammetric processing and provide valuable information to the user about the flight path.

  13. An efficient coding theory for a dynamic trajectory predicts non-uniform allocation of entorhinal grid cells to modules.

    PubMed

    Mosheiff, Noga; Agmon, Haggai; Moriel, Avraham; Burak, Yoram

    2017-06-01

    Grid cells in the entorhinal cortex encode the position of an animal in its environment with spatially periodic tuning curves with different periodicities. Recent experiments established that these cells are functionally organized in discrete modules with uniform grid spacing. Here we develop a theory for efficient coding of position, which takes into account the temporal statistics of the animal's motion. The theory predicts a sharp decrease of module population sizes with grid spacing, in agreement with the trend seen in the experimental data. We identify a simple scheme for readout of the grid cell code by neural circuitry, that can match in accuracy the optimal Bayesian decoder. This readout scheme requires persistence over different timescales, depending on the grid cell module. Thus, we propose that the brain may employ an efficient representation of position which takes advantage of the spatiotemporal statistics of the encoded variable, in similarity to the principles that govern early sensory processing.

  14. An efficient coding theory for a dynamic trajectory predicts non-uniform allocation of entorhinal grid cells to modules

    PubMed Central

    Mosheiff, Noga; Agmon, Haggai; Moriel, Avraham

    2017-01-01

    Grid cells in the entorhinal cortex encode the position of an animal in its environment with spatially periodic tuning curves with different periodicities. Recent experiments established that these cells are functionally organized in discrete modules with uniform grid spacing. Here we develop a theory for efficient coding of position, which takes into account the temporal statistics of the animal’s motion. The theory predicts a sharp decrease of module population sizes with grid spacing, in agreement with the trend seen in the experimental data. We identify a simple scheme for readout of the grid cell code by neural circuitry, that can match in accuracy the optimal Bayesian decoder. This readout scheme requires persistence over different timescales, depending on the grid cell module. Thus, we propose that the brain may employ an efficient representation of position which takes advantage of the spatiotemporal statistics of the encoded variable, in similarity to the principles that govern early sensory processing. PMID:28628647

  15. Aerodynamic analysis of three advanced configurations using the TranAir full-potential code

    NASA Technical Reports Server (NTRS)

    Madson, M. D.; Carmichael, R. L.; Mendoza, J. P.

    1989-01-01

    Computational results are presented for three advanced configurations: the F-16A with wing tip missiles and under wing fuel tanks, the Oblique Wing Research Aircraft, and an Advanced Turboprop research model. These results were generated by the latest version of the TranAir full potential code, which solves for transonic flow over complex configurations. TranAir embeds a surface paneled geometry definition in a uniform rectangular flow field grid, thus avoiding the use of surface conforming grids, and decoupling the grid generation process from the definition of the configuration. The new version of the code locally refines the uniform grid near the surface of the geometry, based on local panel size and/or user input. This method distributes the flow field grid points much more efficiently than the previous version of the code, which solved for a grid that was uniform everywhere in the flow field. TranAir results are presented for the three configurations and are compared with wind tunnel data.

  16. Comparison of NTF Experimental Data with CFD Predictions from the Third AIAA CFD Drag Prediction Workshop

    NASA Technical Reports Server (NTRS)

    Vassberg, John C.; Tinoco, Edward N.; Mani, Mori; Levy, David; Zickuhr, Tom; Mavriplis, Dimitri J.; Wahls, Richard A.; Morrison, Joseph H.; Brodersen, Olaf P.; Eisfeld, Bernhard; hide

    2008-01-01

    Recently acquired experimental data for the DLR-F6 wing-body transonic transport con figuration from the National Transonic Facility (NTF) are compared with the database of computational fluid dynamics (CFD) predictions generated for the Third AIAA CFD Drag Prediction Workshop (DPW-III). The NTF data were collected after the DPW-III, which was conducted with blind test cases. These data include both absolute drag levels and increments associated with this wing-body geometry. The baseline DLR-F6 wing-body geometry is also augmented with a side-of-body fairing which eliminates the flow separation in this juncture region. A comparison between computed and experimentally observed sizes of the side-of-body flow-separation bubble is included. The CFD results for the drag polars and separation bubble sizes are computed on grids which represent current engineering best practices for drag predictions. In addition to these data, a more rigorous attempt to predict absolute drag at the design point is provided. Here, a series of three grid densities are utilized to establish an asymptotic trend of computed drag with respect to grid convergence. This trend is then extrapolated to estimate a grid-converged absolute drag level.

  17. Mechanically modulated dewetting by atomic force microscope for micro- and nano- droplet array fabrication.

    PubMed

    Wang, Feifei; Li, Pan; Wang, Dong; Li, Longhai; Xie, Shuangxi; Liu, Lianqing; Wang, Yuechao; Li, Wen Jung

    2014-10-06

    Organizing a material into well-defined patterns during the dewetting process provides an attractive micro-/nano-fabrication method without using a conventional lithographic process, and hence, offers potential applications in organic electronics, optics systems, and memory devices. We report here how the mechanical modification of polymer surface by an Atomic Force Microscope (AFM) can be used to guide thin film dewetting evolution and break the intrinsic spatial correlation of spontaneous instability. An AFM is used to implement the mechanical modification of progressively narrow grids to investigate the influence of pattern size on the modulation of ultrathin polystyrene films dewetting evolution. For films with different initial thicknesses, when grid size is close to or below the characteristic wavelength of instability, the spinodal dewetting is suppressed, and film rupture is restricted to the cutting trench. We will show in this paper it is possible to generate only one droplet per gridded area on a thin film subsequent to nucleation dominated dewetting on a non-patterned substrate. Furthermore, when the grid periodicity exceeds the spinodal length, the number of droplets in predefined areas gradually approaches that associated with unconfined dewetting.

  18. Mechanically Modulated Dewetting by Atomic Force Microscope for Micro- and Nano- Droplet Array Fabrication

    PubMed Central

    Wang, Feifei; Li, Pan; Wang, Dong; Li, Longhai; Xie, Shuangxi; Liu, Lianqing; Wang, Yuechao; Li, Wen Jung

    2014-01-01

    Organizing a material into well-defined patterns during the dewetting process provides an attractive micro-/nano-fabrication method without using a conventional lithographic process, and hence, offers potential applications in organic electronics, optics systems, and memory devices. We report here how the mechanical modification of polymer surface by an Atomic Force Microscope (AFM) can be used to guide thin film dewetting evolution and break the intrinsic spatial correlation of spontaneous instability. An AFM is used to implement the mechanical modification of progressively narrow grids to investigate the influence of pattern size on the modulation of ultrathin polystyrene films dewetting evolution. For films with different initial thicknesses, when grid size is close to or below the characteristic wavelength of instability, the spinodal dewetting is suppressed, and film rupture is restricted to the cutting trench. We will show in this paper it is possible to generate only one droplet per gridded area on a thin film subsequent to nucleation dominated dewetting on a non-patterned substrate. Furthermore, when the grid periodicity exceeds the spinodal length, the number of droplets in predefined areas gradually approaches that associated with unconfined dewetting. PMID:25283744

  19. GIS characterization of spatially distributed lifeline damage

    USGS Publications Warehouse

    Toprak, Selcuk; O'Rourke, Thomas; Tutuncu, Ilker

    1999-01-01

    This paper describes the visualization of spatially distributed water pipeline damage following an earthquake using geographical information systems (GIS). Pipeline damage is expressed as a repair rate (RR). Repair rate contours are developed with GIS by dividing the study area into grid cells (n ?? n), determining the number of particular pipeline repairs in each grid cell, and dividing the number of repairs by the length of that pipeline in each cell area. The resulting contour plot is a two-dimensional visualization of point source damage. High damage zones are defined herein as areas with an RR value greater than the mean RR for the entire study area of interest. A hyperbolic relationship between visual display of high pipeline damage zones and grid size, n, was developed. The relationship is expressed in terms of two dimensionless parameters, threshold area coverage (TAC) and dimensionless grid size (DGS). The relationship is valid over a wide range of different map scales spanning approximately 1,200 km2 for the largest portion of the Los Angeles water distribution system to 1 km2 for the Marina in San Francisco. This relationship can aid GIS users to get sufficiently refined, but easily visualized, maps of damage patterns.

  20. Texture analysis of aeromagnetic data for enhancing geologic features using co-occurrence matrices in Elallaqi area, South Eastern Desert of Egypt

    NASA Astrophysics Data System (ADS)

    Eldosouky, Ahmed M.; Elkhateeb, Sayed O.

    2018-06-01

    Enhancement of aeromagnetic data for qualitative purposes depends on the variations of texture and amplitude to outline various geologic features within the data. The texture of aeromagnetic data consists continuity of adjacent anomalies, size, and pattern. Variations in geology, or particularly rock magnetization, in a study area cause fluctuations in texture. In the present study, the anomalous features of Elallaqi area were extracted from aeromagnetic data. In order to delineate textures from the aeromagnetic data, the Red, Green, and Blue Co-occurrence Matrices (RGBCM) were applied to the reduced to the pole (RTP) grid of Elallaqi district in the South Eastern Desert of Egypt. The RGBCM are fashioned of sets of spatial analytical parameters that transform magnetic data into texture forms. Six texture features (parameters), i.e. Correlation, Contrast, Entropy, Homogeneity, Second Moment, and Variance, of RGB Co-occurrence Matrices (RGBCM) are used for analyzing the texture of the RTP grid in this study. These six RGBCM texture characteristics were mixed into a single image using principal component analysis. The calculated texture images present geologic characteristics and structures with much greater sidelong resolution than the original RTP grid. The estimated texture images enabled us to distinguish multiple geologic regions and structures within Elallaqi area including geologic terranes, lithologic boundaries, cracks, and faults. The faults of RGBCM maps were more represented than those of magnetic derivatives providing enhancement of the fine structures of Elallaqi area like the NE direction which scattered WNW metavolcanics and metasediments trending in the northwestern division of Elallaqi area.

Top