TABULATED EQUIVALENT SDR FLAMELET (TESF) MODEFL
DOE Office of Scientific and Technical Information (OSTI.GOV)
KUNDU, PRITHWISH; AMEEN, mUHSIN MOHAMMED; UNNIKRISHNAN, UMESH
The code consists of an implementation of a novel tabulated combustion model for non-premixed flames in CFD solvers. This novel technique/model is used to implement an unsteady flamelet tabulation without using progress variables for non-premixed flames. It also has the capability to include history effects which is unique within tabulated flamelet models. The flamelet table generation code can be run in parallel to generate tables with large chemistry mechanisms in relatively short wall clock times. The combustion model/code reads these tables. This framework can be coupled with any CFD solver with RANS as well as LES turbulence models. This frameworkmore » enables CFD solvers to run large chemistry mechanisms with large number of grids at relatively lower computational costs. Currently it has been coupled with the Converge CFD code and validated against available experimental data. This model can be used to simulate non-premixed combustion in a variety of applications like reciprocating engines, gas turbines and industrial burners operating over a wide range of fuels.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Begovich, C.L.; Eckerman, K.F.; Schlatter, E.C.
1981-08-01
The DARTAB computer code combines radionuclide environmental exposure data with dosimetric and health effects data to generate tabulations of the predicted impact of radioactive airborne effluents. DARTAB is independent of the environmental transport code used to generate the environmental exposure data and the codes used to produce the dosimetric and health effects data. Therefore human dose and risk calculations need not be added to every environmental transport code. Options are included in DARTAB to permit the user to request tabulations by various topics (e.g., cancer site, exposure pathway, etc.) to facilitate characterization of the human health impacts of the effluents.more » The DARTAB code was written at ORNL for the US Environmental Protection Agency, Office of Radiation Programs.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cullen, D.E.
1977-01-12
A code, SIGMA1, has been designed to Doppler broaden evaluated cross sections in the ENDF/B format. The code can only be applied to tabulated data that vary linearly in energy and cross section between tabulated points. This report describes the methods used in the code and serves as a user's guide to the code.
Geographic Access to Specialty Mental Health Care Across High- and Low-Income US Communities.
Cummings, Janet R; Allen, Lindsay; Clennon, Julie; Ji, Xu; Druss, Benjamin G
2017-05-01
With the future of the Affordable Care Act and Medicaid program unclear, it is critical to examine the geographic availability of specialty mental health treatment resources that serve low-income populations across local communities. To examine the geographic availability of community-based specialty mental health treatment resources and how these resources are distributed by community socioeconomic status. Measures of the availability of specialty mental health treatment resources were derived using national data for 31 836 zip code tabulation areas from 2013 to 2015. Analyses examined the association between community socioeconomic status (assessed by median household income quartiles) and resource availability using logistic regressions. Models controlled for zip code tabulation area-level demographic characteristics and state indicators. Dichotomous indicators for whether a zip code tabulation area had any (1) outpatient mental health treatment facility (more than nine-tenths of which offer payment arrangements for low-income populations), (2) office-based practice of mental health specialist physician(s), (3) office-based practice of nonphysician mental health professionals (eg, therapists), and (4) mental health facility or office-based practice (ie, any community-based resource). Of the 31 836 zip code tabulation areas in the study, more than four-tenths (3382 of 7959 [42.5%]) of communities in the highest income quartile (mean income, $81 207) had any community-based mental health treatment resource vs 23.1% of communities (1841 of 7959) in the lowest income quartile (mean income, $30 534) (adjusted odds ratio, 1.74; 95% CI, 1.50-2.03). When examining the distribution of mental health professionals, 25.3% of the communities (2014 of 7959) in the highest income quartile had a mental health specialist physician practice vs 8.0% (637 of 7959) of those in the lowest income quartile (adjusted odds ratio, 3.04; 95% CI, 2.53-3.66). Similarly, 35.1% of the communities (2792 of 7959) in the highest income quartile had a nonphysician mental health professional practice vs 12.9% (1029 of 7959) of those in the lowest income quartile (adjusted odds ratio, 2.77; 95%, 2.35-3.26). In contrast, outpatient mental health treatment facilities were less likely to be located in the communities in the highest vs lowest income quartiles (12.9% [1025 of 7959] vs 16.5% [1317 of 7959]; adjusted odds ratio, 0.43; 95% CI, 0.37-0.51). More than seven-tenths of the lowest income communities with any resource (71.5% [1317 of 1841]) had an outpatient mental health treatment facility. Mental health treatment facilities are more likely to be located in poorer communities, whereas office-based practices of mental health professionals are more likely to be located in higher-income communities. These findings indicate that mental health treatment facilities constitute the backbone of the specialty mental health treatment infrastructure in low-income communities. Policies are needed to support and expand available resources for this critical infrastructure.
NASA Astrophysics Data System (ADS)
Laune, Jordan; Tzeferacos, Petros; Feister, Scott; Fatenejad, Milad; Yurchak, Roman; Flocke, Norbert; Weide, Klaus; Lamb, Donald
2017-10-01
Thermodynamic and opacity properties of materials are necessary to accurately simulate laser-driven laboratory experiments. Such data are compiled in tabular format since the thermodynamic range that needs to be covered cannot be described with one single theoretical model. Moreover, tabulated data can be made available prior to runtime, reducing both compute cost and code complexity. This approach is employed by the FLASH code. Equation of state (EoS) and opacity data comes in various formats, matrix-layouts, and file-structures. We discuss recent developments on opacplot2, an open-source Python module that manipulates tabulated EoS and opacity data. We present software that builds upon opacplot2 and enables easy-to-use conversion of different table formats into the IONMIX format, the native tabular input used by FLASH. Our work enables FLASH users to take advantage of a wider range of accurate EoS and opacity tables in simulating HELP experiments at the National Laser User Facilities.
Verification of low-Mach number combustion codes using the method of manufactured solutions
NASA Astrophysics Data System (ADS)
Shunn, Lee; Ham, Frank; Knupp, Patrick; Moin, Parviz
2007-11-01
Many computational combustion models rely on tabulated constitutive relations to close the system of equations. As these reactive state-equations are typically multi-dimensional and highly non-linear, their implications on the convergence and accuracy of simulation codes are not well understood. In this presentation, the effects of tabulated state-relationships on the computational performance of low-Mach number combustion codes are explored using the method of manufactured solutions (MMS). Several MMS examples are developed and applied, progressing from simple one-dimensional configurations to problems involving higher dimensionality and solution-complexity. The manufactured solutions are implemented in two multi-physics hydrodynamics codes: CDP developed at Stanford University and FUEGO developed at Sandia National Laboratories. In addition to verifying the order-of-accuracy of the codes, the MMS problems help highlight certain robustness issues in existing variable-density flow-solvers. Strategies to overcome these issues are briefly discussed.
Radiological performance assessment for the E-Area Vaults Disposal Facility. Appendices A through M
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cook, J.R.
1994-04-15
These document contains appendices A-M for the performance assessment. They are A: details of models and assumptions, B: computer codes, C: data tabulation, D: geochemical interactions, E: hydrogeology of the Savannah River Site, F: software QA plans, G: completeness review guide, H: performance assessment peer review panel recommendations, I: suspect soil performance analysis, J: sensitivity/uncertainty analysis, K: vault degradation study, L: description of naval reactor waste disposal, M: porflow input file. (GHH)
Mapping forest types in Worcester County, Maryland, using LANDSAT data
NASA Technical Reports Server (NTRS)
Burtis, J., Jr.; Witt, R. G.
1981-01-01
The feasibility of mapping Level 2 forest cover types for a county-sized area on Maryland's Eastern Shore was demonstrated. A Level 1 land use/land cover classification was carried out for all of Worcester County as well. A June 1978 LANDSAT scene was utilized in a classification which employed two software packages on different computers (IDIMS on an HP 3000 and ASTEP-II on a Univac 1108). A twelve category classification scheme was devised for the study area. Resulting products include black and white line printer maps, final color coded classification maps, digitally enhanced color imagery and tabulated acreage statistics for all land use and land cover types.
NASA Astrophysics Data System (ADS)
Kotchenova, Svetlana Y.; Vermote, Eric F.; Matarrese, Raffaella; Klemm, Frank J., Jr.
2006-09-01
A vector version of the 6S (Second Simulation of a Satellite Signal in the Solar Spectrum) radiative transfer code (6SV1), which enables accounting for radiation polarization, has been developed and validated against a Monte Carlo code, Coulson's tabulated values, and MOBY (Marine Optical Buoy System) water-leaving reflectance measurements. The developed code was also tested against the scalar codes SHARM, DISORT, and MODTRAN to evaluate its performance in scalar mode and the influence of polarization. The obtained results have shown a good agreement of 0.7% in comparison with the Monte Carlo code, 0.2% for Coulson's tabulated values, and 0.001-0.002 for the 400-550 nm region for the MOBY reflectances. Ignoring the effects of polarization led to large errors in calculated top-of-atmosphere reflectances: more than 10% for a molecular atmosphere and up to 5% for an aerosol atmosphere. This new version of 6S is intended to replace the previous scalar version used for calculation of lookup tables in the MODIS (Moderate Resolution Imaging Spectroradiometer) atmospheric correction algorithm.
Kotchenova, Svetlana Y; Vermote, Eric F; Matarrese, Raffaella; Klemm, Frank J
2006-09-10
A vector version of the 6S (Second Simulation of a Satellite Signal in the Solar Spectrum) radiative transfer code (6SV1), which enables accounting for radiation polarization, has been developed and validated against a Monte Carlo code, Coulson's tabulated values, and MOBY (Marine Optical Buoy System) water-leaving reflectance measurements. The developed code was also tested against the scalar codes SHARM, DISORT, and MODTRAN to evaluate its performance in scalar mode and the influence of polarization. The obtained results have shown a good agreement of 0.7% in comparison with the Monte Carlo code, 0.2% for Coulson's tabulated values, and 0.001-0.002 for the 400-550 nm region for the MOBY reflectances. Ignoring the effects of polarization led to large errors in calculated top-of-atmosphere reflectances: more than 10% for a molecular atmosphere and up to 5% for an aerosol atmosphere. This new version of 6S is intended to replace the previous scalar version used for calculation of lookup tables in the MODIS (Moderate Resolution Imaging Spectroradiometer) atmospheric correction algorithm.
Empirical analysis of knowledge bases to support structured output in the Arden syntax.
Jenders, Robert A
2013-01-01
Structured output has been suggested for the Arden Syntax to facilitate interoperability. Tabulate the components of WRITE statements in a corpus of medical logic modules (MLMs)in order to validate requiring structured output. WRITE statements were tabulated in 258 MLMs from 2 organizations. In a total of 351 WRITE statements, email destinations (226) predominated, and 39 orders and 40 coded output elements also were tabulated. Free-text strings predominated as the message data. Arden WRITE statements contain considerable potentially structured data now included as free text. A future, normative structured WRITE statement must address a variety of data types and destinations.
Wang, Yueyan; Ponce, Ninez A; Wang, Pan; Opsomer, Jean D; Yu, Hongjian
2015-12-01
We propose a method to meet challenges in generating health estimates for granular geographic areas in which the survey sample size is extremely small. Our generalized linear mixed model predicts health outcomes using both individual-level and neighborhood-level predictors. The model's feature of nonparametric smoothing function on neighborhood-level variables better captures the association between neighborhood environment and the outcome. Using 2011 to 2012 data from the California Health Interview Survey, we demonstrate an empirical application of this method to estimate the fraction of residents without health insurance for Zip Code Tabulation Areas (ZCTAs). Our method generated stable estimates of uninsurance for 1519 of 1765 ZCTAs (86%) in California. For some areas with great socioeconomic diversity across adjacent neighborhoods, such as Los Angeles County, the modeled uninsured estimates revealed much heterogeneity among geographically adjacent ZCTAs. The proposed method can increase the value of health surveys by providing modeled estimates for health data at a granular geographic level. It can account for variations in health outcomes at the neighborhood level as a result of both socioeconomic characteristics and geographic locations.
An X-Ray Analysis Database of Photoionization Cross Sections Including Variable Ionization
NASA Technical Reports Server (NTRS)
Wang, Ping; Cohen, David H.; MacFarlane, Joseph J.; Cassinelli, Joseph P.
1997-01-01
Results of research efforts in the following areas are discussed: review of the major theoretical and experimental data of subshell photoionization cross sections and ionization edges of atomic ions to assess the accuracy of the data, and to compile the most reliable of these data in our own database; detailed atomic physics calculations to complement the database for all ions of 17 cosmically abundant elements; reconciling the data from various sources and our own calculations; and fitting cross sections with functional approximations and incorporating these functions into a compact computer code.Also, efforts included adapting an ionization equilibrium code, tabulating results, and incorporating them into the overall program and testing the code (both ionization equilibrium and opacity codes) with existing observational data. The background and scientific applications of this work are discussed. Atomic physics cross section models and calculations are described. Calculation results are compared with available experimental data and other theoretical data. The functional approximations used for fitting cross sections are outlined and applications of the database are discussed.
Lesko, Mehdi M; Woodford, Maralyn; White, Laura; O'Brien, Sarah J; Childs, Charmaine; Lecky, Fiona E
2010-08-06
The purpose of Abbreviated Injury Scale (AIS) is to code various types of Traumatic Brain Injuries (TBI) based on their anatomical location and severity. The Marshall CT Classification is used to identify those subgroups of brain injured patients at higher risk of deterioration or mortality. The purpose of this study is to determine whether and how AIS coding can be translated to the Marshall Classification Initially, a Marshall Class was allocated to each AIS code through cross-tabulation. This was agreed upon through several discussion meetings with experts from both fields (clinicians and AIS coders). Furthermore, in order to make this translation possible, some necessary assumptions with regards to coding and classification of mass lesions and brain swelling were essential which were all approved and made explicit. The proposed method involves two stages: firstly to determine all possible Marshall Classes which a given patient can attract based on allocated AIS codes; via cross-tabulation and secondly to assign one Marshall Class to each patient through an algorithm. This method can be easily programmed in computer softwares and it would enable future important TBI research programs using trauma registry data.
2010-01-01
Background The purpose of Abbreviated Injury Scale (AIS) is to code various types of Traumatic Brain Injuries (TBI) based on their anatomical location and severity. The Marshall CT Classification is used to identify those subgroups of brain injured patients at higher risk of deterioration or mortality. The purpose of this study is to determine whether and how AIS coding can be translated to the Marshall Classification Methods Initially, a Marshall Class was allocated to each AIS code through cross-tabulation. This was agreed upon through several discussion meetings with experts from both fields (clinicians and AIS coders). Furthermore, in order to make this translation possible, some necessary assumptions with regards to coding and classification of mass lesions and brain swelling were essential which were all approved and made explicit. Results The proposed method involves two stages: firstly to determine all possible Marshall Classes which a given patient can attract based on allocated AIS codes; via cross-tabulation and secondly to assign one Marshall Class to each patient through an algorithm. Conclusion This method can be easily programmed in computer softwares and it would enable future important TBI research programs using trauma registry data. PMID:20691038
Code of Federal Regulations, 2010 CFR
2010-04-01
... for collection by the FHWA as requested. A tabulation of this data is contained in the SI&A sheet distributed by the FHWA as part of the “Recording and Coding Guide for the Structure Inventory and Appraisal... version. Report the data using FHWA established procedures as outlined in the “Recording and Coding Guide...
Three computer codes to read, plot and tabulate operational test-site recorded solar data
NASA Technical Reports Server (NTRS)
Stewart, S. D.; Sampson, R. S., Jr.; Stonemetz, R. E.; Rouse, S. L.
1980-01-01
Computer programs used to process data that will be used in the evaluation of collector efficiency and solar system performance are described. The program, TAPFIL, reads data from an IBM 360 tape containing information (insolation, flowrates, temperatures, etc.) from 48 operational solar heating and cooling test sites. Two other programs, CHPLOT and WRTCNL, plot and tabulate the data from the direct access, unformatted TAPFIL file. The methodology of the programs, their inputs, and their outputs are described.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Not Available
1976-07-01
The memorandum contains many detailed tabulations, cross tabulations, and major conclusions for policy assessment resulting from a survey taken in connection with a research effort examining the role of individuals attitudes and perceptions in deciding whether or not to carpool. The research was based upon a survey of commuters in 3 major urban areas and has resulted in a sizeable new data base on respondents' socio-economic and worktrip characteristics, travel perceptions, and travel preferences.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cullen, D.E.
1978-07-04
The code SIGMA1 Doppler broadens evaluated cross sections in the ENDF/B format. The code can be applied only to data that vary as a linear function of energy and cross section between tabulated points. This report describes the methods used in the code and serves as a user's guide to the code. 6 figures, 2 tables.
Thermodynamic and transport properties of gaseous tetrafluoromethane in chemical equilibrium
NASA Technical Reports Server (NTRS)
Hunt, J. L.; Boney, L. R.
1973-01-01
Equations and in computer code are presented for the thermodynamic and transport properties of gaseous, undissociated tetrafluoromethane (CF4) in chemical equilibrium. The computer code calculates the thermodynamic and transport properties of CF4 when given any two of five thermodynamic variables (entropy, temperature, volume, pressure, and enthalpy). Equilibrium thermodynamic and transport property data are tabulated and pressure-enthalpy diagrams are presented.
Motor vehicle crashes in New Zealand, 2008
DOT National Transportation Integrated Search
2009-01-01
This statistical statement contains tabulations of information coded from Traffic Crash Reports. To put these data into context, the following is a brief description of the process : which has resulted in this publication. When a road traffic crash i...
Motor vehicle crashes in New Zealand, 2009
DOT National Transportation Integrated Search
2010-01-01
This statistical statement contains tabulations of information coded from Traffic Crash Reports. To put these data into context, the following is a brief description of the process : which has resulted in this publication. When a road traffic crash i...
NASA Technical Reports Server (NTRS)
Goldberg, Robert K.; Carney, Kelly S.; Dubois, Paul; Hoffarth, Canio; Khaled, Bilal; Shyamsunder, Loukham; Rajan, Subramaniam; Blankenhorn, Gunther
2017-01-01
The need for accurate material models to simulate the deformation, damage and failure of polymer matrix composites under impact conditions is becoming critical as these materials are gaining increased use in the aerospace and automotive communities. The aerospace community has identified several key capabilities which are currently lacking in the available material models in commercial transient dynamic finite element codes. To attempt to improve the predictive capability of composite impact simulations, a next generation material model is being developed for incorporation within the commercial transient dynamic finite element code LS-DYNA. The material model, which incorporates plasticity, damage and failure, utilizes experimentally based tabulated input to define the evolution of plasticity and damage and the initiation of failure as opposed to specifying discrete input parameters such as modulus and strength. The plasticity portion of the orthotropic, three-dimensional, macroscopic composite constitutive model is based on an extension of the Tsai-Wu composite failure model into a generalized yield function with a non-associative flow rule. For the damage model, a strain equivalent formulation is used to allow for the uncoupling of the deformation and damage analyses. For the failure model, a tabulated approach is utilized in which a stress or strain based invariant is defined as a function of the location of the current stress state in stress space to define the initiation of failure. Failure surfaces can be defined with any arbitrary shape, unlike traditional failure models where the mathematical functions used to define the failure surface impose a specific shape on the failure surface. In the current paper, the complete development of the failure model is described and the generation of a tabulated failure surface for a representative composite material is discussed.
South Carolina traffic collision fact book, 2006
DOT National Transportation Integrated Search
2006-01-01
In accordance with Section 56-5-1350 of the South Carolina Code of Laws, a : tabulation and analysis of collision reports has been completed for the year : 2006 as disclosed in this publication. : The number of traffic fatalities decreased from 1,093...
South Carolina traffic collision fact book, 2007
DOT National Transportation Integrated Search
2007-01-01
In accordance with Section 56-5-1350 of the South Carolina Code of Laws, a : tabulation and analysis of collision reports has been completed for the year : 2007 as disclosed in this publication. : The number of traffic fatalities increased from 1,044...
Annual peak discharges from small drainage areas in Montana through September 1978
Omang, R.J.; Parrett, C.; Hull, J.A.
1979-01-01
Annual peak stage and discharge data have been collected and tabulated for crest-stage gaging sites in Montana. The crest-stage program was begun in July 1955 to investigate the magnitude and frequency of floods from small drainage areas. The program has expanded from 45 crest-stage gaging stations initially to 173 stations maintained in 1978. Data are tabulated for the period of record. (Woodard-USGS)
Radiative transfer codes for atmospheric correction and aerosol retrieval: intercomparison study.
Kotchenova, Svetlana Y; Vermote, Eric F; Levy, Robert; Lyapustin, Alexei
2008-05-01
Results are summarized for a scientific project devoted to the comparison of four atmospheric radiative transfer codes incorporated into different satellite data processing algorithms, namely, 6SV1.1 (second simulation of a satellite signal in the solar spectrum, vector, version 1.1), RT3 (radiative transfer), MODTRAN (moderate resolution atmospheric transmittance and radiance code), and SHARM (spherical harmonics). The performance of the codes is tested against well-known benchmarks, such as Coulson's tabulated values and a Monte Carlo code. The influence of revealed differences on aerosol optical thickness and surface reflectance retrieval is estimated theoretically by using a simple mathematical approach. All information about the project can be found at http://rtcodes.ltdri.org.
Radiative transfer codes for atmospheric correction and aerosol retrieval: intercomparison study
NASA Astrophysics Data System (ADS)
Kotchenova, Svetlana Y.; Vermote, Eric F.; Levy, Robert; Lyapustin, Alexei
2008-05-01
Results are summarized for a scientific project devoted to the comparison of four atmospheric radiative transfer codes incorporated into different satellite data processing algorithms, namely, 6SV1.1 (second simulation of a satellite signal in the solar spectrum, vector, version 1.1), RT3 (radiative transfer), MODTRAN (moderate resolution atmospheric transmittance and radiance code), and SHARM (spherical harmonics). The performance of the codes is tested against well-known benchmarks, such as Coulson's tabulated values and a Monte Carlo code. The influence of revealed differences on aerosol optical thickness and surface reflectance retrieval is estimated theoretically by using a simple mathematical approach. All information about the project can be found at http://rtcodes.ltdri.org.
Coding coarse grained polymer model for LAMMPS and its application to polymer crystallization
NASA Astrophysics Data System (ADS)
Luo, Chuanfu; Sommer, Jens-Uwe
2009-08-01
We present a patch code for LAMMPS to implement a coarse grained (CG) model of poly(vinyl alcohol) (PVA). LAMMPS is a powerful molecular dynamics (MD) simulator developed at Sandia National Laboratories. Our patch code implements tabulated angular potential and Lennard-Jones-9-6 (LJ96) style interaction for PVA. Benefited from the excellent parallel efficiency of LAMMPS, our patch code is suitable for large-scale simulations. This CG-PVA code is used to study polymer crystallization, which is a long-standing unsolved problem in polymer physics. By using parallel computing, cooling and heating processes for long chains are simulated. The results show that chain-folded structures resembling the lamellae of polymer crystals are formed during the cooling process. The evolution of the static structure factor during the crystallization transition indicates that long-range density order appears before local crystalline packing. This is consistent with some experimental observations by small/wide angle X-ray scattering (SAXS/WAXS). During the heating process, it is found that the crystalline regions are still growing until they are fully melted, which can be confirmed by the evolution both of the static structure factor and average stem length formed by the chains. This two-stage behavior indicates that melting of polymer crystals is far from thermodynamic equilibrium. Our results concur with various experiments. It is the first time that such growth/reorganization behavior is clearly observed by MD simulations. Our code can be easily used to model other type of polymers by providing a file containing the tabulated angle potential data and a set of appropriate parameters. Program summaryProgram title: lammps-cgpva Catalogue identifier: AEDE_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEDE_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: GNU's GPL No. of lines in distributed program, including test data, etc.: 940 798 No. of bytes in distributed program, including test data, etc.: 12 536 245 Distribution format: tar.gz Programming language: C++/MPI Computer: Tested on Intel-x86 and AMD64 architectures. Should run on any architecture providing a C++ compiler Operating system: Tested under Linux. Any other OS with C++ compiler and MPI library should suffice Has the code been vectorized or parallelized?: Yes RAM: Depends on system size and how many CPUs are used Classification: 7.7 External routines: LAMMPS ( http://lammps.sandia.gov/), FFTW ( http://www.fftw.org/) Nature of problem: Implementing special tabular angle potentials and Lennard-Jones-9-6 style interactions of a coarse grained polymer model for LAMMPS code. Solution method: Cubic spline interpolation of input tabulated angle potential data. Restrictions: The code is based on a former version of LAMMPS. Unusual features.: Any special angular potential can be used if it can be tabulated. Running time: Seconds to weeks, depending on system size, speed of CPU and how many CPUs are used. The test run provided with the package takes about 5 minutes on 4 AMD's opteron (2.6 GHz) CPUs. References:D. Reith, H. Meyer, F. Müller-Plathe, Macromolecules 34 (2001) 2335-2345. H. Meyer, F. Müller-Plathe, J. Chem. Phys. 115 (2001) 7807. H. Meyer, F. Müller-Plathe, Macromolecules 35 (2002) 1241-1252.
Annual peak discharges from small drainage areas in Montana through September 1977
Omang, R.J.; Hull, J.A.
1978-01-01
Annual peak stage and stream-discharge data have been collected and tabulated for crest-stage gaging sites in Montana. The crest-stage program was begun in July 1955 to investigate the magnitude and frequency of floods from small drainage areas. The program has expanded from 45 crest-stage gaging stations initially to 191 stations in 1977. Data are tabulated for 336 sites throughout the period of record. (Woodard-USGS)
Annual peak discharges from small drainage areas in Montana through September 1980
Omang, R.J.; Parrett, Charles; Hull, J.A.
1955-01-01
Annual peak stage and discharge data have been collected and tabulated for crest-stage gaging sites in Montana. The crest-stage program was begun in July 1955 to investigate the magnitude and frequency of floods from small drainage areas. The program has expanded from 45 crest-stage gaging stations initially to 172 stations maintained in 1980. Data in the report are tabulated for the period of record. (USGS)
Annual peak discharges from small drainage areas in Montana through September 1979
Omang, R.J.; Parrett, C.; Hull, J.A.
1955-01-01
Annual peak stage and discharge data have been collected and tabulated for crest-stage gaging sites in Montana. The crest-stage program was begun in July 1955 to investigate the magnitude and frequency of floods from small drainage areas. The program has expanded from 45 crest-stage gaging stations initially to 173 stations maintained in 1979. Data in the report are tabulated for the period of record. (USGS)
Performance analysis of the word synchronization properties of the outer code in a TDRSS decoder
NASA Technical Reports Server (NTRS)
Costello, D. J., Jr.; Lin, S.
1984-01-01
A self-synchronizing coding scheme for NASA's TDRSS satellite system is a concatenation of a (2,1,7) inner convolutional code with a (255,223) Reed-Solomon outer code. Both symbol and word synchronization are achieved without requiring that any additional symbols be transmitted. An important parameter which determines the performance of the word sync procedure is the ratio of the decoding failure probability to the undetected error probability. Ideally, the former should be as small as possible compared to the latter when the error correcting capability of the code is exceeded. A computer simulation of a (255,223) Reed-Solomon code as carried out. Results for decoding failure probability and for undetected error probability are tabulated and compared.
Clusters in irregular areas and lattices.
Wieczorek, William F; Delmerico, Alan M; Rogerson, Peter A; Wong, David W S
2012-01-01
Geographic areas of different sizes and shapes of polygons that represent counts or rate data are often encountered in social, economic, health, and other information. Often political or census boundaries are used to define these areas because the information is available only for those geographies. Therefore, these types of boundaries are frequently used to define neighborhoods in spatial analyses using geographic information systems and related approaches such as multilevel models. When point data can be geocoded, it is possible to examine the impact of polygon shape on spatial statistical properties, such as clustering. We utilized point data (alcohol outlets) to examine the issue of polygon shape and size on visualization and statistical properties. The point data were allocated to regular lattices (hexagons and squares) and census areas for zip-code tabulation areas and tracts. The number of units in the lattices was set to be similar to the number of tract and zip-code areas. A spatial clustering statistic and visualization were used to assess the impact of polygon shape for zip- and tract-sized units. Results showed substantial similarities and notable differences across shape and size. The specific circumstances of a spatial analysis that aggregates points to polygons will determine the size and shape of the areal units to be used. The irregular polygons of census units may reflect underlying characteristics that could be missed by large regular lattices. Future research to examine the potential for using a combination of irregular polygons and regular lattices would be useful.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fields, D.E.; Little, C.A.
1978-11-01
The APORT computer code was developed to apportion variables tabulated for polygon-structured civil districts onto cells of a polar grid. The apportionment is based on fractional overlap between the polygon and the grid cells. Centering the origin of the polar system at a pollutant source site yields results that are very useful for assessing and interpreting the effects of airborne pollutant dissemination. The APOPLT graphics code, which uses the same data set as APORT, provides a convenient visual display of the polygon structure and the extent of the polar grid. The APORT/APOPLT methodology was verified by application to county summariesmore » of cattle population for counties surrounding the Oyster Creek, New Jersey, nuclear power plant. These numerical results, which were obtained using approximately 2-min computer time on an IBM System 360/91 computer, compare favorably to results of manual computations in both speed and accuracy.« less
VLSI Technology for Cognitive Radio
NASA Astrophysics Data System (ADS)
VIJAYALAKSHMI, B.; SIDDAIAH, P.
2017-08-01
One of the most challenging tasks of cognitive radio is the efficiency in the spectrum sensing scheme to overcome the spectrum scarcity problem. The popular and widely used spectrum sensing technique is the energy detection scheme as it is very simple and doesn’t require any previous information related to the signal. We propose one such approach which is an optimised spectrum sensing scheme with reduced filter structure. The optimisation is done in terms of area and power performance of the spectrum. The simulations of the VLSI structure of the optimised flexible spectrum is done using verilog coding by using the XILINX ISE software. Our method produces performance with 13% reduction in area and 66% reduction in power consumption in comparison to the flexible spectrum sensing scheme. All the results are tabulated and comparisons are made. A new scheme for optimised and effective spectrum sensing opens up with our model.
Truncation Depth Rule-of-Thumb for Convolutional Codes
NASA Technical Reports Server (NTRS)
Moision, Bruce
2009-01-01
In this innovation, it is shown that a commonly used rule of thumb (that the truncation depth of a convolutional code should be five times the memory length, m, of the code) is accurate only for rate 1/2 codes. In fact, the truncation depth should be 2.5 m/(1 - r), where r is the code rate. The accuracy of this new rule is demonstrated by tabulating the distance properties of a large set of known codes. This new rule was derived by bounding the losses due to truncation as a function of the code rate. With regard to particular codes, a good indicator of the required truncation depth is the path length at which all paths that diverge from a particular path have accumulated the minimum distance of the code. It is shown that the new rule of thumb provides an accurate prediction of this depth for codes of varying rates.
NASA Technical Reports Server (NTRS)
Bryant, N. A.; Zobrist, A. L.
1978-01-01
The paper describes the development of an image based information system and its use to process a Landsat thematic map showing land use or land cover in conjunction with a census tract polygon file to produce a tabulation of land use acreages per census tract. The system permits the efficient cross-tabulation of two or more geo-coded data sets, thereby setting the stage for the practical implementation of models of diffusion processes or cellular transformation. Characteristics of geographic information systems are considered, and functional requirements, such as data management, geocoding, image data management, and data analysis are discussed. The system is described, and the potentialities of its use are examined.
40 CFR 146.34 - Information to be considered by the Director.
Code of Federal Regulations, 2010 CFR
2010-07-01
... injected; (ii) Average and maximum injection pressure; and (iii) Qualitative analysis and ranges in..., cross sections, tabulations of wells within the area of review, and other data may be included in the... this map. (3) A tabulation of data reasonably available from public records or otherwise known to the...
Annual peak discharges from small drainage areas in Montana for stations discontinued before 1978
Omang, R.J.; Hull, J.A.; Parrett, Charles
1979-01-01
Annual peak stage and discharge data have been tabulated for crest-stage gage sites in Montana. The crest-stage program was begun in July 1955 to investigate the magnitude and frequency of floods from samll drainage areas. The program has expanded from 45 crest-stage gaging stations initially to 172 stations maintained in 1978. From 1955 to 1978, 156 stations have been discontinued. This report is a tabulation of the stage and discharge data for the discontinued stations. (Woodard-USGS)
NASA Astrophysics Data System (ADS)
Gilleron, Franck; Piron, Robin
2015-12-01
We present Dédale, a fast code implementing a simplified non-local-thermodynamic-equilibrium (NLTE) plasma model. In this approach, the stationary collisional-radiative rates equations are solved for a set of well-chosen Layzer complexes in order to determine the ion state populations. The electronic structure is approximated using the screened hydrogenic model (SHM) of More with relativistic corrections. The radiative and collisional cross-sections are based on Kramers and Van Regemorter formula, respectively, which are extrapolated to derive analytical expressions for all the rates. The latter are improved thereafter using Gaunt factors or more accurate tabulated data. Special care is taken for dielectronic rates which are compared and rescaled with quantum calculations from the Averroès code. The emissivity and opacity spectra are calculated under the same assumptions as for the radiative rates, either in a detailed manner by summing the transitions between each pair of complexes, or in a coarser statistical way by summing the one-electron transitions averaged over the complexes. Optionally, nℓ-splitting can be accounted for using a WKB approach in an approximate potential reconstructed analytically from the screened charges. It is also possible to improve the spectra by replacing some transition arrays with more accurate data tabulated using the SCO-RCG or FAC codes. This latter option is particularly useful for K-shell emission spectroscopy. The Dédale code was used to submit neon and tungsten cases in the last NLTE-8 workshop (Santa Fe, November 4-8, 2013). Some of these results are presented, as well as comparisons with Averroès calculations.
A comparison of frame synchronization methods. [Deep Space Network
NASA Technical Reports Server (NTRS)
Swanson, L.
1982-01-01
Different methods are considered for frame synchronization of a concatenated block code/Viterbi link. Synchronization after Viterbi decoding, synchronization before Viterbi decoding based on hard-quantized channel symbols are all compared. For each scheme, the probability under certain conditions of true detection of sync within four 10,000 bit frames is tabulated.
Plascak, Jesse J.; Molina, Yamile; Wu-Georges, Samantha; Idris, Ayah; Thompson, Beti
2016-01-01
The relationship between Latino residential segregation and self-rated health (SRH) is unclear, but might be partially affected by social capital. We investigated the association between Latino residential segregation and SRH while also examining the roles of various social capital measures. Washington State Behavioral Risk Factor Surveillance System (2012–2014) and U.S. Census data were linked by zip code and zip code tabulation area. Multilevel logistic regression models were used to estimate odds of good or better SRH by Latino residential segregation, measured by the Gini coefficient, and controlling for sociodemographic, acculturation and social capital measures of neighborhood ties, collective socialization of children, and social control. The Latino residential segregation – SRH relationship was convex, or ‘U’-shaped, such that increases in segregation among Latinos residing in lower segregation areas was associated with lower SRH while increases in segregation among Latinos residing in higher segregation areas was associated with higher SRH. The social capital measures were independently associated with SRH but had little effect on the relationship between Latino residential segregation and SRH. A convex relationship between Latino residential segregation and SRH could explain mixed findings of previous studies. Although important for SRH, social capital measures of neighborhood ties, collective socialization of children, and social control might not account for the relationship between Latino residential segregation and SRH. PMID:27173739
Plascak, Jesse J; Molina, Yamile; Wu-Georges, Samantha; Idris, Ayah; Thompson, Beti
2016-06-01
The relationship between Latino residential segregation and self-rated health (SRH) is unclear, but might be partially affected by social capital. We investigated the association between Latino residential segregation and SRH while also examining the roles of various social capital measures. Washington State Behavioral Risk Factor Surveillance System (2012-2014) and U.S. Census data were linked by zip code and zip code tabulation area. Multilevel logistic regression models were used to estimate odds of good or better SRH by Latino residential segregation, measured by the Gini coefficient, and controlling for sociodemographic, acculturation and social capital measures of neighborhood ties, collective socialization of children, and social control. The Latino residential segregation - SRH relationship was convex, or 'U'-shaped, such that increases in segregation among Latinos residing in lower segregation areas was associated with lower SRH while increases in segregation among Latinos residing in higher segregation areas was associated with higher SRH. The social capital measures were independently associated with SRH but had little effect on the relationship between Latino residential segregation and SRH. A convex relationship between Latino residential segregation and SRH could explain mixed findings of previous studies. Although important for SRH, social capital measures of neighborhood ties, collective socialization of children, and social control might not account for the relationship between Latino residential segregation and SRH. Copyright © 2016 Elsevier Ltd. All rights reserved.
Tabulated Combustion Model Development For Non-Premixed Flames
NASA Astrophysics Data System (ADS)
Kundu, Prithwish
Turbulent non-premixed flames play a very important role in the field of engineering ranging from power generation to propulsion. The coupling of fluid mechanics and complicated combustion chemistry of fuels pose a challenge for the numerical modeling of these type of problems. Combustion modeling in Computational Fluid Dynamics (CFD) is one of the most important tools used for predictive modeling of complex systems and to understand the basic fundamentals of combustion. Traditional combustion models solve a transport equation of each species with a source term. In order to resolve the complex chemistry accurately it is important to include a large number of species. However, the computational cost is generally proportional to the cube of number of species. The presence of a large number of species in a flame makes the use of CFD computationally expensive and beyond reach for some applications or inaccurate when solved with simplified chemistry. For highly turbulent flows, it also becomes important to incorporate the effects of turbulence chemistry interaction (TCI). The aim of this work is to develop high fidelity combustion models based on the flamelet concept and to significantly advance the existing capabilities. A thorough investigation of existing models (Finite-rate chemistry and Representative Interactive Flamelet (RIF)) and comparative study of combustion models was done initially on a constant volume combustion chamber with diesel fuel injection. The CFD modeling was validated with experimental results and was also successfully applied to a single cylinder diesel engine. The effect of number of flamelets on the RIF model and flamelet initialization strategies were studied. The RIF model with multiple flamelets is computationally expensive and a model was proposed on the frame work of RIF. The new model was based on tabulated chemistry and incorporated TCI effects. A multidimensional tabulated chemistry database generation code was developed based on the 1D diffusion flame solver. The proposed model did not use progress variables like the traditional chemistry tabulation methods. The resulting model demonstrated an order of magnitude computational speed up over the RIF model. The results were validated across a wide range of operating conditions for diesel injections and the results were in close agreement to those of the experimental data. History of scalar dissipation rates plays a very important role in non premixed flames. However, tabulated methods have not been able to incorporate this physics in their models. A comparative approach is developed that can quantify these effects and find correlations with flow variables. A new model is proposed to include these effects in tabulated combustion models. The model is initially validated for 1D counterflow diffusion flame problems at engine conditions. The model is further implemented and validated in a 3D RANS code across a range of operating conditions for spray flames.
NASA Technical Reports Server (NTRS)
Warren, W. H., Jr.
1982-01-01
The contents and format of the machine-readable version of the cataloque distributed by the Astronomical Data Center are described. Coding for the various scales and abbreviations used in the catalogue are tabulated and certain revisions to the machine version made to improve storage efficiency and notation are discussed.
New Millenium Inflatable Structures Technology
NASA Technical Reports Server (NTRS)
Mollerick, Ralph
1997-01-01
Specific applications where inflatable technology can enable or enhance future space missions are tabulated. The applicability of the inflatable technology to large aperture infra-red astronomy missions is discussed. Space flight validation and risk reduction are emphasized along with the importance of analytical tools in deriving structurally sound concepts and performing optimizations using compatible codes. Deployment dynamics control, fabrication techniques, and system testing are addressed.
NASA Technical Reports Server (NTRS)
Goldberg, Robert K.; Carney, Kelly S.; Dubois, Paul; Hoffarth, Canio; Khaled, Bilal; Shyamsunder, Loukham; Rajan, Subramaniam; Blankenhorn, Gunther
2017-01-01
The need for accurate material models to simulate the deformation, damage and failure of polymer matrix composites under impact conditions is becoming critical as these materials are gaining increased use in the aerospace and automotive communities. The aerospace community has identified several key capabilities which are currently lacking in the available material models in commercial transient dynamic finite element codes. To attempt to improve the predictive capability of composite impact simulations, a next generation material model is being developed for incorporation within the commercial transient dynamic finite element code LS-DYNA. The material model, which incorporates plasticity, damage and failure, utilizes experimentally based tabulated input to define the evolution of plasticity and damage and the initiation of failure as opposed to specifying discrete input parameters such as modulus and strength. The plasticity portion of the orthotropic, three-dimensional, macroscopic composite constitutive model is based on an extension of the Tsai-Wu composite failure model into a generalized yield function with a non-associative flow rule. For the damage model, a strain equivalent formulation is used to allow for the uncoupling of the deformation and damage analyses. In the damage model, a semi-coupled approach is employed where the overall damage in a particular coordinate direction is assumed to be a multiplicative combination of the damage in that direction resulting from the applied loads in various coordinate directions. For the failure model, a tabulated approach is utilized in which a stress or strain based invariant is defined as a function of the location of the current stress state in stress space to define the initiation of failure. Failure surfaces can be defined with any arbitrary shape, unlike traditional failure models where the mathematical functions used to define the failure surface impose a specific shape on the failure surface. In the current paper, the complete development of the failure model is described and the generation of a tabulated failure surface for a representative composite material is discussed.
DOE Office of Scientific and Technical Information (OSTI.GOV)
CULLEN, D. E.
2001-06-13
Version 00 As distributed, the original evaluated data include cross sections represented in the form of a combination of resonance parameters and/or tabulated energy dependent cross sections, nominally at 0 Kelvin temperature. For use in applications, these ENDF/B-VI, Release 7 data were processed into the form of temperature dependent cross sections at eight temperatures between 0 and 2100 Kelvin, in steps of 300 Kelvin. At each temperature the cross sections are tabulated and linearly interpolable in energy. POINT2000 contains all of the evaluations in the ENDF/B-VI general purpose library, which contains evaluations for 324 materials (isotopes or naturally occurring elementalmore » mixtures of isotopes). No special purpose ENDF/B-VI libraries, such as fission products, thermal scattering, photon interaction data are included. The majority of these evaluations are complete, in the sense that they include all cross sections over the energy range 10-5 eV to at least 20 MeV. However, the following are only partial evaluations that either only contain single reactions and no total cross section (Mg24, K41, Ti46, Ti47, Ti48, Ti50 and Ni59), or do not include energy dependent cross sections above the resonance region (Ar40, Mo92, Mo98, Mo100, In115, Sn120, Sn122 and Sn124). The CCC-638/TART96 code package will soon be updated to TART2000, which is recommended for use with these data. Codes within TART2000 can be used to display these data or to run calculations using these data.« less
Current research on aviation weather (bibliography), 1979
NASA Technical Reports Server (NTRS)
Turkel, B. S.; Frost, W.
1980-01-01
The titles, managers, supporting organizations, performing organizations, investigators and objectives of 127 current research projects in advanced meteorological instruments, forecasting, icing, lightning, visibility, low level wind shear, storm hazards/severe storms, and turbulence are tabulated and cross-referenced. A list of pertinent reference material produced through the above tabulated research activities is given. The acquired information is assembled in bibliography form to provide a readily available source of information in the area of aviation meteorology.
EXFILE: A program for compiling irradiation data on UN and UC fuel pins
NASA Technical Reports Server (NTRS)
Mayer, J. T.; Smith, R. L.; Weinstein, M. B.; Davison, H. W.
1973-01-01
A FORTRAN-4 computer program for handling fuel pin data is described. Its main features include standardized output, easy access for data manipulation, and tabulation of important material property data. An additional feature allows simplified preparation of input decks for a fuel swelling computer code (CYGRO-2). Data from over 300 high temperature nitride and carbide based fuel pin irradiations are listed.
NASA Technical Reports Server (NTRS)
Tinetti, Ana F.; Maglieri, Domenic J.; Driver, Cornelius; Bobbitt, Percy J.
2011-01-01
A detailed geometric description, in wave drag format, has been developed for the Convair B-58 and North American XB-70-1 delta wing airplanes. These descriptions have been placed on electronic files, the contents of which are described in this paper They are intended for use in wave drag and sonic boom calculations. Included in the electronic file and in the present paper are photographs and 3-view drawings of the two airplanes, tabulated geometric descriptions of each vehicle and its components, and comparisons of the electronic file outputs with existing data. The comparisons include a pictorial of the two airplanes based on the present geometric descriptions, and cross-sectional area distributions for both the normal Mach cuts and oblique Mach cuts above and below the vehicles. Good correlation exists between the area distributions generated in the late 1950s and 1960s and the present files. The availability of these electronic files facilitates further validation of sonic boom prediction codes through the use of two existing data bases on these airplanes, which were acquired in the 1960s and have not been fully exploited.
Inter-comparison of Computer Codes for TRISO-based Fuel Micro-Modeling and Performance Assessment
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brian Boer; Chang Keun Jo; Wen Wu
2010-10-01
The Next Generation Nuclear Plant (NGNP), the Deep Burn Pebble Bed Reactor (DB-PBR) and the Deep Burn Prismatic Block Reactor (DB-PMR) are all based on fuels that use TRISO particles as their fundamental constituent. The TRISO particle properties include very high durability in radiation environments, hence the designs reliance on the TRISO to form the principal barrier to radioactive materials release. This durability forms the basis for the selection of this fuel type for applications such as Deep Bun (DB), which require exposures up to four times those expected for light water reactors. It follows that the study and predictionmore » of the durability of TRISO particles must be carried as part of the safety and overall performance characterization of all the designs mentioned above. Such evaluations have been carried out independently by the performers of the DB project using independently developed codes. These codes, PASTA, PISA and COPA, incorporate models for stress analysis on the various layers of the TRISO particle (and of the intervening matrix material for some of them), model for fission products release and migration then accumulation within the SiC layer of the TRISO particle, just next to the layer, models for free oxygen and CO formation and migration to the same location, models for temperature field modeling within the various layers of the TRISO particle and models for the prediction of failure rates. All these models may be either internal to the code or external. This large number of models and the possibility of different constitutive data and model formulations and the possibility of a variety of solution techniques makes it highly unlikely that the model would give identical results in the modeling of identical situations. The purpose of this paper is to present the results of an inter-comparison between the codes and to identify areas of agreement and areas that need reconciliation. The inter-comparison has been carried out by the cooperating institutions using a set of pre-defined TRISO conditions (burnup levels, temperature or power levels, etc.) and the outcome will be tabulated in the full length paper. The areas of agreement will be pointed out and the areas that require further modeling or reconciliation will be shown. In general the agreement between the codes is good within less than one order of magnitude in the prediction of TRISO failure rates.« less
Geographic Variability in Geocoding Success for West Nile Virus Cases in South Dakota
Wey, Christine L.; Griesse, Jennifer; Kightlinger, Lon; Wimberly, Michael C.
2009-01-01
Background Geocoding, the process of assigning each case a set of coordinates that closely approximates its true location, is an important component of spatial epidemiological studies. The failure to accurately geocode cases adversely affects the validity and strength of conclusions drawn from the analysis. We investigated whether there were differences among geographic locations and demographic classes in the ability to successfully geocode West Nile virus (WNV) cases in South Dakota. We successfully geocoded 1,354 cases (80.8%) to their street address locations and assigned all 1,676 cases to ZIP code tabulation areas (ZCTAs). Using spatial scan statistics, significant clusters of non-geocoded cases were identified in central and western South Dakota. Geocoding success rates were lower in areas of low population density and on Indian reservations than in other portions of the state. Geocoding success rates were lower for Native Americans than for other races. Spatial epidemiological studies should consider the potential biases that may result from excluding non-geocoded cases, particularly in rural portions of the Great Plains that contain large Native American populations. PMID:19577505
Musa, Maizura binti; Harun-Or-Rashid, M D; Sakamoto, Junichi
2011-11-16
Nurse managers have the burden of experiencing frequent ethical issues related to both their managerial and nursing care duties, according to previous international studies. However, no such study was published in Malaysia. The purpose of this study was to explore nurse managers' experience with ethical issues in six government hospitals in Malaysia including learning about the way they dealt with the issues. A cross-sectional study was conducted in August-September, 2010 involving 417 (69.2%) of total 603 nurse managers in the six Malaysian government hospitals. Data were collected using three-part self-administered questionnaire. Part I was regarding participants' demographics. Part II was about the frequency and areas of management where ethical issues were experienced, and scoring of the importance of 11 pre-identified ethical issues. Part III asked how they dealt with ethical issues in general; ways to deal with the 11 pre-identified ethical issues, and perceived stress level. Data were analyzed using descriptive statistics, cross-tabulations and Pearson's Chi-square. A total of 397 (95.2%) participants experienced ethical issues and 47.2% experienced them on weekly to daily basis. Experiencing ethical issues were not associated with areas of practice. Top area of management where ethical issues were encountered was "staff management", but "patient care" related ethical issues were rated as most important. Majority would "discuss with other nurses" in dealing generally with the issues. For pre-identified ethical issues regarding "patient care", "discuss with doctors" was preferred. Only 18.1% referred issues to "ethics committees" and 53.0% to the code of ethics. Nurse managers, regardless of their areas of practice, frequently experienced ethical issues. For dealing with these, team-approach needs to be emphasized. Proper understanding of the code of ethics is needed to provide basis for reasoning.
2006-07-01
precision of the determination of Rmax, we established a refined method based on the model of bubble formation described above in section 3.6.1 and the...development can be modeled by hydrodynamic codes based on tabulated equation-of-state data . This has previously demonstrated on ps optical breakdown...per response, including the time for reviewing instructions, searching existing data sources, gathering and maintaining the data needed, and
Innovative Techniques to Predict Atmospheric Effects on Sensor Performance
2009-10-15
since acquiring the MRO data, extensive tabulation of all of the data from all visible satellites (generally, non- resolved ) was also accomplished...efficient code has been written to run multiple OSC simulations in less time . Data from many passes of the same satellite is useful for SOI, whether it is...the data analyzed. Questions about the data were resolved using OSC to determine solar phase angle (SPA), range, time of penumbra entrance/exit and
NASA Astrophysics Data System (ADS)
Kaur, Rajnish; Kumar, Anil; Osan, Janos; Czyzycki, M.; Karydas, A. G.; Puri, Sanjiv
2017-07-01
The absolute values of the mass attenuation coefficients have been measured at sixty two photon energies across the Li (i=1-3) sub-shell absorption edges of 66Dy covering the region 7.6-14.0 keV in order to investigate the influence of near-edge processes on the attenuation coefficients. The present measured attenuation coefficients are found to be higher by up to 10% than the theoretical values evaluated from the computer code XCOM (Berger et al., 2010) and the self-consistent Dirac-Hartree-Slater (DHS) model based values tabulated by Chantler (1995) over the energy region 7.6-14.0 keV, except at energies in vicinity (few eV) of the Li (i=1-3) sub-shell absorption edge energies where the measured values are significantly higher (up to 37%) than both the sets of theoretical values. Further, the Li (i=1-3) sub-shell photoionization cross sections, (σLiP)exp, deduced from the present measured mass attenuation coefficients are compared with the non-relativistic Hartree-Fock-Slater (HFS) model based values tabulated by Scofield (1973) and those evaluated from the theoretical total photoionization attenuation coefficients tabulated by Chantler (1995). The deduced (σLiP)exp(i=1-3) values are found to be in better agreement with those evaluated from the tabulations given by Chantler (1995) than the values given by Scofield (1973) over the energy region 7.8 - 14.0 keV included in this study. However, at photon energies up to few eV above the Li edges, the deduced (σLiP)exp(i=1-3) values are found to be significantly higher (up to 32%) than both the sets of theoretical values.
DOE Office of Scientific and Technical Information (OSTI.GOV)
CULLEN, D. E.
2005-02-21
Version 00 As distributed, the original evaluated data include cross sections represented in the form of a combination of resonance parameters and/or tabulated energy dependent cross sections, nominally at 0 Kelvin temperature. For use in applications this library has been processed into the form of temperature dependent cross sections at eight neutron reactor like temperatures, between 0 and 2100 Kelvin, in steps of 300 Kelvin. It has also been processed to five astrophysics like temperatures, 1, 10, 100 eV, 1 and 10 keV. For reference purposes, 300 Kelvin is approximately 1/40 eV, so that 1 eV is approximately 12,000 Kelvin.more » At each temperature the cross sections are tabulated and linearly interpolable in energy. POINT2004 contains all of the evaluations in the ENDF/B-VI general purpose library, which contains evaluations for 328 materials (isotopes or naturally occurring elemental mixtures of isotopes). No special purpose ENDF/B-VI libraries, such as fission products, thermal scattering, or photon interaction data are included. The majority of these evaluations are complete, in the sense that they include all cross sections over the energy range 10-5 eV to at least 20 MeV. However, the following are only partial evaluations that either contain only single reactions and no total cross section (Mg24, K41, Ti46, Ti47, Ti48, Ti50 and Ni59), or do not include energy dependent cross sections above the resonance region (Ar40, Mo92, Mo98, Mo100, In115, Sn120, Sn122 and Sn124). The CCC-638/TART20002 code package is recommended for use with these data. Codes within TART can be used to display these data or to run calculations using these data.« less
NASA Technical Reports Server (NTRS)
Goldberg, Robert K.; Carney, Kelly S.; DuBois, Paul; Hoffarth, Canio; Rajan, Subramaniam; Blackenhorn, Gunther
2015-01-01
The need for accurate material models to simulate the deformation, damage and failure of polymer matrix composites under impact conditions is becoming critical as these materials are gaining increased usage in the aerospace and automotive industries. While there are several composite material models currently available within commercial transient dynamic finite element codes, several features have been identified as being lacking in the currently available material models that could substantially enhance the predictive capability of the impact simulations. A specific desired feature pertains to the incorporation of both plasticity and damage within the material model. Another desired feature relates to using experimentally based tabulated stress-strain input to define the evolution of plasticity and damage as opposed to specifying discrete input properties (such as modulus and strength) and employing analytical functions to track the response of the material. To begin to address these needs, a combined plasticity and damage model suitable for use with both solid and shell elements is being developed for implementation within the commercial code LS-DYNA. The plasticity model is based on extending the Tsai-Wu composite failure model into a strain-hardening based orthotropic plasticity model with a non-associative flow rule. The evolution of the yield surface is determined based on tabulated stress-strain curves in the various normal and shear directions and is tracked using the effective plastic strain. The effective plastic strain is computed by using the non-associative flow rule in combination with appropriate numerical methods. To compute the evolution of damage, a strain equivalent semi-coupled formulation is used, in which a load in one direction results in a stiffness reduction in multiple coordinate directions. A specific laminated composite is examined to demonstrate the process of characterizing and analyzing the response of a composite using the developed model.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rohatgi, U.S.; Cheng, H.S.; Khan, H.J.
This document is the User`s Manual for the Boiling Water Reactor (BWR), and Simplified Boiling Water Reactor (SBWR) systems transient code RAMONA-4B. The code uses a three-dimensional neutron-kinetics model coupled with a multichannel, nonequilibrium, drift-flux, phase-flow model of the thermal hydraulics of the reactor vessel. The code is designed to analyze a wide spectrum of BWR core and system transients. Chapter 1 gives an overview of the code`s capabilities and limitations; Chapter 2 describes the code`s structure, lists major subroutines, and discusses the computer requirements. Chapter 3 is on code, auxillary codes, and instructions for running RAMONA-4B on Sun SPARCmore » and IBM Workstations. Chapter 4 contains component descriptions and detailed card-by-card input instructions. Chapter 5 provides samples of the tabulated output for the steady-state and transient calculations and discusses the plotting procedures for the steady-state and transient calculations. Three appendices contain important user and programmer information: lists of plot variables (Appendix A) listings of input deck for sample problem (Appendix B), and a description of the plotting program PAD (Appendix C). 24 refs., 18 figs., 11 tabs.« less
Compact storage of medical images with patient information.
Acharya, R; Anand, D; Bhat, S; Niranjan, U C
2001-12-01
Digital watermarking is a technique of hiding specific identification data for copyright authentication. This technique is adapted here for interleaving patient information with medical images to reduce storage and transmission overheads. The text data are encrypted before interleaving with images to ensure greater security. The graphical signals are compressed and subsequently interleaved with the image. Differential pulse-code-modulation and adaptive-delta-modulation techniques are employed for data compression, and encryption and results are tabulated for a specific example.
Computation of Thermally Perfect Compressible Flow Properties
NASA Technical Reports Server (NTRS)
Witte, David W.; Tatum, Kenneth E.; Williams, S. Blake
1996-01-01
A set of compressible flow relations for a thermally perfect, calorically imperfect gas are derived for a value of c(sub p) (specific heat at constant pressure) expressed as a polynomial function of temperature and developed into a computer program, referred to as the Thermally Perfect Gas (TPG) code. The code is available free from the NASA Langley Software Server at URL http://www.larc.nasa.gov/LSS. The code produces tables of compressible flow properties similar to those found in NACA Report 1135. Unlike the NACA Report 1135 tables which are valid only in the calorically perfect temperature regime the TPG code results are also valid in the thermally perfect, calorically imperfect temperature regime, giving the TPG code a considerably larger range of temperature application. Accuracy of the TPG code in the calorically perfect and in the thermally perfect, calorically imperfect temperature regimes are verified by comparisons with the methods of NACA Report 1135. The advantages of the TPG code compared to the thermally perfect, calorically imperfect method of NACA Report 1135 are its applicability to any type of gas (monatomic, diatomic, triatomic, or polyatomic) or any specified mixture of gases, ease-of-use, and tabulated results.
Kotchenova, Svetlana Y; Vermote, Eric F
2007-07-10
This is the second part of the validation effort of the recently developed vector version of the 6S (Second Simulation of a Satellite Signal in the Solar Spectrum) radiative transfer code (6SV1), primarily used for the calculation of look-up tables in the Moderate Resolution Imaging Spectroradiometer (MODIS) atmospheric correction algorithm. The 6SV1 code was tested against a Monte Carlo code and Coulson's tabulated values for molecular and aerosol atmospheres bounded by different Lambertian and anisotropic surfaces. The code was also tested in scalar mode against the scalar code SHARM to resolve the previous 6S accuracy issues in the case of an anisotropic surface. All test cases were characterized by good agreement between the 6SV1 and the other codes: The overall relative error did not exceed 0.8%. The study also showed that ignoring the effects of radiation polarization in the atmosphere led to large errors in the simulated top-of-atmosphere reflectances: The maximum observed error was approximately 7.2% for both Lambertian and anisotropic surfaces.
NASA Astrophysics Data System (ADS)
Kotchenova, Svetlana Y.; Vermote, Eric F.
2007-07-01
This is the second part of the validation effort of the recently developed vector version of the 6S (Second Simulation of a Satellite Signal in the Solar Spectrum) radiative transfer code (6SV1), primarily used for the calculation of look-up tables in the Moderate Resolution Imaging Spectroradiometer (MODIS) atmospheric correction algorithm. The 6SV1 code was tested against a Monte Carlo code and Coulson's tabulated values for molecular and aerosol atmospheres bounded by different Lambertian and anisotropic surfaces. The code was also tested in scalar mode against the scalar code SHARM to resolve the previous 6S accuracy issues in the case of an anisotropic surface. All test cases were characterized by good agreement between the 6SV1 and the other codes: The overall relative error did not exceed 0.8%. The study also showed that ignoring the effects of radiation polarization in the atmosphere led to large errors in the simulated top-of-atmosphere reflectances: The maximum observed error was approximately 7.2% for both Lambertian and anisotropic surfaces.
Comparison of EGS4 and MCNP Monte Carlo codes when calculating radiotherapy depth doses.
Love, P A; Lewis, D G; Al-Affan, I A; Smith, C W
1998-05-01
The Monte Carlo codes EGS4 and MCNP have been compared when calculating radiotherapy depth doses in water. The aims of the work were to study (i) the differences between calculated depth doses in water for a range of monoenergetic photon energies and (ii) the relative efficiency of the two codes for different electron transport energy cut-offs. The depth doses from the two codes agree with each other within the statistical uncertainties of the calculations (1-2%). The relative depth doses also agree with data tabulated in the British Journal of Radiology Supplement 25. A discrepancy in the dose build-up region may by attributed to the different electron transport algorithims used by EGS4 and MCNP. This discrepancy is considerably reduced when the improved electron transport routines are used in the latest (4B) version of MCNP. Timing calculations show that EGS4 is at least 50% faster than MCNP for the geometries used in the simulations.
NASA Technical Reports Server (NTRS)
Zehe, Michael J.; Gordon, Sanford; McBride, Bonnie J.
2002-01-01
For several decades the NASA Glenn Research Center has been providing a file of thermodynamic data for use in several computer programs. These data are in the form of least-squares coefficients that have been calculated from tabular thermodynamic data by means of the NASA Properties and Coefficients (PAC) program. The source thermodynamic data are obtained from the literature or from standard compilations. Most gas-phase thermodynamic functions are calculated by the authors from molecular constant data using ideal gas partition functions. The Coefficients and Properties (CAP) program described in this report permits the generation of tabulated thermodynamic functions from the NASA least-squares coefficients. CAP provides considerable flexibility in the output format, the number of temperatures to be tabulated, and the energy units of the calculated properties. This report provides a detailed description of input preparation, examples of input and output for several species, and a listing of all species in the current NASA Glenn thermodynamic data file.
CAP: A Computer Code for Generating Tabular Thermodynamic Functions from NASA Lewis Coefficients
NASA Technical Reports Server (NTRS)
Zehe, Michael J.; Gordon, Sanford; McBride, Bonnie J.
2001-01-01
For several decades the NASA Glenn Research Center has been providing a file of thermodynamic data for use in several computer programs. These data are in the form of least-squares coefficients that have been calculated from tabular thermodynamic data by means of the NASA Properties and Coefficients (PAC) program. The source thermodynamic data are obtained from the literature or from standard compilations. Most gas-phase thermodynamic functions are calculated by the authors from molecular constant data using ideal gas partition functions. The Coefficients and Properties (CAP) program described in this report permits the generation of tabulated thermodynamic functions from the NASA least-squares coefficients. CAP provides considerable flexibility in the output format, the number of temperatures to be tabulated, and the energy units of the calculated properties. This report provides a detailed description of input preparation, examples of input and output for several species, and a listing of all species in the current NASA Glenn thermodynamic data file.
A new catalog of planetary maps
NASA Technical Reports Server (NTRS)
Batson, R. M.; Inge, J. L.
1991-01-01
A single, concise reference to all existing planetary maps, including lunar ones, is being prepared that will allow map users to identify and locate maps of their areas of interest. This will be the first such comprehensive listing of planetary maps. Although the USGS shows index maps on the collar of each map sheet, periodically publishes index maps of Mars, and provides informal listings of the USGS map database, no tabulation exists that identifies all planetary maps, including those published by DMA and other organizations. The catalog will consist of a booklet containing small-scale image maps with superimposed quadrangle boundaries and map data tabulations.
DOT National Transportation Integrated Search
1980-01-01
This publication brings together annual series of selected statistical tabulations relating to highway transportation in three major areas: (1) Highway use--the ownership and operation of motor vehicles; (2) highway finance--the receipts and expendit...
DOT National Transportation Integrated Search
2007-01-01
This publication brings together annual series of selected statistical tabulations relating to highway transportation in four major areas: (1) highway infrastructure and users; (2) motor vehicles and fuel; (3) highway finance--the receipts and expend...
Improving Hospital Reporting of Patient Race and Ethnicity--Approaches to Data Auditing.
Zingmond, David S; Parikh, Punam; Louie, Rachel; Lichtensztajn, Daphne Y; Ponce, Ninez; Hasnain-Wynia, Romana; Gomez, Scarlett Lin
2015-08-01
To investigate new metrics to improve the reporting of patient race and ethnicity (R/E) by hospitals. California Patient Discharge Database (PDD) and birth registry, 2008-2009, Healthcare and Cost Utilization Project's State Inpatient Database, 2008-2011, cancer registry 2000-2008, and 2010 US Census Summary File 2. We examined agreement between hospital reported R/E versus self-report among mothers delivering babies and a cancer cohort in California. Metrics were created to measure root mean squared differences (RMSD) by hospital between reported R/E distribution and R/E estimates using R/E distribution within each patient's zip code of residence. RMSD comparisons were made to corresponding "gold standard" facility-level measures within the maternal cohort for California and six comparison states. Maternal birth hospitalization (linked to the state birth registry) and cancer cohort records linked to preceding and subsequent hospitalizations. Hospital discharges were linked to the corresponding Census zip code tabulation area using patient zip code. Overall agreement between the PDD and the gold standard for the maternal cohort was 86 percent for the combined R/E measure and 71 percent for race alone. The RMSD measure is modestly correlated with the summary level gold standard measure for R/E (r = 0.44). The RMSD metric revealed general improvement in data agreement and completeness across states. "Other" and "unknown" categories were inconsistently applied within inpatient databases. Comparison between reported R/E and R/E estimates using zip code level data may be a reasonable first approach to evaluate and track hospital R/E reporting. Further work should focus on using more granular geocoded data for estimates and tracking data to improve hospital collection of R/E data. © Health Research and Educational Trust.
Yasaitis, Laura C; Pajerowski, William; Polsky, Daniel; Werner, Rachel M
2016-08-01
Early evidence suggested that accountable care organizations (ACOs) could improve health care quality while constraining costs, and ACOs are expanding throughout the United States. However, if disadvantaged patients have unequal access to physicians who participate in ACOs, that expansion may exacerbate health care disparities. We examined the relationship between physicians' participation in both Medicare and commercial ACOs across the country and the sociodemographic characteristics of their likely patient populations. Physicians' participation in ACOs varied widely across hospital referral regions, from nearly 0 percent to over 85 percent. After we adjusted for individual physician and practice characteristics, we found that physicians who practiced in ZIP Code Tabulation Areas where a higher percentage of the population was black, living in poverty, uninsured, or disabled or had less than a high school education-compared to other areas-had significantly lower rates of ACO participation than other physicians. Our findings suggest that vulnerable populations' access to physicians participating in ACOs may not be as great as access for other groups, which could exacerbate existing disparities in health care quality. Project HOPE—The People-to-People Health Foundation, Inc.
The Epidemiology of Vascular Injury in the Wars in Iraq and Afghanistan
2011-06-01
scale (AIS) and In- ternational Classification of Diseases , Ninth Revision (ICD-9) codes for vascular injury (arterial and venous) and vascular injury...denominator of significant wounding in the tabulation of rates. Nonbattle-related injuries (ie, disease nonbattle or DNBI) were not included in the...Coronary 2 0.13 Celiac 3 0.19 Superior mesenteric artery 13 0.83 Aorta 45 2.9 Vena cava (n = 21) Superior 5 0.32 Inferior 16 1.1 Iliac (n = 61) Iliac
The woody biomass resource of Tennessee, 1989
James F. Rosson
1993-01-01
Tabulates fresh and dry biomass estimates of major trees in Tennessee by forest type, ownership, species, stand basal area, tree class, diameter, and height. Information is presented for total tree, stem, and crown components.
The woody biomass resource of Louisiana, 1991
James F. Rosson
1993-01-01
Tabulates fresh and dry biomass estimates of major trees in Louisiana by forest type, ownership, species, stand basal area, tree class, diameter, and height. Information is presented for total tree, stem, and crown components.
Do socially deprived urban areas have lesser supplies of cancer care services?
Lamont, Elizabeth B; He, Yulei; Subramanian, S V; Zaslavsky, Alan M
2012-09-10
Area social deprivation is associated with unfavorable health outcomes of residents across the full clinical course of cancer from the stage at diagnosis through survival. We sought to determine whether area social factors are associated with the area health care supply. We studied the area supply of health services required for the provision of guideline-recommended care for patients with breast cancer and colorectal cancer (CRC) in each of the following three distinct clinical domains: screening, treatment, and post-treatment surveillance. We characterized area social factors in 3,096 urban zip code tabulation areas by using Census Bureau data and the health care supply in the corresponding 465 hospital service areas by using American Hospital Association, American Medical Association, and US Food and Drug Administration data. In two-level hierarchical models, we assessed associations between social factors and the supply of health services across areas. We found no clear associations between area social factors and the supply of health services essential to the provision of guideline recommended breast cancer and CRC care in urban areas. The measures of health service included the supply of physicians who facilitate screening, treatment, and post-treatment care and the supply of facilities required for the same services. Because we found that the supply of types of health care required for the provision of guideline-recommended cancer care for patients with breast cancer and CRC did not vary with markers of area socioeconomic disadvantage, it is possible that previously reported unfavorable breast cancer and CRC outcomes among individuals living in impoverished areas may have occurred despite an apparent adequate area health care supply.
SOSPAC- SOLAR SPACE POWER ANALYSIS CODE
NASA Technical Reports Server (NTRS)
Selcuk, M. K.
1994-01-01
The Solar Space Power Analysis Code, SOSPAC, was developed to examine the solar thermal and photovoltaic power generation options available for a satellite or spacecraft in low earth orbit. SOSPAC is a preliminary systems analysis tool and enables the engineer to compare the areas, weights, and costs of several candidate electric and thermal power systems. The configurations studied include photovoltaic arrays and parabolic dish systems to produce electricity only, and in various combinations to provide both thermal and electric power. SOSPAC has been used for comparison and parametric studies of proposed power systems for the NASA Space Station. The initial requirements are projected to be about 40 kW of electrical power, and a similar amount of thermal power with temperatures above 1000 degrees Centigrade. For objects in low earth orbit, the aerodynamic drag caused by suitably large photovoltaic arrays is very substantial. Smaller parabolic dishes can provide thermal energy at a collection efficiency of about 80%, but at increased cost. SOSPAC allows an analysis of cost and performance factors of five hybrid power generating systems. Input includes electrical and thermal power requirements, sun and shade durations for the satellite, and unit weight and cost for subsystems and components. Performance equations of the five configurations are derived, and the output tabulates total weights of the power plant assemblies, area of the arrays, efficiencies, and costs. SOSPAC is written in FORTRAN IV for batch execution and has been implemented on an IBM PC computer operating under DOS with a central memory requirement of approximately 60K of 8 bit bytes. This program was developed in 1985.
Forest statistics of western Kentucky
The Forest Survey Organization Central States Forest Experiment Station
1950-01-01
This Survey Release presents the more significant preliminary statistics on the forest area and timber volume for the western region of Kentucky. Similar reports for the remainder of the state will be published as soon as statistical tabulations are completed. Later, an analytical report for the state will be published which will interpret forest area, timber volume,...
Wood fuel potential from harvested areas in the eastern United States.
Eugene M. Carpenter
1980-01-01
Estimates amount of wood fiber that could be available for fuel from forest residues on harvested areas in the eastern United States. Includes a key to resource data published by the USDA Forest Service and factors for estimating amounts of cull, bark, tops, and limbs from inventory and product output tabulations.
Hydrologic data for the Larimer-Weld regional water-monitoring program, Colorado, 1975-82
Blakely, S.R.; Steinheimer, J.T.
1984-01-01
The Larimer-Weld, Colorado, regional Monitoring Program was begun in 1976 to provide information on the quality and quantity of the surface-water resources in the area. Three stations on the big Thompson River and five stations on the Cache La Poudre River were selected for a data-collection network. Four previously established stations were added to complete the data-collection network: Horsetooth Reservoir, Joe Wright Creek above and below Joe Wright Reservoir, and Michigan River near Cameron Pass. Station description, location, and period of record are given for each station. A statistical summary of the water-quality data for each station is tabulated. Frequency of occurrence is given at the 95th, 75th, 50th, and 25th percentiles. Monthly water-quality data and daily average streamflow data are tabulated for each streamflow station for which this data was collected; Monthly contents data are presented for Horsetooth Reservoir. All data tabulated and summarized are from the period October 1, 1975, through September 30, 1982. (USGS)
Energy data report: Solar collector manufacturing activity, July - December 1980
NASA Astrophysics Data System (ADS)
1981-03-01
Statistics on solar collector manufacturing activity for both solar thermal collectors and photovoltaic modules through 1980 are presented. Summary data are given for the number of manufacturers and collector area produced each year from 1974 through 1980; data for collector type are included, i.e., low temperature or medium temperature and special collectors. Producer shipments are tabulated according to location of company headquarters, producer size, and collector types. headquarters The number of companies engaged in activities related to solar collector manufacturing for 1978 through 1980 are listed; and the number of manufacturers and market sector are tabulated for photovoltaic modules manufacturing activities.
Small-area Variation in Hypertension Prevalence among Black and White Medicaid Enrollees.
White, Kellee; Stewart, John E; Lòpez-DeFede, Ana; Wilkerson, Rebecca C
2016-07-21
To examine within-state geographic heterogeneity in hypertension prevalence and evaluate associations between hypertension prevalence and small-area contextual characteristics for Black and White South Carolina Medicaid enrollees in urban vs rural areas. Ecological. South Carolina, United States. Hypertension prevalence. Data representing adult South Carolina Medicaid recipients enrolled in fiscal year 2013 (N=409,907) and ZIP Code Tabulation Area (ZCTA)-level contextual measures (racial segregation, rurality, poverty, educational attainment, unemployment and primary care physician adequacy) were linked in a spatially referenced database. Optimized Getis-Ord hotspot mapping was used to visualize geographic clustering of hypertension prevalence. Spatial regression was performed to examine the association between hypertension prevalence and small-area contextual indicators. Significant (alpha=.05) hotspot spatial clustering patterns were similar for Blacks and Whites. Black isolation was significantly associated with hypertension among Blacks and Whites in both urban (Black, b=1.34, P<.01; White, b=.66, P<.01) and rural settings (Black, b=.71, P=.02; White, b=.70, P<.01). Primary care physician adequacy was associated with hypertension among urban Blacks (b=-2.14, P<.01) and Whites (b=-1.74, P<.01). The significant geographic overlap of hypertension prevalence hotspots for Black and White Medicaid enrollees provides an opportunity for targeted health intervention. Provider adequacy findings suggest the value of ACA network adequacy standards for Medicaid managed care plans in ensuring health care accessibility for persons with hypertension and related chronic conditions.
An analysis of MCNP cross-sections and tally methods for low-energy photon emitters.
Demarco, John J; Wallace, Robert E; Boedeker, Kirsten
2002-04-21
Monte Carlo calculations are frequently used to analyse a variety of radiological science applications using low-energy (10-1000 keV) photon sources. This study seeks to create a low-energy benchmark for the MCNP Monte Carlo code by simulating the absolute dose rate in water and the air-kerma rate for monoenergetic point sources with energies between 10 keV and 1 MeV. The analysis compares four cross-section datasets as well as the tally method for collision kerma versus absorbed dose. The total photon attenuation coefficient cross-section for low atomic number elements has changed significantly as cross-section data have changed between 1967 and 1989. Differences of up to 10% are observed in the photoelectric cross-section for water at 30 keV between the standard MCNP cross-section dataset (DLC-200) and the most recent XCOM/NIST tabulation. At 30 keV, the absolute dose rate in water at 1.0 cm from the source increases by 7.8% after replacing the DLC-200 photoelectric cross-sections for water with those from the XCOM/NIST tabulation. The differences in the absolute dose rate are analysed when calculated with either the MCNP absorbed dose tally or the collision kerma tally. Significant differences between the collision kerma tally and the absorbed dose tally can occur when using the DLC-200 attenuation coefficients in conjunction with a modern tabulation of mass energy-absorption coefficients.
Calculating sediment discharge from a highway construction site in central Pennsylvania
Reed, L.A.; Ward, J.R.; Wetzel, K.L.
1985-01-01
The Pennsylvania Department of Transportation, the Federal Highway Administration, and the U.S. Geological Survey have cooperated in a study to evaluate two methods of predicting sediment yields during highway construction. Sediment yields were calculated using the Universal Soil Loss and the Younkin Sediment Prediction Equations. Results were compared to the actual measured values, and standard errors and coefficients of correlation were calculated. Sediment discharge from the construction area was determined for storms that occurred during construction of Interstate 81 in a 0.38-square mile basin near Harrisburg, Pennsylvania. Precipitation data tabulated included total rainfall, maximum 30-minute rainfall, kinetic energy, and the erosive index of the precipitation. Highway construction data tabulated included the area disturbed by clearing and grubbing, the area in cuts and fills, the average depths of cuts and fills, the area seeded and mulched, and the area paved. Using the Universal Soil Loss Equation, sediment discharge from the construction area was calculated for storms. The standard error of estimate was 0.40 (about 105 percent), and the coefficient of correlation was 0.79. Sediment discharge from the construction area was also calculated using the Younkin Equation. The standard error of estimate of 0.42 (about 110 percent), and the coefficient of correlation of 0.77 are comparable to those from the Universal Soil Loss Equation.
Lòpez-De Fede, Ana; Stewart, John E; Hardin, James W; Mayfield-Smith, Kathy
2016-06-10
Measures of small-area deprivation may be valuable in geographically targeting limited resources to prevent, diagnose, and effectively manage chronic conditions in vulnerable populations. We developed a census-based small-area socioeconomic deprivation index specifically to predict chronic disease burden among publically insured Medicaid recipients in South Carolina, a relatively poor state in the southern United States. We compared the predictive ability of the new index with that of four other small-area deprivation indicators. To derive the ZIP Code Tabulation Area-Level Palmetto Small-Area Deprivation Index (Palmetto SADI), we evaluated ten census variables across five socioeconomic deprivation domains, identifying the combination of census indicators most highly correlated with a set of five chronic disease conditions among South Carolina Medicaid enrollees. In separate validation studies, we used both logistic and spatial regression methods to assess the ability of Palmetto SADI to predict chronic disease burden among state Medicaid recipients relative to four alternative small-area socioeconomic deprivation measures: the Townsend index of material deprivation; a single-variable poverty indicator; and two small-area designations of health care resource deprivation, Primary Care Health Professional Shortage Area and Medically Underserved Area/Medically Underserved Population. Palmetto SADI was the best predictor of chronic disease burden (presence of at least one condition and presence of two or more conditions) among state Medicaid recipients compared to all alternative deprivation measures tested. A low-cost, regionally optimized socioeconomic deprivation index, Palmetto SADI can be used to identify areas in South Carolina at high risk for chronic disease burden among Medicaid recipients and other low-income Medicaid-eligible populations for targeted prevention, screening, diagnosis, disease self-management, and care coordination activities.
2011-01-01
Background Nurse managers have the burden of experiencing frequent ethical issues related to both their managerial and nursing care duties, according to previous international studies. However, no such study was published in Malaysia. The purpose of this study was to explore nurse managers' experience with ethical issues in six government hospitals in Malaysia including learning about the way they dealt with the issues. Methods A cross-sectional study was conducted in August-September, 2010 involving 417 (69.2%) of total 603 nurse managers in the six Malaysian government hospitals. Data were collected using three-part self-administered questionnaire. Part I was regarding participants' demographics. Part II was about the frequency and areas of management where ethical issues were experienced, and scoring of the importance of 11 pre-identified ethical issues. Part III asked how they dealt with ethical issues in general; ways to deal with the 11 pre-identified ethical issues, and perceived stress level. Data were analyzed using descriptive statistics, cross-tabulations and Pearson's Chi-square. Results A total of 397 (95.2%) participants experienced ethical issues and 47.2% experienced them on weekly to daily basis. Experiencing ethical issues were not associated with areas of practice. Top area of management where ethical issues were encountered was "staff management", but "patient care" related ethical issues were rated as most important. Majority would "discuss with other nurses" in dealing generally with the issues. For pre-identified ethical issues regarding "patient care", "discuss with doctors" was preferred. Only 18.1% referred issues to "ethics committees" and 53.0% to the code of ethics. Conclusions Nurse managers, regardless of their areas of practice, frequently experienced ethical issues. For dealing with these, team-approach needs to be emphasized. Proper understanding of the code of ethics is needed to provide basis for reasoning. PMID:22085735
Stephens, J Mark; Bensink, Mark; Bowers, Charles; Hollenbeak, Christopher S
2017-07-31
Prophylaxis with granulocyte colony-stimulating factors (G-CSFs) is recommended for patients receiving myelosuppressive chemotherapy regimens with a high risk of febrile neutropenia (FN). G-CSFs should be administered starting the day after chemotherapy, necessitating return trips to the oncology clinic at the end of each cycle. We examined the travel burden related to prophylactic G-CSF injections after chemotherapy in the US. We used 2012-2014 Medicare claims data to identify a national cohort of beneficiaries age 65+ with non-myeloid cancers who received both chemotherapy and prophylactic G-CSFs. Patient travel origin was based on residence ZIP code. Oncologist practice locations and hospital addresses were obtained from the Medicare Physician Compare and Hospital Compare websites and geocoded using the Google Maps Application Programming Interface (API). Driving distance and time to the care site from each patient ZIP code tabulation area (ZCTA) were calculated using Open Street Maps road networks. Geographic and socio-economic characteristics of each ZCTA from the US Census Bureau's American Community Survey were used to stratify and analyze travel estimates. The mean one-way driving distance to the G-CSF provider was 23.8 (SD 30.1) miles and the mean one-way driving time was 33.3 (SD 37.8) minutes. When stratified by population density, the mean one-way travel time varied from 12.1 (SD 10.1) minutes in Very Dense Urban areas to 76.7 (SD 72.1) minutes in Super Rural areas. About 48% of patients had one-way travel times of <20 minutes, but 19% of patients traveled ≥50 minutes one way for G-CSF prophylaxis. Patients in areas with above average concentrations of aged, poor or disabled residents were more likely to experience longer travel. Administration of G-CSF therapy after chemotherapy can present a significant travel burden for cancer patients. Technological improvements in the form and methods of drug delivery for G-CSFs might significantly reduce this travel burden.
Use of Tabulated Thermochemical Data for Pure Compounds
NASA Technical Reports Server (NTRS)
Jacobson, Nathan S.
1999-01-01
Thermodynamic data for inorganic compounds is found in a variety of tabulations and computer databases. An extensive listing of sources of inorganic thermodynamic data is provided. The three major tabulations are the JANAF tables. Thermodynamic Properties of Individual Substances, and the tabulation by Barin. The notation and choice of standard states is different in each of these tabulations, so combining data from the different tabulations is often a problem. By understanding the choice of standard states, it is possible to develop simple equations for conversion of the data from one form to another.
Performance comparison of AV1, HEVC, and JVET video codecs on 360 (spherical) video
NASA Astrophysics Data System (ADS)
Topiwala, Pankaj; Dai, Wei; Krishnan, Madhu; Abbas, Adeel; Doshi, Sandeep; Newman, David
2017-09-01
This paper compares the coding efficiency performance on 360 videos, of three software codecs: (a) AV1 video codec from the Alliance for Open Media (AOM); (b) the HEVC Reference Software HM; and (c) the JVET JEM Reference SW. Note that 360 video is especially challenging content, in that one codes full res globally, but typically looks locally (in a viewport), which magnifies errors. These are tested in two different projection formats ERP and RSP, to check consistency. Performance is tabulated for 1-pass encoding on two fronts: (1) objective performance based on end-to-end (E2E) metrics such as SPSNR-NN, and WS-PSNR, currently developed in the JVET committee; and (2) informal subjective assessment of static viewports. Constant quality encoding is performed with all the three codecs for an unbiased comparison of the core coding tools. Our general conclusion is that under constant quality coding, AV1 underperforms HEVC, which underperforms JVET. We also test with rate control, where AV1 currently underperforms the open source X265 HEVC codec. Objective and visual evidence is provided.
Clinical Data Interchange Standards Consortium (CDISC) is an international, non-profit organization that develops and supports global data standards for medical research. CDISC is working actively with EVS to develop and support controlled terminology in several areas, notably CDISC's Study Data Tabulation Model (SDTM).
Early Yields Of Slash Pine Planted On a Cutover Site At Various Spacings
W.F. Mann
1971-01-01
Tabulates basal areas, cordwood and cubic-foot volumes, average d.b.h., and diameter distributions for 14-year-old slash pine planted in central Louisiana. Also gives regression equations developed to predict these parameters.
1980-08-01
Compression factor and error sensitivity together with statistical data have also been tabulated. This TIB is a companion drcument to NCS TIB’s 79-7...vu donner la priorit6 pour lour r~alisation. Chaque application est conf ice A un " chef do projet", responsable successivoment do sa conception. de son...pilote depend des r~sultats obtenus et fait I’objet d’une d~cision- de ’.a Direction Gdnerale. Ndanmoins, le chef do projet doit dOs le d~part consid~rer
Software quality: Process or people
NASA Technical Reports Server (NTRS)
Palmer, Regina; Labaugh, Modenna
1993-01-01
This paper will present data related to software development processes and personnel involvement from the perspective of software quality assurance. We examine eight years of data collected from six projects. Data collected varied by project but usually included defect and fault density with limited use of code metrics, schedule adherence, and budget growth information. The data are a blend of AFSCP 800-14 and suggested productivity measures in Software Metrics: A Practioner's Guide to Improved Product Development. A software quality assurance database tool, SQUID, was used to store and tabulate the data.
Konrad, Christopher; Sevier, Maria
2014-01-01
Geospatial information for the active streamflow gaging network in the Puget Sound Basin was compiled to support regional monitoring of stormwater effects to small streams. The compilation includes drainage area boundaries and physiographic and land use attributes that affect hydrologic processes. Three types of boundaries were used to tabulate attributes: Puget Sound Watershed Characterization analysis units (AU); the drainage area of active streamflow gages; and the catchments of Regional Stream Monitoring Program (RSMP) sites. The active streamflow gaging network generally includes sites that represent the ranges of attributes for lowland AUs, although there are few sites with low elevations (less than 60 meters), low precipitation (less than 1 meter year), or high stream density (greater than 5 kilometers per square kilometers). The active streamflow gaging network can serve to provide streamflow information in some AUs and RSMP sites, particularly where the streamflow gage measures streamflow generated from a part of the AU or that drains to the RSMP site, and that part of the AU or RSMP site is a significant fraction of the drainage area of the streamgage. The maximum fraction of each AU or RSMP catchment upstream of a streamflow gage and the maximum fraction of any one gaged basin in an AU or RSMP along with corresponding codes are provided in the attribute tables.
The woody biomass resource of major tree taxa for the Midsouth States
James F. Rosson
1992-01-01
Fresh and dry biomass estimates of major trees in seven Midsouth States by forest type, ownership, species, stand basal area, tree class, diameter, and height are tabulated. Information is presented for total tree, bole, and crown components.
Ambient noise level measurements in proposed Florida airport area
DOT National Transportation Integrated Search
1972-12-31
This report documents the measurement results made at ten locations near the three remaining sites being studied for the "South Florida Regional Report." Tabulated data display a summary of the measured noise levels at each location expressed as nois...
The Influence of Stand Density and Structure on Growth of Northern Hardwoods in New England
Dale S. Solomon
1977-01-01
Growth of northern hardwoods over a 10-year period was studied in plots that were treated to produce residual densities of 40, 60, 80, and 100 square feet of basal area per acre with stand structures of 30, 45 and 60 percent sawtimber. Both diameter and basal area growth are tabulated by treatment and species.
A Survey of Radio Listenership in the Davao Provinces of Mindanao, The Philippines.
ERIC Educational Resources Information Center
Spain, Peter
To identify the listening habits and preferences of the audience of a radio-braodcasting station, DXCD, in a rural area of the Philippines, a questionnaire-based survey was conducted. A total sample of 960 respondents was interviewed. The methodology of the survey is presented, and the demographics of the area described. Results are tabulated in…
48 CFR 908.7117 - Tabulating machine cards.
Code of Federal Regulations, 2011 CFR
2011-10-01
... 48 Federal Acquisition Regulations System 5 2011-10-01 2011-10-01 false Tabulating machine cards. 908.7117 Section 908.7117 Federal Acquisition Regulations System DEPARTMENT OF ENERGY COMPETITION... Tabulating machine cards. DOE offices shall acquire tabulating machine cards in accordance with FPMR 41 CFR...
48 CFR 908.7117 - Tabulating machine cards.
Code of Federal Regulations, 2010 CFR
2010-10-01
... 48 Federal Acquisition Regulations System 5 2010-10-01 2010-10-01 false Tabulating machine cards. 908.7117 Section 908.7117 Federal Acquisition Regulations System DEPARTMENT OF ENERGY COMPETITION... Tabulating machine cards. DOE offices shall acquire tabulating machine cards in accordance with FPMR 41 CFR...
2012-05-22
tabulation of the reduced space is performed using the In Situ Adaptive Tabulation ( ISAT ) algorithm. In addition, we use x2f mpi – a Fortran library...for parallel vector-valued function evaluation (used with ISAT in this context) – to efficiently redistribute the chemistry workload among the...Constrained-Equilibrium (RCCE) method, and tabulation of the reduced space is performed using the In Situ Adaptive Tabulation ( ISAT ) algorithm. In addition
1989-09-30
26 QUESTIONNAIRE INSTRUMENT ri. -I., DATA TABULATION VOLUMES This material provides information for use by readers to interpret...The second longitudinal Tabulation Volume reports the 1988 questionnaire responses of the junior enlisted "stayers" who were used as the sample to...the specific crossing variables used for the cross-sectional and longitudinal Tabulation Volumes. Cross-Sectional Tabulation Volumes. Demographic
Geography:The TIGER Line Files are feature classes and related database files (.) that are an extract of selected geographic and cartographic information from the U.S. Census Bureau's Master Address File / Topologically Integrated Geographic Encoding and Referencing (MAF/TIGER) Database (MTDB). The MTDB represents a seamless national file with no overlaps or gaps between parts, however, each TIGER Line File is designed to stand alone as an independent data set, or they can be combined to cover the entire nation. Census Blocks are statistical areas bounded on all sides by visible features, such as streets, roads, streams, and railroad tracks, and/or by non visible boundaries such as city, town, township, and county limits, and short line-of-sight extensions of streets and roads. Census blocks are relatively small in area; for example, a block in a city bounded by streets. However, census blocks in remote areas are often large and irregular and may even be many square miles in area. A common misunderstanding is that data users think census blocks are used geographically to build all other census geographic areas, rather all other census geographic areas are updated and then used as the primary constraints, along with roads and water features, to delineate the tabulation blocks. As a result, all 2010 Census blocks nest within every other 2010 Census geographic area, so that Census Bureau statistical data can be tabulated at the block level and aggregated up t
Upper Kalamazoo watershed land cover inventory. [based on remote sensing
NASA Technical Reports Server (NTRS)
Richason, B., III; Enslin, W.
1973-01-01
Approximately 1000 square miles of the eastern portion of the watershed were inventoried based on remote sensing imagery. The classification scheme, imagery and interpretation procedures, and a cost analysis are discussed. The distributions of land cover within the area are tabulated.
NASA Technical Reports Server (NTRS)
Talcott, N. A., Jr.
1977-01-01
Equations and computer code are given for the thermodynamic properties of gaseous fluorocarbons in chemical equilibrium. In addition, isentropic equilibrium expansions of two binary mixtures of fluorocarbons and argon are included. The computer code calculates the equilibrium thermodynamic properties and, in some cases, the transport properties for the following fluorocarbons: CCl2F, CCl2F2, CBrF3, CF4, CHCl2F, CHF3, CCL2F-CCl2F, CCLF2-CClF2, CF3-CF3, and C4F8. Equilibrium thermodynamic properties are tabulated for six of the fluorocarbons(CCl3F, CCL2F2, CBrF3, CF4, CF3-CF3, and C4F8) and pressure-enthalpy diagrams are presented for CBrF3.
Yasaitis, Laura C.; Pajerowski, William; Polsky, Daniel; Werner, Rachel M.
2016-01-01
Early evidence suggested that accountable care organizations (ACOs) could improve health care quality while constraining costs, and ACOs are expanding throughout the United States. However, if disadvantaged patients have unequal access to physicians who participate in ACOs, that expansion may exacerbate health care disparities. We examined the relationship between physician participation in both Medicare and commercial ACOs across the country and the sociodemographic characteristics of their likely patient populations. Physician participation in ACOs varied widely across hospital referral regions, from nearly 0 percent to over 85 percent. After we adjusted for individual physician and practice characteristics, we found that physicians who practiced in ZIP Code Tabulation Areas where a higher percentage of the population was black, living in poverty, uninsured, or disabled or had less than a high school education—compared to other areas—had significantly lower rates of ACO participation than other physicians. Our findings suggest that vulnerable populations may not have as great access as other groups to physicians participating in ACOs, which could exacerbate existing disparities in health care quality. PMID:27503961
NASA Astrophysics Data System (ADS)
Wulder, M. A.
1998-03-01
Forest stand data are normally stored in a geographic information system (GIS) on the basis of areas of similar species combinations. Polygons are created based upon species assemblages and given labels relating the percentage of areal coverage by each significant species type within the specified area. As a result, estimation of leaf area index (LAI) from the digital numbers found within GIS-stored polygons lack accuracy as the predictive equations for LAI are normally developed for individual species, not species assemblages. A Landsat TM image was acquired to enable a classification which allows for the decomposition of forest-stand polygons into greater species detail. Knowledge of the actual internal composition of the stand polygons provides for computation of LAI values based upon the appropriate predictive equation resulting in higher accuracy of these estimates. To accomplish this goal it was necessary to extract, for each cover type in each polygon, descriptive values to represent the digital numbers located in that portion of the polygon. The classified image dictates the species composition of the various portions of the polygon and within these areas the raster pixel values are tabulated and averaged. Due to a lack of existing software tools to assess the raster values occurring within GIS polygons a combination of remote sensing, GIS, UNIX, and specifically coded C programs were necessary. Such tools are frequently used by the spatial analyst and indicate the complexity of what may appear to be a straight-forward spatial analysis problem.
Rural AIDS Diagnoses in Florida: Changing Demographics and Factors Associated With Survival
Trepka, Mary Jo; Niyonsenga, Theophile; Maddox, Lorene M.; Lieb, Spencer
2012-01-01
Purpose To compare demographic characteristics and predictors of survival of rural residents diagnosed with acquired immunodeficiency syndrome (AIDS) with those of urban residents. Methods Florida surveillance data for people diagnosed with AIDS during 1993–2007 were merged with 2000 Census data using ZIP code tabulation areas (ZCTA). Rural status was classified based on the ZCTA’s rural-urban commuting area classification. Survival rates were compared between rural and urban areas using survival curves and Cox proportional hazards models controlling for demographic, clinical, and area-level socioeconomic and health care access factors. Findings Of the 73,590 people diagnosed with AIDS, 1,991 (2.7%) resided in rural areas. People in the most recent rural cohorts were more likely than those in earlier cohorts to be female, non-Hispanic black, older, and have a reported transmission mode of heterosexual sex. There were no statistically significant differences in the 3-, 5-, or 10-year survival rates between rural and urban residents. Older age at the time of diagnosis, diagnosis during the 1993–1995 period, other/unknown transmission mode, and lower CD4 count/percent categories were associated with lower survival in both rural and urban areas. In urban areas only, being non-Hispanic black or Hispanic, being US born, more poverty, less community social support, and lower physician density were also associated with lower survival. Conclusions In rural Florida, the demographic characteristics of people diagnosed with AIDS have been changing, which may necessitate modifications in the delivery of AIDS-related services. Rural residents diagnosed with AIDS did not have a significant survival disadvantage relative to urban residents. PMID:23802929
41 CFR 101-26.509 - Tabulating machine cards.
Code of Federal Regulations, 2011 CFR
2011-07-01
... 41 Public Contracts and Property Management 2 2011-07-01 2007-07-01 true Tabulating machine cards... PROGRAM 26.5-GSA Procurement Programs § 101-26.509 Tabulating machine cards. Procurement by Federal agencies of tabulating machine cards shall be made in accordance with the provisions of this § 101-26.509...
Coady, Sean A; Johnson, Norman J; Hakes, Jahn K; Sorlie, Paul D
2014-07-09
The Medicare program provides universal access to hospital care for the elderly; however, mortality disparities may still persist in this population. The association of individual education and area income with survival and recurrence post Myocardial Infarction (MI) was assessed in a national sample. Individual level education from the National Longitudinal Mortality Study was linked to Medicare and National Death Index records over the period of 1991-2001 to test the association of individual education and zip code tabulation area median income with survival and recurrence post-MI. Survival was partitioned into 3 periods: in-hospital, discharge to 1 year, and 1 year to 5 years and recurrence was partitioned into two periods: 28 day to 1 year, and 1 year to 5 years. First MIs were found in 8,043 women and 7,929 men. In women and men 66-79 years of age, less than a high school education compared with a college degree or more was associated with 1-5 year mortality in both women (HRR 1.61, 95% confidence interval 1.03-2.50) and men (HRR 1.37, 1.06-1.76). Education was also associated with 1-5 year recurrence in men (HRR 1.68, 1.18-2.41, < High School compared with college degree or more), but not women. Across the spectrum of survival and recurrence periods median zip code level income was inconsistently associated with outcomes. Associations were limited to discharge-1 year survival (RR lowest versus highest quintile 1.31, 95% confidence interval 1.03-1.67) and 28 day-1 year recurrence (RR lowest versus highest quintile 1.72, 95% confidence interval 1.14-2.57) in older men. Despite the Medicare entitlement program, disparities related to individual socioeconomic status remain. Additional research is needed to elucidate the barriers and mechanisms to eliminating health disparities among the elderly.
Henry, Kevin A; Swiecki-Sikora, Allison L; Stroup, Antoinette M; Warner, Echo L; Kepka, Deanna
2017-07-14
This study is the first to examine associations between several area-based socioeconomic factors and human papillomavirus (HPV) vaccine uptake among boys in the United States (U.S.). Data from the 2012-2013 National Immunization Survey-Teen restricted-use data were analyzed to examine associations of HPV vaccination initiation (receipt of ≥1 dose) and series completion (receipt of three doses) among boys aged 13-17 years (N = 19,518) with several individual-level and ZIP Code Tabulation Area (ZCTA) census measures. Multivariable logistic regression was used to estimate the odds of HPV vaccination initiation and series completion separately. In 2012-2013 approximately 27.9% (95% CI 26.6%-29.2%) of boys initiated and 10.38% (95% CI 9.48%-11.29%) completed the HPV vaccine series. Area-based poverty was not statistically significantly associated with HPV vaccination initiation. It was, however, associated with series completion, with boys living in high-poverty areas (≥20% of residents living below poverty) having higher odds of completing the series (AOR 1.22, 95% CI 1.01-1.48) than boys in low-poverty areas (0-4.99%). Interactions between race/ethnicity and ZIP code-level poverty indicated that Hispanic boys living in high-poverty areas had a statistically significantly higher odds of HPV vaccine initiation (AOR 1.43, 95% CI 1.03-1.97) and series completion (AOR 1.56, 95% CI 1.05-2.32) than Hispanic boys in low-poverty areas. Non-Hispanic Black boys in high poverty areas had higher odds of initiation (AOR 2.23, 95% CI 1.33-3.75) and completion (AOR 2.61, 95% CI 1.06-6.44) than non-Hispanic Black boys in low-poverty areas. Rural/urban residence and population density were also significant factors, with boys from urban or densely populated areas having higher odds of initiation and completion compared to boys living in non-urban, less densely populated areas. Higher HPV vaccination coverage in urban areas and among racial/ethnic minorities in areas with high poverty may be attributable to factors such as vaccine acceptance, health-care practices, and their access to HPV vaccines through the Vaccines for Children Program, which provides free vaccines to uninsured and under-insured children. Given the low HPV vaccination rates among boys in the U.S., these results provide important evidence to inform public health interventions to increase HPV vaccination.
Electronic stopping in oxides beyond Bragg additivity
NASA Astrophysics Data System (ADS)
Sigmund, P.; Schinner, A.
2018-01-01
We present stopping cross sections calculated by our PASS code for several ions in metal oxides and SiO2 over a wide energy range. Input takes into account changes in the valence structure by assigning two additional electrons to the 2p shell of oxygen and removing the appropriate number of electrons from the outer shells of the metal atom. Results are compared with tabulated experimental values and with two versions of Bragg's additivity rule. Calculated stopping cross sections are applied in testing a recently-proposed scaling rule, which relates the stopping cross section to the number of oxygen atoms per molecule.
Estimate of Space Radiation-Induced Cancer Risks for International Space Station Orbits
NASA Technical Reports Server (NTRS)
Wu, Honglu; Atwell, William; Cucinotta, Francis A.; Yang, Chui-hsu
1996-01-01
Excess cancer risks from exposures to space radiation are estimated for various orbits of the International Space Station (ISS). Organ exposures are computed with the transport codes, BRYNTRN and HZETRN, and the computerized anatomical male and computerized anatomical female models. Cancer risk coefficients in the National Council on Radiation Protection and Measurements report No. 98 are used to generate lifetime excess cancer incidence and cancer mortality after a one-month mission to ISS. The generated data are tabulated to serve as a quick reference for assessment of radiation risk to astronauts on ISS missions.
Overview of NASA supported Stirling thermodynamic loss research
NASA Technical Reports Server (NTRS)
Tew, Roy C.; Geng, Steven M.
1992-01-01
NASA is funding research to characterize Stirling machine thermodynamic losses. NASA's primary goal is to improve Stirling design codes to support engine development for space and terrestrial power. However, much of the fundamental data is applicable to Stirling cooling and heat pump applications. The research results are reviewed. Much was learned about oscillating flow hydrodynamics, including laminar/turbulent transition, and tabulated data was documented for further analysis. Now, with a better understanding of the oscillating flow field, it is time to begin measuring the effects of oscillating flow and oscillating pressure level on heat transfer in heat exchanger flow passages and in cylinders.
Code of Federal Regulations, 2014 CFR
2014-01-01
... 15 Commerce and Foreign Trade 1 2014-01-01 2014-01-01 false Report of tabulations of population to... DECENNIAL CENSUS POPULATION INFORMATION § 101.1 Report of tabulations of population to states and localities... the methodology to be used in calculating the tabulations of population reported to States and...
Code of Federal Regulations, 2012 CFR
2012-01-01
... 15 Commerce and Foreign Trade 1 2012-01-01 2012-01-01 false Report of tabulations of population to... DECENNIAL CENSUS POPULATION INFORMATION § 101.1 Report of tabulations of population to states and localities... the methodology to be used in calculating the tabulations of population reported to States and...
Code of Federal Regulations, 2013 CFR
2013-01-01
... 15 Commerce and Foreign Trade 1 2013-01-01 2013-01-01 false Report of tabulations of population to... DECENNIAL CENSUS POPULATION INFORMATION § 101.1 Report of tabulations of population to states and localities... the methodology to be used in calculating the tabulations of population reported to States and...
Design study for electronic system for Jupiter Orbit Probe (JOP)
NASA Technical Reports Server (NTRS)
Elero, B. P., Jr.; Carignan, G. R.
1978-01-01
The conceptual design of the Jupiter probe spectrometer is presented. Block and circuit diagrams are presented along with tabulated parts lists. Problem areas are considered to be (1) the schedule, (2) weight limitations for the electronic systems, and (3) radiation hardness of the electronic devices.
The "prudent layperson" definition of an emergency medical condition.
Li, James; Galvin, Hannah K; Johnson, Sandra C
2002-01-01
The study objectives, based on federal and state legislative language, were to objectively define symptoms and signs commonly agreed on by "prudent laypersons" as "emergency medical conditions." After comprehensive tabulation of symptom classifications from the International Classification of Diseases (ICD-9), we performed a survey of nonmedical laypersons. Data analysis included descriptive statistics, proportional calculations, and 95% confidence intervals. A minority of symptoms and signs (25/87, 29%) were considered emergency medical conditions by more than half of nonmedical survey respondents who were self-defined as prudent laypersons. The leading conditions deemed emergencies were loss of consciousness, seizure, no recognition of one side of the body, paralysis, shock, gangrene, coughing blood, trouble breathing, chest pain, and choking. Pain, except for renal colic or chest pain, was not considered an emergency. No symptoms or signs specifically related to gynecologic disorders were considered emergencies. Most symptoms and signs tabulated in the diagnostic coding manual, ICD-9, are not considered emergency medical conditions by self-designated prudent laypersons. These include many conditions that are commonly investigated and treated in the emergency department setting. Use of the prudent layperson standard for reimbursable emergency health services may not reflect the actual scope of symptoms necessitating emergency care.
TIGER/Line Shapefile, 2010, 2010 Census Block State-based
The TIGER/Line Files are shapefiles and related database files (.dbf) that are an extract of selected geographic and cartographic information from the U.S. Census Bureau's Master Address File / Topologically Integrated Geographic Encoding and Referencing (MAF/TIGER) Database (MTDB). The MTDB represents a seamless national file with no overlaps or gaps between parts, however, each TIGER/Line File is designed to stand alone as an independent data set, or they can be combined to cover the entire nation. Census Blocks are statistical areas bounded on all sides by visible features, such as streets, roads, streams, and railroad tracks, and/or by nonvisible boundaries such as city, town, township, and county limits, and short line-of-sight extensions of streets and roads. Census blocks are relatively small in area; for example, a block in a city bounded by streets. However, census blocks in remote areas are often large and irregular and may even be many square miles in area. A common misunderstanding is that data users think census blocks are used geographically to build all other census geographic areas, rather all other census geographic areas are updated and then used as the primary constraints, along with roads and water features, to delineate the tabulation blocks. As a result, all 2010 Census blocks nest within every other 2010 Census geographic area, so that Census Bureau statistical data can be tabulated at the block level and aggregated up to the appropr
Trogdon, Justin G; Ahn, Thomas
2015-03-01
The purpose of this study was to explore geospatial patterns in influenza vaccination. We conducted an ecological analysis of publicly funded influenza vaccinations at the ZIP code tabulation area (ZCTA) level using secondary data for publicly funded influenza vaccinations among eligible school-aged children (age range, 5-17 years) for the 2010-2011 and 2011-2012 influenza seasons from the North Carolina Immunization Registry (NCIR). NCIR data were merged by ZCTA with other publicly available data. We tested for spatial autocorrelation in unadjusted influenza vaccination rates using choropleth maps and Moran's I. We estimated nonspatial and spatial negative binomial models with spatially correlated random effects adjusted for demographic, economic, and health care variables. The study was conducted at the University of North Carolina at Chapel Hill in the spring of 2014. The NCIR demonstrated spatial autocorrelation in publicly funded influenza vaccinations among uninsured and means-tested, publicly insured school-aged children; ZCTAs tended to have influenza vaccination rates that were similar to their neighbors. This result was partially explained by included ZCTA characteristics, but not wholly. To the extent that the geospatial clustering of vaccination rates is the result of social influences, targeting interventions to increase influenza vaccination among school-aged children in one area could also lead to increases in neighboring areas. Copyright © 2015 Association for Professionals in Infection Control and Epidemiology, Inc. Published by Elsevier Inc. All rights reserved.
NASA Technical Reports Server (NTRS)
Hoffarth, C.; Khaled, B.; Rajan, S. D.; Goldberg, R.; Carney, K.; DuBois, P.; Blankenhorn, Gunther
2016-01-01
An orthotropic elasto-plastic-damage three-dimensional model with tabulated input has been developed to analyze the impact response of composite materials. The theory has been implemented as MAT 213 into a tailored version of LS-DYNA being developed under a joint effort of the FAA and NASA and has the following features: (a) the theory addresses any composite architecture that can be experimentally characterized as an orthotropic material and includes rate and temperature sensitivities, (b) the formulation is applicable for solid as well as shell element implementations and utilizes input data in a tabulated form directly from processed experimental data, (c) deformation and damage mechanics are both accounted for within the material model, (d) failure criteria are established that are functions of strain and damage parameters, and mesh size dependence is included, and (e) the theory can be efficiently implemented into a commercial code for both sequential and parallel executions. The salient features of the theory as implemented in LS-DYNA are illustrated using a widely used composite - the T800S/3900-2B[P2352W-19] BMS8-276 Rev-H-Unitape fiber/resin unidirectional composite. First, the experimental tests to characterize the deformation, damage and failure parameters in the material behavior are discussed. Second, the MAT213 input model and implementation details are presented with particular attention given to procedures that have been incorporated to ensure that the yield surfaces in the rate and temperature dependent plasticity model are convex. Finally, the paper concludes with a validation test designed to test the stability, accuracy and efficiency of the implemented model.
ERIC Educational Resources Information Center
Broward County Schools, Fort Lauderdale, FL.
THE SCIENCE PROGRAM AT NOVA HIGH SCHOOL IS STRUCTURED TO MEET THE NEEDS OF A SCIENTIFIC SOCIETY. THE JUNIOR HIGH PROGRAM EMPHASIZES THE ACCUMLATION, TABULATION, AND ANALYSIS OF DATA. THE SENIOR HIGH PROGRAM MAKES USE OF THESE PROCESSES IN THE INVESTIGATION OF VARIOUS SUBJECT AREAS. THE UNITS COMPOSING THE HIGH SCHOOL CURRICULUM ARE DESCRIBED. UNIT…
DOE Office of Scientific and Technical Information (OSTI.GOV)
Not Available
1979-06-01
In this compendium each profile of a nuclear facility is a capsule summary of pertinent facts regarding that particular installation. The facilities described include the entire fuel cycle in the broadest sense, encompassing resource recovery through waste management. Power plants and all US facilities have been excluded. To facilitate comparison the profiles have been recorded in a standard format. Because of the breadth of the undertaking some data fields do not apply to the establishment under discussion and accordingly are blank. The set of nuclear facility profiles occupies four volumes; the profiles are ordered by country name, and then bymore » facility code. Each nuclear facility profile volume contains two complete indexes to the information. The first index aggregates the facilities alphabetically by country. It is further organized by category of facility, and then by the four-character facility code. It provides a quick summary of the nuclear energy capability or interest in each country and also an identifier, the facility code, which can be used to access the information contained in the profile.« less
Pacific Northwest residential energy survey. Volume 3. Question-by-question results
DOE Office of Scientific and Technical Information (OSTI.GOV)
None
1980-07-01
Tabulations are presented of responses to approximately 105 questions. Results are tabulated by 9 geographic regions: the four states of Idaho, Montana, Oregon, and Washington; four climate zones in the region; and a weighted Pacific Northwest total. A description of the tabulated data is given in the Introduction. Tabulated data deal with questions on dwelling characteristics; heating and air-conditioning systems; water heating; appliances; demographic and swelling characteristics; and insulation.
Over Ten Million Served: Gendered Service in Language and Literature Workplaces
ERIC Educational Resources Information Center
Masse, Michelle A., Ed.; Hogan, Katie J., Ed.
2010-01-01
All tenured and tenure-track faculty know the trinity of promotion and tenure criteria: research, teaching, and service. While teaching and research are relatively well-defined areas of institutional focus and evaluation, service work is rarely tabulated or analyzed as a key aspect of higher education's political economy. Instead, service, silent…
Forest statistics for Arkansas counties - 1979
Renewable Resources Evaluation Research Work Unit
1979-01-01
This report tabulates information from a new forest survey of Arkansas completed in 1979 by the Renewable Resources Evaluation Research Unit of the Southern Forest Experiment Station. Forest area was estimated from aerial photos with an adjustment for ground truth at selected locations. Sample plots were systematically established at three-mile intervals using a grid...
Assessing Needs for Gerontological Education in Urban and Rural Areas of Ohio
ERIC Educational Resources Information Center
Van Dussen, Daniel J.; Leson, Suzanne M.; Emerick, Eric S.; Voytek, Joseph A.; Ewen, Heidi H.
2016-01-01
Purpose of the Study: This project surveyed health care professionals from both urban and rural care settings in Ohio and examined differences in professionals' needs and interests in continuing gerontological education. Design and Methods: The survey data were analyzed for 766 health care professionals descriptively, using cross-tabulations and…
Assessing soil impacts related to forest harvest operations
E.A. Carter; John M. III. Grace
2011-01-01
Three studies conducted in Alabama evaluated impacts associated with a clear cut harvest in three physiographic regions. Machine impacts were assessed via tabulation of soil disturbance classes, measurement of bulk density and soil strength, or a combination of the two. Soil disturbance classes were similar among all locations with untrafficked areas comprising...
Integration of Computer Related Instruction in Texas Vocational Agriculture Programs. Final Report.
ERIC Educational Resources Information Center
Cepica, M. J.; And Others
A study examined current usage of microcomputers, projected software needs, and teacher inservice training needs in Texas vocational agriculture programs. Questionnaires were mailed to each of the 922 vocational agriculture departments in Texas. Data from the 446 usable instruments returned were tabulated by geographical area and school size.…
NASA Astrophysics Data System (ADS)
Madrucci, Vanessa; Taioli, Fabio; de Araújo, Carlos César
2008-08-01
SummaryThis paper presents the groundwater favorability mapping on a fractured terrain in the eastern portion of São Paulo State, Brazil. Remote sensing, airborne geophysical data, photogeologic interpretation, geologic and geomorphologic maps and geographic information system (GIS) techniques have been used. The results of cross-tabulation between these maps and well yield data allowed groundwater prospective parameters in a fractured-bedrock aquifer. These prospective parameters are the base for the favorability analysis whose principle is based on the knowledge-driven method. The multicriteria analysis (weighted linear combination) was carried out to give a groundwater favorability map, because the prospective parameters have different weights of importance and different classes of each parameter. The groundwater favorability map was tested by cross-tabulation with new well yield data and spring occurrence. The wells with the highest values of productivity, as well as all the springs occurrence are situated in the excellent and good favorability mapped areas. It shows good coherence between the prospective parameters and the well yield and the importance of GIS techniques for definition of target areas for detail study and wells location.
Blood Glucose Meters That Are Accessible to Blind and Visually Impaired Persons
Uslan, Mark M.; Burton, Darren M.; Clements, Charles W.
2008-01-01
Blood glucose meters (BGMs) that can be used nonvisually or with a visual limitation were introduced in the mid-1990s, but it was not until 2006 and 2007 that a new set of meters with accessibility features were introduced: Prodigy, Prodigy Autocode, and Prodigy Voice (Diagnostic Devices, Charlotte, NC), and Advocate and Advocate Redi-Code (TaiDoc, Taiwan). Accessibility attributes of the newly introduced BGMs were tabulated, and product design features were examined and documented. The Prodigy Voice was found to be the only one of these new BGMs that is fully usable by blind and visually impaired persons. PMID:19885356
Blood glucose meters that are accessible to blind and visually impaired persons.
Uslan, Mark M; Burton, Darren M; Clements, Charles W
2008-03-01
Blood glucose meters (BGMs) that can be used nonvisually or with a visual limitation were introduced in the mid-1990s, but it was not until 2006 and 2007 that a new set of meters with accessibility features were introduced: Prodigy, Prodigy Autocode, and Prodigy Voice (Diagnostic Devices, Charlotte, NC), and Advocate and Advocate Redi-Code (TaiDoc, Taiwan). Accessibility attributes of the newly introduced BGMs were tabulated, and product design features were examined and documented. The Prodigy Voice was found to be the only one of these new BGMs that is fully usable by blind and visually impaired persons.
Interpolation of hard and soft dilepton rates
NASA Astrophysics Data System (ADS)
Ghisoiu, I.; Laine, M.
2014-10-01
Strict next-to-leading order (NLO) results for the dilepton production rate from a QCD plasma at temperatures above a few hundred MeV suffer from a breakdown of the loop expansion in the regime of soft invariant masses M 2 ≪ ( πT)2. In this regime an LPM resummation is needed for obtaining the correct leading-order result. We show how to construct an interpolation between the hard NLO and the leading-order LPM expression, which is theoretically consistent in both regimes and free from double counting. The final numerical results are presented in a tabulated form, suitable for insertion into hydrodynamical codes.
Study of X-ray photoionized Fe plasma and comparisons with astrophysical modeling codes
DOE Office of Scientific and Technical Information (OSTI.GOV)
Foord, M E; Heeter, R F; Chung, H
The charge state distributions of Fe, Na and F are determined in a photoionized laboratory plasma using high resolution x-ray spectroscopy. Independent measurements of the density and radiation flux indicate the ionization parameter {zeta} in the plasma reaches values {zeta} = 20-25 erg cm s{sup -1} under near steady-state conditions. A curve-of-growth analysis, which includes the effects of velocity gradients in a one-dimensional expanding plasma, fits the observed line opacities. Absorption lines are tabulated in the wavelength region 8-17 {angstrom}. Initial comparisons with a number of astrophysical x-ray photoionization models show reasonable agreement.
Effect of the tubular-fan drum shapes on the performance of cleaning head module
NASA Astrophysics Data System (ADS)
Hong, C. K.; Y Cho, M.; Kim, Y. J.
2013-12-01
The geometrical effects of a tubular-fan drum on the performance improvement of the cleaning head module of a vacuum cleaner were investigated. In this study, the number of blades and the width of the blade were selected as the design parameters. Static pressure, eccentric vortex, turbulence kinetic energy (TKE) and suction efficiency were analysed and tabulated. Three-dimensional computational fluid dynamics method was used with an SST (Shear Stress Transfer) turbulence model to simulate the flow field at the suction of the cleaning head module using the commercial code ANSYS-CFX. Suction pressure distributions were graphically depicted for different values of the design parameters.
1989-09-30
information for use by readers to interpret the tabulation volumes accompanying the final project report: 1989 Survey of U.S. Army Reserve (USAR) Troop...34stayers" who were used as the sample to generate the first longitudinal Tabulation Volume. Comparing questionnaire response frequencies between the...as described below). Detailed below are the specific crossing variables used for the cross-sectional and longitudinal Tabulation Volumes. Cross
DOE Office of Scientific and Technical Information (OSTI.GOV)
Watson, C.R.
The SNODOG Glossary is used by the DOE-supported life-span beagle studies to describe medical observations in a standardized format. It is an adaptation of the human medical glossary, SNOMED, which lists 107,165 terms. Each of the five laboratories, Argonne National Laboratory, the Inhalation Toxicology Research Institute, the Pacific Northwest Laboratory, the University of California at Davis, and the University of Utah, has selected an appropriate subset from the published SNOMED glossary and added beagle and research-specific terms. The National Radiobiology Archives is the coordinator of these enhancements, and periodically distributes SNODOG to the respective laboratories. Information donated by Colorado Statemore » University and Oak Ridge National Laboratory has been related to SNODOG and is available in a standardized format. This document is designed for the database manager and the scientist who will be managing or coding medical observations. It is also designed for the scientist analyzing coded information. The document includes: an overview of the NRA and the SNODOG glossary, a discussion of hardware requirements, a review of the SNODOG code structure and printed lists of the 4,770 terms which have been used at least once. Instructions for obtaining electronic copies of the glossary and for nominating additional terms are provided. This document describes the origins and structure of the SNODOG codes, explains code usage at each participating institution, and presents a usage frequency tabulation of the terms for neoplasia. A diskette or magnetic tape containing 15,641 SNODOG codes and translations is available on request.« less
SNODOG Glossary: Part 1, Introduction
DOE Office of Scientific and Technical Information (OSTI.GOV)
Watson, C.R.
The SNODOG Glossary is used by the DOE-supported life-span beagle studies to describe medical observations in a standardized format. It is an adaptation of the human medical glossary, SNOMED, which lists 107,165 terms. Each of the five laboratories, Argonne National Laboratory, the Inhalation Toxicology Research Institute, the Pacific Northwest Laboratory, the University of California at Davis, and the University of Utah, has selected an appropriate subset from the published SNOMED glossary and added beagle and research-specific terms. The National Radiobiology Archives is the coordinator of these enhancements, and periodically distributes SNODOG to the respective laboratories. Information donated by Colorado Statemore » University and Oak Ridge National Laboratory has been related to SNODOG and is available in a standardized format. This document is designed for the database manager and the scientist who will be managing or coding medical observations. It is also designed for the scientist analyzing coded information. The document includes: an overview of the NRA and the SNODOG glossary, a discussion of hardware requirements, a review of the SNODOG code structure and printed lists of the 4,770 terms which have been used at least once. Instructions for obtaining electronic copies of the glossary and for nominating additional terms are provided. This document describes the origins and structure of the SNODOG codes, explains code usage at each participating institution, and presents a usage frequency tabulation of the terms for neoplasia. A diskette or magnetic tape containing 15,641 SNODOG codes and translations is available on request.« less
Berger, D.L.; Schaefer, D.H.; Frick, E.A.
1990-01-01
Principal facts for 537 gravity stations in the carbonate-rock province of eastern and southern Nevada are tabulated and presented. The gravity data were collected in support of groundwater studies in several valleys. The study areas include the Elko area, northern Steptoe Valley, Coyote Spring Valley, and the western Sheep Range area. The data for each site include values for latitude, longitude, altitude, observed gravity, free- air anomaly, terrain correction, and Bouguer anomaly (calculated at a bedrock density of 2.67 g/cu cm. (USGS)
Report list Arizona's oil, gas potential
Rauzi, S.L.
2001-01-01
This article is a summary of Arizona geological survey circular 29, which addresses the petroleum geology of Arizona, USA. Eight areas have been identified with fair to excellent oil and gas potential, and some Tertiary basins have evidence of source or reservoir rocks. The following are considered here: production history, lands status and services, regulation and permitting, petroleum geology, hydrocarbon indications, and areas with hydrocarbon potential and their petroleum geology and characteristics. The full report contains detailed figures of each of these basin areas, a descriptive tabulation of seeps and petroliferous rocks and extensive references.
Ground-water data for the Beryl-Enterprise area, Escalante Desert, Utah
Mower, R.W.
1981-01-01
This report contains a compilation of selected ground-water data for the Beryl-Enterprise area, Iron and Washington Counties, Utah. The records of the wells include such information as driller 's logs, yield, drawdown, use, and temperature of the well water. There are also records of water levels in selected wells for the period 1973-79, chemical analyses of ground water, records of selected springs, and a tabulation of ground-water withdrawals for 1937-78. (USGS)
Flood of June 7-9, 2008, in Central and Southern Indiana
Morlock, Scott E.; Menke, Chad D.; Arvin, Donald V.; Kim, Moon H.
2008-01-01
On June 6-7, 2008, heavy rainfall of 2 to more than 10 inches fell upon saturated soils and added to already high streamflows from a wetter than normal spring in central and southern Indiana. The heavy rainfall resulted in severe flooding on many streams within the White River Basin during June 7-9, causing three deaths, evacuation of thousands of residents, and hundreds of millions of dollars of damage to residences, businesses, infrastructure, and agricultural lands. In all, 39 Indiana counties were declared Federal disaster areas. U.S. Geological Survey (USGS) streamgages at nine locations recorded new record peak streamflows for the respective periods of record as a result of the heavy rainfall. Recurrence intervals of flood-peak streamflows were estimated to be greater than 100 years at five streamgages and 50-100 years at two streamgages. Peak-gage-height data, peak-streamflow data, and recurrence intervals are tabulated for 19 USGS streamgages in central and southern Indiana. Peak-streamflow estimates are tabulated for four ungaged locations, and estimated recurrence intervals are tabulated for three ungaged locations. The estimated recurrence interval for an ungaged location on Haw Creek in Columbus was greater than 100 years and for an ungaged location on Hurricane Creek in Franklin was 50-100 years. Because flooding was particularly severe in the communities of Columbus, Edinburgh, Franklin, Paragon, Seymour, Spencer, Martinsville, Newberry, and Worthington, high-water-mark data collected after the flood were tabulated for those communities. Flood peak inundation maps and water-surface profiles for selected streams were made in a geographic information system by combining the high-water-mark data with the highest-resolution digital elevation model data available.
Code of Federal Regulations, 2011 CFR
2011-07-01
... 41 Public Contracts and Property Management 2 2011-07-01 2007-07-01 true Requisitioning tabulating... Contracts and Property Management Federal Property Management Regulations System FEDERAL PROPERTY MANAGEMENT... electrical and mechanical contact tabulating machines, including aperture cards and copy cards. Federal...
Code of Federal Regulations, 2010 CFR
2010-07-01
... 41 Public Contracts and Property Management 2 2010-07-01 2010-07-01 true Requisitioning tabulating... Contracts and Property Management Federal Property Management Regulations System FEDERAL PROPERTY MANAGEMENT... electrical and mechanical contact tabulating machines, including aperture cards and copy cards. Federal...
Forest resources of the eastern Ozark Region in Missouri
The Forest Survey Organization Central States Forest Experiment Station
1948-01-01
This Survey Release presents the more significant statistics on forest area and timber volume in 14 counties in the Eastern Ozark region of Missouri. As soon as statistical tabulations have been completed other releases will be issued giving similar information for the other important subdivisions of the State. Later an analytical report for the entire State will be...
Growth models for ponderosa pine: I. Yield of unthinned plantations in northern California.
William W. Oliver; Robert F. Powers
1978-01-01
Yields for high-survival, unthinned ponderosa pine (Pinus ponderosa Laws.) plantations in northern California are estimated. Stems of 367 trees in 12 plantations were analyzed to produce a growth model simulating stand yields. Diameter, basal area, and net cubic volume yields by Site Indices50 40 through 120 are tabulated for...
Forest statistics of central Kentucky
The Forest Survey Organization Central States Forest Experiment Station
1950-01-01
This Survey Release presents the more significant preliminary statistics on the forest area and timber volume for each of the four regions of Central Kentucky. A similar report has been published for the Western Kentucky region and a release for the eastern region will be issued as soon as field work and tabulations are completed. Later an analytical report for the...
ERIC Educational Resources Information Center
Who's Who among American High School Students, Northbrook, IL.
The report, fact sheets, and news releases cite findings of a survey involving 22,000 Ss to determine the attitudes of high school student leaders. Responses were tabulated for the following areas: demographics, government mandates/taxes, American institutions, domestic issues, foreign issues/defense, religious beliefs/practices, marriage/sex,…
23 CFR 635.113 - Bid opening and bid tabulations.
Code of Federal Regulations, 2013 CFR
2013-04-01
... permitted. (b) The STD shall prepare and forward tabulations of bids to the Division Administrator. These tabulations shall be certified by a responsible STD official and shall show: (1) Bid item details for at least... opened and reviewed in accordance with the terms of the solicitation. The STD must use its own procedures...
23 CFR 635.113 - Bid opening and bid tabulations.
Code of Federal Regulations, 2014 CFR
2014-04-01
... permitted. (b) The STD shall prepare and forward tabulations of bids to the Division Administrator. These tabulations shall be certified by a responsible STD official and shall show: (1) Bid item details for at least... opened and reviewed in accordance with the terms of the solicitation. The STD must use its own procedures...
23 CFR 635.113 - Bid opening and bid tabulations.
Code of Federal Regulations, 2012 CFR
2012-04-01
... permitted. (b) The STD shall prepare and forward tabulations of bids to the Division Administrator. These tabulations shall be certified by a responsible STD official and shall show: (1) Bid item details for at least... opened and reviewed in accordance with the terms of the solicitation. The STD must use its own procedures...
23 CFR 635.113 - Bid opening and bid tabulations.
Code of Federal Regulations, 2011 CFR
2011-04-01
... permitted. (b) The STD shall prepare and forward tabulations of bids to the Division Administrator. These tabulations shall be certified by a responsible STD official and shall show: (1) Bid item details for at least... opened and reviewed in accordance with the terms of the solicitation. The STD must use its own procedures...
NASA Technical Reports Server (NTRS)
Jasperson, W. H.; Holdeman, J. D.
1984-01-01
Tabulations are given of GASP ambient ozone mean, standard deviation, median, 84th percentile, and 98th percentile values, by month, flight level, and geographical region. These data are tabulated to conform to the temporal and spatial resolution required by FAA Advisory Circular 120-38 (monthly by 2000 ft in altitude by 5 deg in latitude) for climatological data used to show compliance with cabin ozone regulations. In addition seasonal x 10 deg latitude tabulations are included which are directly comparable to and supersede the interim GASP ambient ozone tabulations given in appendix B of FAA-EE-80-43 (NASA TM-81528). Selected probability variations are highlighted to illustrate the spatial and temporal variability of ambient ozone and to compare results from the coarse and fine grid analyses.
Congdon, Peter
2009-01-30
Estimates of disease prevalence for small areas are increasingly required for the allocation of health funds according to local need. Both individual level and geographic risk factors are likely to be relevant to explaining prevalence variations, and in turn relevant to the procedure for small area prevalence estimation. Prevalence estimates are of particular importance for major chronic illnesses such as cardiovascular disease. A multilevel prevalence model for cardiovascular outcomes is proposed that incorporates both survey information on patient risk factors and the effects of geographic location. The model is applied to derive micro area prevalence estimates, specifically estimates of cardiovascular disease for Zip Code Tabulation Areas in the USA. The model incorporates prevalence differentials by age, sex, ethnicity and educational attainment from the 2005 Behavioral Risk Factor Surveillance System survey. Influences of geographic context are modelled at both county and state level, with the county effects relating to poverty and urbanity. State level influences are modelled using a random effects approach that allows both for spatial correlation and spatial isolates. To assess the importance of geographic variables, three types of model are compared: a model with person level variables only; a model with geographic effects that do not interact with person attributes; and a full model, allowing for state level random effects that differ by ethnicity. There is clear evidence that geographic effects improve statistical fit. Geographic variations in disease prevalence partly reflect the demographic composition of area populations. However, prevalence variations may also show distinct geographic 'contextual' effects. The present study demonstrates by formal modelling methods that improved explanation is obtained by allowing for distinct geographic effects (for counties and states) and for interaction between geographic and person variables. Thus an appropriate methodology to estimate prevalence at small area level should include geographic effects as well as person level demographic variables.
Congdon, Peter
2009-01-01
Background Estimates of disease prevalence for small areas are increasingly required for the allocation of health funds according to local need. Both individual level and geographic risk factors are likely to be relevant to explaining prevalence variations, and in turn relevant to the procedure for small area prevalence estimation. Prevalence estimates are of particular importance for major chronic illnesses such as cardiovascular disease. Methods A multilevel prevalence model for cardiovascular outcomes is proposed that incorporates both survey information on patient risk factors and the effects of geographic location. The model is applied to derive micro area prevalence estimates, specifically estimates of cardiovascular disease for Zip Code Tabulation Areas in the USA. The model incorporates prevalence differentials by age, sex, ethnicity and educational attainment from the 2005 Behavioral Risk Factor Surveillance System survey. Influences of geographic context are modelled at both county and state level, with the county effects relating to poverty and urbanity. State level influences are modelled using a random effects approach that allows both for spatial correlation and spatial isolates. Results To assess the importance of geographic variables, three types of model are compared: a model with person level variables only; a model with geographic effects that do not interact with person attributes; and a full model, allowing for state level random effects that differ by ethnicity. There is clear evidence that geographic effects improve statistical fit. Conclusion Geographic variations in disease prevalence partly reflect the demographic composition of area populations. However, prevalence variations may also show distinct geographic 'contextual' effects. The present study demonstrates by formal modelling methods that improved explanation is obtained by allowing for distinct geographic effects (for counties and states) and for interaction between geographic and person variables. Thus an appropriate methodology to estimate prevalence at small area level should include geographic effects as well as person level demographic variables. PMID:19183458
Verification of unfold error estimates in the UFO code
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fehl, D.L.; Biggs, F.
Spectral unfolding is an inverse mathematical operation which attempts to obtain spectral source information from a set of tabulated response functions and data measurements. Several unfold algorithms have appeared over the past 30 years; among them is the UFO (UnFold Operator) code. In addition to an unfolded spectrum, UFO also estimates the unfold uncertainty (error) induced by running the code in a Monte Carlo fashion with prescribed data distributions (Gaussian deviates). In the problem studied, data were simulated from an arbitrarily chosen blackbody spectrum (10 keV) and a set of overlapping response functions. The data were assumed to have anmore » imprecision of 5% (standard deviation). 100 random data sets were generated. The built-in estimate of unfold uncertainty agreed with the Monte Carlo estimate to within the statistical resolution of this relatively small sample size (95% confidence level). A possible 10% bias between the two methods was unresolved. The Monte Carlo technique is also useful in underdetemined problems, for which the error matrix method does not apply. UFO has been applied to the diagnosis of low energy x rays emitted by Z-Pinch and ion-beam driven hohlraums.« less
Maund, Emma; Tendal, Britta; Hróbjartsson, Asbjørn; Lundh, Andreas; Gøtzsche, Peter C
2014-06-04
To assess the effects of coding and coding conventions on summaries and tabulations of adverse events data on suicidality within clinical study reports. Systematic electronic search for adverse events of suicidality in tables, narratives, and listings of adverse events in individual patients within clinical study reports. Where possible, for each event we extracted the original term reported by the investigator, the term as coded by the medical coding dictionary, medical coding dictionary used, and the patient's trial identification number. Using the patient's trial identification number, we attempted to reconcile data on the same event between the different formats for presenting data on adverse events within the clinical study report. 9 randomised placebo controlled trials of duloxetine for major depressive disorder submitted to the European Medicines Agency for marketing approval. Clinical study reports obtained from the EMA in 2011. Six trials used the medical coding dictionary COSTART (Coding Symbols for a Thesaurus of Adverse Reaction Terms) and three used MedDRA (Medical Dictionary for Regulatory Activities). Suicides were clearly identifiable in all formats of adverse event data in clinical study reports. Suicide attempts presented in tables included both definitive and provisional diagnoses. Suicidal ideation and preparatory behaviour were obscured in some tables owing to the lack of specificity of the medical coding dictionary, especially COSTART. Furthermore, we found one event of suicidal ideation described in narrative text that was absent from tables and adverse event listings of individual patients. The reason for this is unclear, but may be due to the coding conventions used. Data on adverse events in tables in clinical study reports may not accurately represent the underlying patient data because of the medical dictionaries and coding conventions used. In clinical study reports, the listings of adverse events for individual patients and narratives of adverse events can provide additional information, including original investigator reported adverse event terms, which can enable a more accurate estimate of harms. © Maund et al 2014.
Experimental validation of beam quality correction factors for proton beams
NASA Astrophysics Data System (ADS)
Gomà, Carles; Hofstetter-Boillat, Bénédicte; Safai, Sairos; Vörös, Sándor
2015-04-01
This paper presents a method to experimentally validate the beam quality correction factors (kQ) tabulated in IAEA TRS-398 for proton beams and to determine the kQ of non-tabulated ionization chambers (based on the already tabulated values). The method is based exclusively on ionometry and it consists in comparing the reading of two ionization chambers under the same reference conditions in a proton beam quality Q and a reference beam quality 60Co. This allows one to experimentally determine the ratio between the kQ of the two ionization chambers. In this work, 7 different ionization chamber models were irradiated under the IAEA TRS-398 reference conditions for 60Co beams and proton beams. For the latter, the reference conditions for both modulated beams (spread-out Bragg peak field) and monoenergetic beams (pseudo-monoenergetic field) were studied. For monoenergetic beams, it was found that the experimental kQ values obtained for plane-parallel chambers are consistent with the values tabulated in IAEA TRS-398; whereas the kQ values obtained for cylindrical chambers are not consistent—being higher than the tabulated values. These results support the suggestion (of previous publications) that the IAEA TRS-398 reference conditions for monoenergetic proton beams should be revised so that the effective point of measurement of cylindrical ionization chambers is taken into account when positioning the reference point of the chamber at the reference depth. For modulated proton beams, the tabulated kQ values of all the ionization chambers studied in this work were found to be consistent with each other—except for the IBA FC65-G, whose experimental kQ value was found to be 0.6% lower than the tabulated one. The kQ of the PTW Advanced Markus chamber, which is not tabulated in IAEA TRS-398, was found to be 0.997 ± 0.042 (k = 2), based on the tabulated value of the PTW Markus chamber.
Evaluation of different flamelet tabulation methods for laminar spray combustion
NASA Astrophysics Data System (ADS)
Luo, Yujuan; Wen, Xu; Wang, Haiou; Luo, Kun; Fan, Jianren
2018-05-01
In this work, three different flamelet tabulation methods for spray combustion are evaluated. Major differences among these methods lie in the treatment of the temperature boundary conditions of the flamelet equations. Particularly, in the first tabulation method ("M1"), both the fuel and oxidizer temperature boundary conditions are set to be fixed. In the second tabulation method ("M2"), the fuel temperature boundary condition is varied while the oxidizer temperature boundary condition is fixed. In the third tabulation method ("M3"), both the fuel and oxidizer temperature boundary conditions are varied and set to be equal. The focus of this work is to investigate whether the heat transfer between the droplet phase and gas phase can be represented by the studied tabulation methods through a priori analyses. To this end, spray flames stabilized in a three-dimensional counterflow are first simulated with detailed chemistry. Then, the trajectory variables are calculated from the detailed chemistry solutions. Finally, the tabulated thermo-chemical quantities are compared to the corresponding values from the detailed chemistry solutions. The comparisons show that the gas temperature cannot be predicted by "M1" with only a mixture fraction and reaction progress variable being the trajectory variables. The gas temperature can be correctly predicted by both "M2" and "M3," in which the total enthalpy is introduced as an additional manifold. In "M2," variations of the oxidizer temperature are considered with a temperature modification technique, which is not required in "M3." Interestingly, it is found that the mass fractions of the reactants and major products are not sensitive to the representation of the interphase heat transfer in the flamelet chemtables, and they can be correctly predicted by all tabulation methods. By contrast, the intermediate species CO and H2 in the premixed flame reaction zone are over-predicted by all tabulation methods.
Polarized Continuum Radiation from Stellar Atmospheres
NASA Astrophysics Data System (ADS)
Harrington, J. Patrick
2015-10-01
Continuum scattering by free electrons can be significant in early type stars, while in late type stars Rayleigh scattering by hydrogen atoms or molecules may be important. Computer programs used to construct models of stellar atmospheres generally treat the scattering of the continuum radiation as isotropic and unpolarized, but this scattering has a dipole angular dependence and will produce polarization. We review an accurate method for evaluating the polarization and limb darkening of the radiation from model stellar atmospheres. We use this method to obtain results for: (i) Late type stars, based on the MARCS code models (Gustafsson et al. 2008), and (ii) Early type stars, based on the NLTE code TLUSTY (Lanz and Hubeny 2003). These results are tabulated at http://www.astro.umd.edu/~jph/Stellar_Polarization.html. While the net polarization vanishes for an unresolved spherical star, this symmetry is broken by rapid rotation or by the masking of part of the star by a binary companion or during the transit of an exoplanet. We give some numerical results for these last cases.
Code of Federal Regulations, 2011 CFR
2011-07-01
... machine cards not available from Federal Supply Schedule contracts. 101-26.509-2 Section 101-26.509-2... Programs § 101-26.509-2 Requisitioning tabulating machine cards not available from Federal Supply Schedule contracts. (a) Requisitions for tabulating machine cards covered by Federal Supply Schedule contracts which...
Ground-water data in the Baker County-northern Malheur County area, Oregon
Collins, C.A.
1979-01-01
Ground-water data for the Baker County-northern Malheur area, Oregon, are tabulated for the Bureau of Land Management. The data include well and spring records, a well-location map, drillers ' logs of wells, observation-well hydrographs, and chemical analyses of ground-water samples. The reported yields of wells and springs in the area ranged from less than 1 to 2 ,500 gallons per minute. Dissolved solids in ground-water samples ranged from 50 to 1,587 milligrams per liter, and arsenic ranged from 0.001 to 0.317 milligrams per liter. (Woodard-USGS)
Environmental monitoring from spacecraft data. [Ohio, Kentucky, and Indiana
NASA Technical Reports Server (NTRS)
Rogers, R. H. (Principal Investigator); Wilson, C. L.; Reed, L. E.; Shah, N. J.; Akeley, R.; Mara, T. G.; Smith, V. E.
1975-01-01
The author has identified the following significant results. LANDSAT was used as a basis for inventorying land use within each of the Ohio-Kentucky-Indiana regional commissions, 225 drainage areas, and nine counties. Computer tabulations were produced to obtain the area covered by each of 16 land use categories within 225 drainage areas. The 16 categories were merged into ten categories and mapped at a scale of 1 inch = 5,000 ft, with detail to 0.44 hectares for the 2,700 sq mi region. These products were produced in less than 90 days, at a cost of $20,000.
NASA Technical Reports Server (NTRS)
Manhardt, P. D.
1982-01-01
The CMC fluid mechanics program system was developed to transmit the theoretical solution of finite element numerical solution methodology, applied to nonlinear field problems into a versatile computer code for comprehensive flow field analysis. Data procedures for the CMC 3 dimensional Parabolic Navier-Stokes (PNS) algorithm are presented. General data procedures a juncture corner flow standard test case data deck is described. A listing of the data deck and an explanation of grid generation methodology are presented. Tabulations of all commands and variables available to the user are described. These are in alphabetical order with cross reference numbers which refer to storage addresses.
Mission Analysis for High Specific Impulse Deep Space Exploration
NASA Technical Reports Server (NTRS)
Adams, Robert B.; Polsgrove, Tara; Brady, Hugh J. (Technical Monitor)
2002-01-01
This paper describes trajectory calculations for high specific impulse engines. Specific impulses on the order of 10,000 to 100,000 sec are predicted in a variety of fusion powered propulsion systems. This paper and its companion paper seek to build on analyses in the literature to yield an analytical routine for determining time of flight and payload fraction to a predetermined destination. The companion paper will compare the results of this analysis to the trajectories determined by several trajectory codes. The major parameters that affect time of flight and payload fraction will be identified and their sensitivities quantified. A review of existing fusion propulsion concepts and their capabilities will also be tabulated.
Forest Resources of the southwestern Ozark region in Missouri
The Forest Survey Organization. Central States Forest Experiment Station
1948-01-01
This Survey Release presents the more significant statistics on forest area and timber volume in 12 counties in the Southwestern Ozark region of Missouri. A similar report has been published for the Eastern Ozark region and releases for the other subdivisions of the State will be issued as soon as statistical tabulations have been completed. Later, an analytical report...
Estimating snow load in California for three recurrence intervals
David L. Azuma
1985-01-01
A key to designing facilities in snowbound areas is knowing what the expected snow load levels are for given recurrence intervals. In California, information about snow load is available only for the Lake Tahoe Basin. About 280 snow courses in the State were analyzed, and snow load estimated and related to elevation on a river basin and statewide level. The tabulated...
Update and evaluation of decay data for spent nuclear fuel analyses
NASA Astrophysics Data System (ADS)
Simeonov, Teodosi; Wemple, Charles
2017-09-01
Studsvik's approach to spent nuclear fuel analyses combines isotopic concentrations and multi-group cross-sections, calculated by the CASMO5 or HELIOS2 lattice transport codes, with core irradiation history data from the SIMULATE5 reactor core simulator and tabulated isotopic decay data. These data sources are used and processed by the code SNF to predict spent nuclear fuel characteristics. Recent advances in the generation procedure for the SNF decay data are presented. The SNF decay data includes basic data, such as decay constants, atomic masses and nuclide transmutation chains; radiation emission spectra for photons from radioactive decay, alpha-n reactions, bremsstrahlung, and spontaneous fission, electrons and alpha particles from radioactive decay, and neutrons from radioactive decay, spontaneous fission, and alpha-n reactions; decay heat production; and electro-atomic interaction data for bremsstrahlung production. These data are compiled from fundamental (ENDF, ENSDF, TENDL) and processed (ESTAR) sources for nearly 3700 nuclides. A rigorous evaluation procedure of internal consistency checks and comparisons to measurements and benchmarks, and code-to-code verifications is performed at the individual isotope level and using integral characteristics on a fuel assembly level (e.g., decay heat, radioactivity, neutron and gamma sources). Significant challenges are presented by the scope and complexity of the data processing, a dearth of relevant detailed measurements, and reliance on theoretical models for some data.
Precambrian basement geology of North and South Dakota.
Klasner, J.S.; King, E.R.
1986-01-01
Combined analysis of drill-hole, gravity and magnetic data indicates that the Precambrian rocks in the basement of the Dakotas may be divided into a series of lithotectonic terrains. On the basis of an analysis of geological and geophysical data in the Dakotas and from the surrounding states and Canada, it is shown how the exposed Precambrian rocks of the adjacent shield areas project into the study area. Brief comments are made on the tectonic implications of this study. Geological and geophysical characteristics of 11 terrains are tabulated. -P.Br.
ERIC Educational Resources Information Center
LERMAN, ALAN; ROSENSTEIN, JOSEPH
THE OBJECTIVES OF THIS TWO YEAR STUDY WERE--(1) TABULATION OF VOCATIONAL STATUS, (2) DETERMINATION OF SCHOOL AND POST-SCHOOL FACTORS LEADING TO JOB SUCCESS OR FAILURE, AND (3) LOCATING PROBLEM AREAS IN TOTAL JOB PROCESS. POST-SCHOOL VOCATIONAL INFORMATION WAS OBTAINED FROM 177 DEAF WOMEN WHO HAD ATTENDED THE LEXINGTON SCHOOL FOR THE DEAF PRIOR TO…
ERIC Educational Resources Information Center
Conway, Lorraine
In an effort to provide science teachers with the tables and scales most often used in teaching earth science, this document was designed to coordinate each table with meaningful activities, projects and experiments. The major areas covered by the booklet are: (1) electromagnetic waves (with activities about light waves and sound waves); (2) the…
Survey of United States Army Reserve (USAR) Troop Program Unit (TPU) soldiers - 1989
1989-09-30
TABULATION VOLUMES This material provides information for use by readers to interpret the tabulation volumes accompanying the final project report...questionnaire responses of the junior enlisted "stayers" who were used as the sample to generate the first longitudinal Tabulation Volume. Comparing...crossed by Iselected demographic variables (as described below). Detailed below are the specific crossing variables used for the cross-sectional and
Ishii, Tadashi; Nakayama, Masaharu; Abe, Michiaki; Takayama, Shin; Kamei, Takashi; Abe, Yoshiko; Yamadera, Jun; Amito, Koichiro; Morino, Kazuma
2016-10-01
Introduction There were 5,385 deceased and 710 missing in the Ishinomaki medical zone following the Great East Japan Earthquake that occurred in Japan on March 11, 2011. The Ishinomaki Zone Joint Relief Team (IZJRT) was formed to unify the relief teams of all organizations joining in support of the Ishinomaki area. The IZJRT expanded relief activity as they continued to manually collect and analyze assessments of essential information for maintaining health in all 328 shelters using a paper-type survey. However, the IZJRT spent an enormous amount of time and effort entering and analyzing these data because the work was vastly complex. Therefore, an assessment system must be developed that can tabulate shelter assessment data correctly and efficiently. The objective of this report was to describe the development and verification of a system to rapidly assess evacuation centers in preparation for the next major disaster. Report Based on experiences with the complex work during the disaster, software called the "Rapid Assessment System of Evacuation Center Condition featuring Gonryo and Miyagi" (RASECC-GM) was developed to enter, tabulate, and manage the shelter assessment data. Further, a verification test was conducted during a large-scale Self-Defense Force (SDF) training exercise to confirm its feasibility, usability, and accuracy. The RASECC-GM comprises three screens: (1) the "Data Entry screen," allowing for quick entry on tablet devices of 19 assessment items, including shelter administrator, living and sanitary conditions, and a tally of the injured and sick; (2) the "Relief Team/Shelter Management screen," for registering information on relief teams and shelters; and (3) the "Data Tabulation screen," which allows tabulation of the data entered for each shelter, as well as viewing and sorting from a disaster headquarters' computer. During the verification test, data of mock shelters entered online were tabulated quickly and accurately on a mock disaster headquarters' computer. Likewise, data entered offline also were tabulated quickly on the mock disaster headquarters' computer when the tablet device was moved into an online environment. The RASECC-GM, a system for rapidly assessing the condition of evacuation centers, was developed. Tests verify that users of the system would be able to easily, quickly, and accurately assess vast quantities of data from multiple shelters in a major disaster and immediately manage the inputted data at the disaster headquarters. Ishii T , Nakayama M , Abe M , Takayama S , Kamei T , Abe Y , Yamadera J , Amito K , Morino K . Development and verification of a mobile shelter assessment system "Rapid Assessment System of Evacuation Center Condition featuring Gonryo and Miyagi (RASECC-GM)" for major disasters. Prehosp Disaster Med. 2016;31(5):539-546.
Annual peak discharges from small drainage areas in Montana through September 1976
Johnson, M.V.; Omang, R.J.; Hull, J.A.
1977-01-01
Annual peak discharge from small drainage areas is tabulated for 336 sites in Montana. The 1976 additions included data collected at 206 sites. The program which investigates the magnitude and frequency of floods from small drainage areas in Montana, was begun July 1, 1955. Originally 45 crest-stage gaging stations were established. The purpose of the program is to collect sufficient peak-flow data, which through analysis could provide methods for estimating the magnitude and frequency of floods at any point in Montana. The ultimate objective is to provide methods for estimating the 100-year flood with the reliability needed for road design. (Woodard-USGS)
Goswami, Neela D; Schmitz, Michelle M; Sanchez, Travis; Dasgupta, Sharoda; Sullivan, Patrick; Cooper, Hannah; Rane, Deepali; Kelly, Jane; Del Rio, Carlos; Waller, Lance A
2016-05-01
Engagement in care is central to reducing mortality for HIV-infected persons and achieving the White House National AIDS Strategy of 80% viral suppression in the US by 2020. Where an HIV-infected person lives impacts his or her ability to achieve viral suppression. Reliable transportation access for healthcare may be a key determinant of this place-suppression relationship. ZIP code tabulation areas (ZCTAs) were the units of analysis. We used geospatial and ecologic analyses to examine spatial distributions of neighborhood-level variables (eg, transportation accessibility) and associations with: (1) community linkage to care, and (2) community viral suppression. Among Atlanta ZCTAs with data for newly diagnosed HIV cases (2006-2010), we used Moran I to evaluate spatial clustering and linear regression models to evaluate associations between neighborhood variables and outcomes. In 100 ZCTAs with 8413 newly diagnosed HIV-positive residents, a median of 60 HIV cases were diagnosed per ZCTA during the 5-year period. We found significant clustering of ZCTAs with low linkage to care and viral suppression (Moran I = 0.218, P < 0.05). In high-poverty ZCTAs, a 10% point increase in ZCTA-level household vehicle ownership was associated with a 4% point increase in linkage to care (P = 0.02, R = 0.16). In low-poverty ZCTAs, a 10% point increase in ZCTA-level household vehicle ownership was associated with a 30% point increase in ZCTA-level viral suppression (P = 0.01, R = 0.08). Correlations between transportation variables and community-level care linkage and viral suppression vary by area poverty level and provide opportunities for interventions beyond individual-level factors.
O'Lenick, Cassandra R; Winquist, Andrea; Mulholland, James A; Friberg, Mariel D; Chang, Howard H; Kramer, Michael R; Darrow, Lyndsey A; Sarnat, Stefanie Ebelt
2017-02-01
A broad literature base provides evidence of association between air pollution and paediatric asthma. Socioeconomic status (SES) may modify these associations; however, previous studies have found inconsistent evidence regarding the role of SES. Effect modification of air pollution-paediatric asthma morbidity by multiple indicators of neighbourhood SES was examined in Atlanta, Georgia. Emergency department (ED) visit data were obtained for 5-18 years old with a diagnosis of asthma in 20-county Atlanta during 2002-2008. Daily ZIP Code Tabulation Area (ZCTA)-level concentrations of ozone, nitrogen dioxide, fine particulate matter and elemental carbon were estimated using ambient monitoring data and emissions-based chemical transport model simulations. Pollutant-asthma associations were estimated using a case-crossover approach, controlling for temporal trends and meteorology. Effect modification by ZCTA-level (neighbourhood) SES was examined via stratification. We observed stronger air pollution-paediatric asthma associations in 'deprivation areas' (eg, ≥20% of the ZCTA population living in poverty) compared with 'non-deprivation areas'. When stratifying analyses by quartiles of neighbourhood SES, ORs indicated stronger associations in the highest and lowest SES quartiles and weaker associations among the middle quartiles. Our results suggest that neighbourhood-level SES is a factor contributing vulnerability to air pollution-related paediatric asthma morbidity in Atlanta. Children living in low SES environments appear to be especially vulnerable given positive ORs and high underlying asthma ED rates. Inconsistent findings of effect modification among previous studies may be partially explained by choice of SES stratification criteria, and the use of multiplicative models combined with differing baseline risk across SES populations. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/.
NASA Technical Reports Server (NTRS)
Dixon, G. V.; Barringer, S. R.; Gray, C. E.; Leatherman, A. D.
1975-01-01
Computer programs and resulting tabulations are presented of pipeline length-to-diameter ratios as a function of Mach number and pressure ratios for compressible flow. The tabulations are applicable to air, nitrogen, oxygen, and hydrogen for compressible isothermal flow with friction and compressible adiabatic flow with friction. Also included are equations for the determination of weight flow. The tabulations presented cover a wider range of Mach numbers for choked, adiabatic flow than available from commonly used engineering literature. Additional information presented, but which is not available from this literature, is unchoked, adiabatic flow over a wide range of Mach numbers, and choked and unchoked, isothermal flow for a wide range of Mach numbers.
Boundary-layer measurements on a transonic low-aspect ratio wing
NASA Technical Reports Server (NTRS)
Keener, Earl R.
1985-01-01
Tabulations and plots are presented of boundary-layer velocity and flow-direction surveys from wind-tunnel tests of a large-scale (0.90 m semi-span) model of the NASA/Lockheed Wing C. This wing is a generic, transonic, supercritical, highly three-dimensional, low-aspect-ratio configuration designed with the use of a three-dimensional, transonic full-potential-flow wing code (FLO22). Tests were conducted at the design angle of attack of 5 deg over a Mach number range from 0.25 to 0.96 and a Reynolds number range of 3.4x10 to the 6th power. Wing pressures were measured at five span stations, and boundary-layer surveys were measured at the midspan station. The data are presented without analysis.
Population ecology of the mallard: VII. Distribution and derivation of the harvest
Munro, Robert E.; Kimball, Charles F.
1982-01-01
This is the seventh in a series of comprehensive reports on population ecology of the mallard (Anas platyrhynchos) in North America. Banding records for 1961-1975 were used, together with information from previous reports in this series, to estimate annual and average preseason age and sex structure of the mallard population and patterns of harvest distribution and derivation. Age ratios in the pre-season population averaged 0.98 immatures per adult and ranged from 0.75 to 1.44. The adult male per female ration averaged 1.42. The young male per female ratio average 1.01. Geographic and annual differences in recovery distributions were associated with age, sex, and years after banding. Such variation might indicate that survival or band recovery rates, or both, change as a function of number of years after banding, and that estimates of these rates might thus be affected. Distribution of the mallard harvest from 16 major breeding ground reference areas to States, Provinces, and flyways is tabulated and illustrated. Seasonal (weekly) breeding ground derivation of the harvest within States and Provinces from the 16 reference areas also is tabulated. Harvest distributions, derivation, and similarity of derivation between harvest areas are summarily illustrated with maps. Derivation of harvest appears to be consistent throughout the hunting season in the middle and south central United States, encompassing States in both the Central and Mississippi flyways. However, weekly derivation patterns for most northern States suggest that early dates of hunting result in relatively greater harvest of locally derived mallard, in contrast to birds from more northern breeding areas.
NASA Astrophysics Data System (ADS)
Cuntz, Matthias; Mai, Juliane; Samaniego, Luis; Clark, Martyn; Wulfmeyer, Volker; Branch, Oliver; Attinger, Sabine; Thober, Stephan
2016-09-01
Land surface models incorporate a large number of process descriptions, containing a multitude of parameters. These parameters are typically read from tabulated input files. Some of these parameters might be fixed numbers in the computer code though, which hinder model agility during calibration. Here we identified 139 hard-coded parameters in the model code of the Noah land surface model with multiple process options (Noah-MP). We performed a Sobol' global sensitivity analysis of Noah-MP for a specific set of process options, which includes 42 out of the 71 standard parameters and 75 out of the 139 hard-coded parameters. The sensitivities of the hydrologic output fluxes latent heat and total runoff as well as their component fluxes were evaluated at 12 catchments within the United States with very different hydrometeorological regimes. Noah-MP's hydrologic output fluxes are sensitive to two thirds of its applicable standard parameters (i.e., Sobol' indexes above 1%). The most sensitive parameter is, however, a hard-coded value in the formulation of soil surface resistance for direct evaporation, which proved to be oversensitive in other land surface models as well. Surface runoff is sensitive to almost all hard-coded parameters of the snow processes and the meteorological inputs. These parameter sensitivities diminish in total runoff. Assessing these parameters in model calibration would require detailed snow observations or the calculation of hydrologic signatures of the runoff data. Latent heat and total runoff exhibit very similar sensitivities because of their tight coupling via the water balance. A calibration of Noah-MP against either of these fluxes should therefore give comparable results. Moreover, these fluxes are sensitive to both plant and soil parameters. Calibrating, for example, only soil parameters hence limit the ability to derive realistic model parameters. It is thus recommended to include the most sensitive hard-coded model parameters that were exposed in this study when calibrating Noah-MP.
Environmental Justice Assessment for Transportation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mills, G.S.; Neuhauser, K.S.
1999-04-05
Application of Executive Order 12898 to risk assessment of highway or rail transport of hazardous materials has proven difficult; the location and conditions affecting the propagation of a plume of hazardous material released in a potential accident are unknown, in general. Therefore, analyses have only been possible in geographically broad or approximate manner. The advent of geographic information systems and development of software enhancements at Sandia National Laboratories have made kilometer-by-kilometer analysis of populations tallied by U.S. Census Blocks along entire routes practicable. Tabulations of total, or racially/ethnically distinct, populations close to a route, its alternatives, or the broader surroundingmore » area, can then be compared and differences evaluated statistically. This paper presents methods of comparing populations and their racial/ethnic compositions using simple tabulations, histograms and Chi Squared tests for statistical significance of differences found. Two examples of these methods are presented: comparison of two routes and comparison of a route with its surroundings.« less
Tendal, Britta; Hróbjartsson, Asbjørn; Lundh, Andreas; Gøtzsche, Peter C
2014-01-01
Objective To assess the effects of coding and coding conventions on summaries and tabulations of adverse events data on suicidality within clinical study reports. Design Systematic electronic search for adverse events of suicidality in tables, narratives, and listings of adverse events in individual patients within clinical study reports. Where possible, for each event we extracted the original term reported by the investigator, the term as coded by the medical coding dictionary, medical coding dictionary used, and the patient’s trial identification number. Using the patient’s trial identification number, we attempted to reconcile data on the same event between the different formats for presenting data on adverse events within the clinical study report. Setting 9 randomised placebo controlled trials of duloxetine for major depressive disorder submitted to the European Medicines Agency for marketing approval. Data sources Clinical study reports obtained from the EMA in 2011. Results Six trials used the medical coding dictionary COSTART (Coding Symbols for a Thesaurus of Adverse Reaction Terms) and three used MedDRA (Medical Dictionary for Regulatory Activities). Suicides were clearly identifiable in all formats of adverse event data in clinical study reports. Suicide attempts presented in tables included both definitive and provisional diagnoses. Suicidal ideation and preparatory behaviour were obscured in some tables owing to the lack of specificity of the medical coding dictionary, especially COSTART. Furthermore, we found one event of suicidal ideation described in narrative text that was absent from tables and adverse event listings of individual patients. The reason for this is unclear, but may be due to the coding conventions used. Conclusion Data on adverse events in tables in clinical study reports may not accurately represent the underlying patient data because of the medical dictionaries and coding conventions used. In clinical study reports, the listings of adverse events for individual patients and narratives of adverse events can provide additional information, including original investigator reported adverse event terms, which can enable a more accurate estimate of harms. PMID:24899651
NASA Technical Reports Server (NTRS)
Best, R. G.; Moore, D. G.; Myers, V. I.
1977-01-01
An inventory of wetlands for the area included in the proposed Oahe irrigation project was conducted to provide supplemental data for the wildlife mitigation plan. Interpretation techniques for inventoring small wetlands in the low relief terrain of the Lake Dakota Plain were documented and data summaries included. The data were stored and tabulated in a computerized spatial data analysis system.
The Final Skylab Mission: Man at Home and at Work in Space
NASA Technical Reports Server (NTRS)
1974-01-01
The accomplishments of the Skylab 4 mission are discussed. The medical experiments and dietary aspects of the mission are reported. The observation of the Comet Kohoutek is described. The remote sensing of earth resources is examined to show the areas of coverage. The repair of the space station and the accomplishment of unscheduled requirements are discussed. Statistical data of all the Skylab missions are tabulated.
Engine-Airframe Integration for Rotorcraft.
1986-05-01
detailed technical literature. Initial attempts were to apply this existing fixed-wing data ; but, in general, only small selective areas were applicable...tabulated loss coefficients for a great variety of geometries; two excellent ones are the SAE Manual and GE Data books (References 2 and 3). The loss...pressure drop due to friction is somewhat more complicated because friction factor data given in the literature applies to fully developed duct flow
The Effects of Tank Crew Turbulence on Tank Gunnery Performance
1978-09-01
complete. Crewmen’s responses were converted to mouths for all itqms and tabulated for analysis. Because data was tabulated to two digits a maximum...two- digit data tabulation, mean and standard deviation statistics are somewhat conserva- tive for items 8, 9, and 10. There were 14-1S% of the TCs who...Benjamin Harrison. ATTN Libary I HQOA (DAMA-ARI I USAPACDC. Ft SBenjamuun Haritson, ATTN; ATCP-IHR I HOCIA OAPE HRE PO) IUSA Comrm- Elect Sch
Techniques for land use change detection using Landsat imagery
NASA Technical Reports Server (NTRS)
Angelici, G. L.; Bryant, N. A.; Friedman, S. Z.
1977-01-01
A variety of procedures were developed for the delineation of areas of land use change using Landsat Multispectral Scanner data and the generation of statistics revealing the nature of the changes involved (i.e., number of acres changed from rural to urban). Techniques of the Image Based Information System were utilized in all stages of the procedure, from logging the Landsat data and registering two frames of imagery, to extracting the changed areas and printing tabulations of land use change in acres. Two alternative methods of delineating land use change are presented while enumerating the steps of the entire process. The Houston, Texas urban area, and the Orlando, Florida urban area, are used as illustrative examples of various procedures.
Danskin, Wesley R.; McPherson, Kelly R.; Woolfenden, Linda R.
2006-01-01
The San Bernardino area of southern California has complex water-management issues. As an aid to local water managers, this report provides an integrated analysis of the surface-water and ground-water systems, documents ground-water flow and constrained optimization models, and provides seven examples using the models to better understand and manage water resources of the area. As an aid to investigators and water managers in other areas, this report provides an expanded description of constrained optimization techniques and how to use them to better understand the local hydrogeology and to evaluate inter-related water-management problems. In this report, the hydrology of the San Bernardino area, defined as the Bunker Hill and Lytle Creek basins, is described and quantified for calendar years 1945-98. The major components of the surface-water system are identified, and a routing diagram of flow through these components is provided. Annual surface-water inflow and outflow for the area are tabulated using gaged measurements and estimated values derived from linear-regression equations. Average inflow for the 54-year period (1945-98) was 146,452 acre-feet per year; average outflow was 67,931 acre-feet per year. The probability of exceedance for annual surface-water inflow is calculated using a Log Pearson Type III analysis. Cumulative surface-water inflow and outflow and ground-water-level measurements indicate that the relation between the surface-water system and the ground-water system changed in about 1951, in about 1979, and again in about 1992. Higher ground-water levels prior to 1951 and between 1979 and 1992 induced ground-water discharge to Warm Creek. This discharge was quantified using streamflow measurements and can be estimated for other time periods using ground-water levels from a monitoring well (1S/4W-3Q1) and a logarithmic-regression equation. Annual wastewater discharge from the area is tabulated for the major sewage and power-plant facilities. More...
1987-09-01
Pollutants by Gas Chromatographic Headspace Analysis. J. Chrom . 260:23-32. Miller, R. E. 1984. Confidence Intervals and Hypothesis Tests. Chem. Engr...tabulation of the injection peak areas, Henry’s law constant estimates, and Coefficient of Variation (COV) values for the component at five temperatures...I 15.1897 (4) I 14.5788 I 19.7121 1 16 6428 Injection: (1) 1 16158 I 2596 38628 Peak Area] (2) 1 154846 1 281438 1 261148 (3) 4673 1 64736 1 63322 (4
Making GRADE accessible: a proposal for graphic display of evidence quality assessments.
Khan, Khalid S; Borowiack, Ewa; Roos, Carolien; Kowalska, Monika; Zapalska, Anna; Mol, Ben W; Mignini, Luciano; Meads, Catherine; Walczak, Jacek
2011-06-01
When generating guidelines, quality of evidence is frequently reported in tabulated form capturing several domains, for example, study design, risk of bias and heterogeneity. Increasingly, this is done using the Grading of Recommendations Assessment, Development and Evaluation approach. As assimilating large amount of tabulated data across several comparisons and outcomes spread over many pages (sometimes hundreds) is not easy, there is a need to present evidence summaries in a more effective way. A graphic display plotting the several domains used in evidence grading on equiangular spokes starting from the same point, the data length of each spoke proportional to the magnitude of the quality, succinctly captures tabulated information. These plots allow easy identification of deficiencies, outliers and similarities in evidence quality for individual and multiple comparisons and outcomes, paving the way for their routine use alongside tabulated information.
ERIC Educational Resources Information Center
Ferguson, Ronald F.
This survey presents statistics on secondary school student culture by school district and race/ethnicity, using data from approximately forty thousand secondary school students in Minority Student Achievement Network districts. Information is presented in six areas: (1) "Family Background Resources and Living Arrangements (racial/ethnic…
47 CFR 52.19 - Area code relief.
Code of Federal Regulations, 2010 CFR
2010-10-01
... new area codes within their states. Such matters may include, but are not limited to: Directing... realignment; establishing new area code boundaries; establishing necessary dates for the implementation of... code relief planning encompasses all functions related to the implementation of new area codes that...
47 CFR 52.19 - Area code relief.
Code of Federal Regulations, 2011 CFR
2011-10-01
... new area codes within their states. Such matters may include, but are not limited to: Directing... realignment; establishing new area code boundaries; establishing necessary dates for the implementation of... code relief planning encompasses all functions related to the implementation of new area codes that...
Organ doses from radionuclides on the ground. Part I. Simple time dependences
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jacob, P.; Paretzke, H.G.; Rosenbaum, H.
1988-06-01
Organ dose equivalents of mathematical, anthropomorphical phantoms ADAM and EVA for photon exposures from plane sources on the ground have been calculated by Monte Carlo photon transport codes and tabulated in this article. The calculation takes into account the air-ground interface and a typical surface roughness, the energy and angular dependence of the photon fluence impinging on the phantom and the time dependence of the contributions from daughter nuclides. Results are up to 35% higher than data reported in the literature for important radionuclides. This manuscript deals with radionuclides, for which the time dependence of dose equivalent rates and dosemore » equivalents may be approximated by a simple exponential. A companion manuscript treats radionuclides with non-trivial time dependences.« less
Fifty year canon of solar eclipses: 1986 - 2035
NASA Technical Reports Server (NTRS)
Espenak, Fred
1987-01-01
A complete catalog is presented, listing the general characteristics of every solar eclipse from 1901 through 2100. To complement this catalog, a detailed set of cylindrical projection world maps shows the umbral paths of every solar eclipse over the 200 year interval. Focusing in on the next 50 years, accurate geodetic path coordinates and local circumstances for the 71 central eclipses from 1987 through 2035 are tabulated. Finally, the geodetic paths of the umbral and penumbral shadows of all 109 solar eclipses in this period are plotted on orthographic projection maps of the Earth. Appendices are included which discuss eclipse geometry, eclipse frequency and occurrence, modern eclipse prediction and time determination. Finally, code for a simple Fortran program is given to predict the occurrence and characteristics of solar eclipses.
1989-09-30
AD-A237 531 1989 SURVEY OF UNITED STATES ARMY RESERVE (USAR) TROOP PROGRAM UNIT (TPU) SOLDIERS Tabulation of Questionnaire Responses: Longitudinal...Program Unit (TPU) Soldiers . The Tabulation Volumes list questionnaire items and the percent of respondents (weighted to population estimates) who have...Reserve population eligible for selection was defined by the number of personnel rec,,rds on a Dpeber 1988 SIDPERS data base; this totalled 280,265
THE McGill PLANAR HYDROGEN ATMOSPHERE CODE (McPHAC)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Haakonsen, Christian Bernt; Turner, Monica L.; Tacik, Nick A.
2012-04-10
The McGill Planar Hydrogen Atmosphere Code (McPHAC) v1.1 calculates the hydrostatic equilibrium structure and emergent spectrum of an unmagnetized hydrogen atmosphere in the plane-parallel approximation, at surface gravities appropriate for neutron stars. McPHAC incorporates several improvements over previous codes for which tabulated model spectra are available: (1) Thomson scattering is treated anisotropically, which is shown to result in a 0.2%-3% correction in the emergent spectral flux across the 0.1-5 keV passband; (2) the McPHAC source code is made available to the community, allowing it to be scrutinized and modified by other researchers wishing to study or extend its capabilities; andmore » (3) the numerical uncertainty resulting from the discrete and iterative solution is studied as a function of photon energy, indicating that McPHAC is capable of producing spectra with numerical uncertainties <0.01%. The accuracy of the spectra may at present be limited to {approx}1%, but McPHAC enables researchers to study the impact of uncertain inputs and additional physical effects, thereby supporting future efforts to reduce those inaccuracies. Comparison of McPHAC results with spectra from one of the previous model atmosphere codes (NSA) shows agreement to {approx}<1% near the peaks of the emergent spectra. However, in the Wien tail a significant deficit of flux in the spectra of the previous model is revealed, determined to be due to the previous work not considering large enough optical depths at the highest photon frequencies. The deficit is most significant for spectra with T{sub eff} < 10{sup 5.6} K, though even there it may not be of much practical importance for most observations.« less
FUEL-FLEXIBLE GASIFICATION-COMBUSTION TECHNOLOGY FOR PRODUCTION OF H2 AND SEQUESTRATION-READY CO2
DOE Office of Scientific and Technical Information (OSTI.GOV)
George Rizeq; Janice West; Arnaldo Frydman
Further development of a combustion Large Eddy Simulation (LES) code for the design of advanced gaseous combustion systems is described in this sixth quarterly report. CFD Research Corporation (CFDRC) is developing the LES module within the parallel, unstructured solver included in the commercial CFD-ACE+ software. In this quarter, in-situ adaptive tabulation (ISAT) for efficient chemical rate storage and retrieval was implemented and tested within the Linear Eddy Model (LEM). ISAT type 3 is being tested so that extrapolation can be performed and further improve the retrieval rate. Further testing of the LEM for subgrid chemistry was performed for parallel applicationsmore » and for multi-step chemistry. Validation of the software on backstep and bluff-body reacting cases were performed. Initial calculations of the SimVal experiment at Georgia Tech using their LES code were performed. Georgia Tech continues the effort to parameterize the LEM over composition space so that a neural net can be used efficiently in the combustion LES code. A new and improved Artificial Neural Network (ANN), with log-transformed output, for the 1-step chemistry was implemented in CFDRC's LES code and gave reasonable results. This quarter, the 2nd consortium meeting was held at CFDRC. Next quarter, LES software development and testing will continue. Alpha testing of the code will continue to be performed on cases of interest to the industrial consortium. Optimization of subgrid models will be pursued, particularly with the ISAT approach. Also next quarter, the demonstration of the neural net approach, for multi-step chemical kinetics speed-up in CFD-ACE+, will be accomplished.« less
The McGill Planar Hydrogen Atmosphere Code (McPHAC)
NASA Astrophysics Data System (ADS)
Haakonsen, Christian Bernt; Turner, Monica L.; Tacik, Nick A.; Rutledge, Robert E.
2012-04-01
The McGill Planar Hydrogen Atmosphere Code (McPHAC) v1.1 calculates the hydrostatic equilibrium structure and emergent spectrum of an unmagnetized hydrogen atmosphere in the plane-parallel approximation, at surface gravities appropriate for neutron stars. McPHAC incorporates several improvements over previous codes for which tabulated model spectra are available: (1) Thomson scattering is treated anisotropically, which is shown to result in a 0.2%-3% correction in the emergent spectral flux across the 0.1-5 keV passband; (2) the McPHAC source code is made available to the community, allowing it to be scrutinized and modified by other researchers wishing to study or extend its capabilities; and (3) the numerical uncertainty resulting from the discrete and iterative solution is studied as a function of photon energy, indicating that McPHAC is capable of producing spectra with numerical uncertainties <0.01%. The accuracy of the spectra may at present be limited to ~1%, but McPHAC enables researchers to study the impact of uncertain inputs and additional physical effects, thereby supporting future efforts to reduce those inaccuracies. Comparison of McPHAC results with spectra from one of the previous model atmosphere codes (NSA) shows agreement to lsim1% near the peaks of the emergent spectra. However, in the Wien tail a significant deficit of flux in the spectra of the previous model is revealed, determined to be due to the previous work not considering large enough optical depths at the highest photon frequencies. The deficit is most significant for spectra with T eff < 105.6 K, though even there it may not be of much practical importance for most observations.
McPHAC: McGill Planar Hydrogen Atmosphere Code
NASA Astrophysics Data System (ADS)
Haakonsen, Christian Bernt; Turner, Monica L.; Tacik, Nick A.; Rutledge, Robert E.
2012-10-01
The McGill Planar Hydrogen Atmosphere Code (McPHAC) v1.1 calculates the hydrostatic equilibrium structure and emergent spectrum of an unmagnetized hydrogen atmosphere in the plane-parallel approximation at surface gravities appropriate for neutron stars. McPHAC incorporates several improvements over previous codes for which tabulated model spectra are available: (1) Thomson scattering is treated anisotropically, which is shown to result in a 0.2%-3% correction in the emergent spectral flux across the 0.1-5 keV passband; (2) the McPHAC source code is made available to the community, allowing it to be scrutinized and modified by other researchers wishing to study or extend its capabilities; and (3) the numerical uncertainty resulting from the discrete and iterative solution is studied as a function of photon energy, indicating that McPHAC is capable of producing spectra with numerical uncertainties <0.01%. The accuracy of the spectra may at present be limited to ~1%, but McPHAC enables researchers to study the impact of uncertain inputs and additional physical effects, thereby supporting future efforts to reduce those inaccuracies. Comparison of McPHAC results with spectra from one of the previous model atmosphere codes (NSA) shows agreement to lsim1% near the peaks of the emergent spectra. However, in the Wien tail a significant deficit of flux in the spectra of the previous model is revealed, determined to be due to the previous work not considering large enough optical depths at the highest photon frequencies. The deficit is most significant for spectra with T eff < 105.6 K, though even there it may not be of much practical importance for most observations.
Nowlan, G.A.; Ficklin, Walter H.; Dover, Robert A.
1985-01-01
This report presents results of geochemical studies carried out in June and July of 1982 in the Buffalo Peaks Wilderness Study Area, Colo. (see index map). Samples of water were collected from 84 streams and 18 springs draining the study area. Tabulations of the analyses and a sample locality map are in Ficklin and others (1984). The geochemistry of stream sediments and panned concentrates of the study area is in Nowlan and Gerstel (1985). The geology of the study area and vicinity is in Hedlund (1985). The mineral resource potential of the study area is described in Hedlund and others (1983). This report (1) assists in the assessment of the mineral resource potential of the Buffalo Peaks Wilderness Study Area; and (2) compares analyses of water samples with analyses of stream-sediment and panned-concentrate samples (Nowlan and Gerstel, 1985).
23 CFR 635.113 - Bid opening and bid tabulations.
Code of Federal Regulations, 2010 CFR
2010-04-01
... CONSTRUCTION AND MAINTENANCE Contract Procedures § 635.113 Bid opening and bid tabulations. (a) All bids... contractors, during the period following the opening of bids and before the award of the contract shall not be...
Reduced description of reactive flows with tabulation of chemistry
NASA Astrophysics Data System (ADS)
Ren, Zhuyin; Goldin, Graham M.; Hiremath, Varun; Pope, Stephen B.
2011-12-01
The direct use of large chemical mechanisms in multi-dimensional Computational Fluid Dynamics (CFD) is computationally expensive due to the large number of chemical species and the wide range of chemical time scales involved. To meet this challenge, a reduced description of reactive flows in combination with chemistry tabulation is proposed to effectively reduce the computational cost. In the reduced description, the species are partitioned into represented species and unrepresented species; the reactive system is described in terms of a smaller number of represented species instead of the full set of chemical species in the mechanism; and the evolution equations are solved only for the represented species. When required, the unrepresented species are reconstructed assuming that they are in constrained chemical equilibrium. In situ adaptive tabulation (ISAT) is employed to speed the chemistry calculation through tabulating information of the reduced system. The proposed dimension-reduction / tabulation methodology determines and tabulates in situ the necessary information of the nr-dimensional reduced system based on the ns-species detailed mechanism. Compared to the full description with ISAT, the reduced descriptions achieve additional computational speed-up by solving fewer transport equations and faster ISAT retrieving. The approach is validated in both a methane/air premixed flame and a methane/air non-premixed flame. With the GRI 1.2 mechanism consisting of 31 species, the reduced descriptions (with 12 to 16 represented species) achieve a speed-up factor of up to three compared to the full description with ISAT, with a relatively moderate decrease in accuracy compared to the full description.
Shen, Xinglei; Showalter, Timothy N; Mishra, Mark V; Barth, Sanford; Rao, Vijay; Levin, David; Parker, Laurence
2014-07-01
We evaluated long-term changes in the volume and payments for radiation oncology services in the intensity-modulated radiation therapy (IMRT) era from 2000 to 2010 using a database of Medicare claims. We used the Medicare Physician/Supplier Procedure Summary Master File (PSPSMF) for each year from 2000 to 2010 to tabulate the volume and payments for radiation oncology services. This database provides a summary of each billing code submitted to Medicare part B. We identified all codes used in radiation oncology services and categorized billing codes by treatment modality and place of service. We focused our analysis on office-based practices. Total office-based patient volume increased 8.2% from 2000 to 2010, whereas total payments increased 217%. Increase in overall payments increased dramatically from 2000 to 2007, but subsequently plateaued from 2008 to 2010. Increases in complexity of care, and image guidance in particular, have also resulted in higher payments. The cost of radiation oncology services increased from 2000 to 2010, mostly due to IMRT, but also with significant contribution from increased overall complexity of care. A cost adjustment occurred after 2007, limiting further growth of payments. Future health policy studies should explore the potential for further cost containment, including differences in use between freestanding and hospital outpatient facilities. Copyright © 2014 by American Society of Clinical Oncology.
Combustor Computations for CO2-Neutral Aviation
NASA Technical Reports Server (NTRS)
Hendricks, Robert C.; Brankovic, Andreja; Ryder, Robert C.; Huber, Marcia
2011-01-01
Knowing the pure component C(sub p)(sup 0) or mixture C(sub p) (sup 0) as computed by a flexible code such as NIST-STRAPP or McBride-Gordon, one can, within reasonable accuracy, determine the thermophysical properties necessary to predict the combustion characteristics when there are no tabulated or computed data for those fluid mixtures 3or limited results for lower temperatures. (Note: C(sub p) (sup 0) is molar heat capacity at constant pressure.) The method can be used in the determination of synthetic and biological fuels and blends using the NIST code to compute the C(sub p) (sup 0) of the mixture. In this work, the values of the heat capacity were set at zero pressure, which provided the basis for integration to determine the required combustor properties from the injector to the combustor exit plane. The McBride-Gordon code was used to determine the heat capacity at zero pressure over a wide range of temperatures (room to 6,000 K). The selected fluids were Jet-A, 224TMP (octane), and C12. It was found that each heat capacity loci were form-similar. It was then determined that the results [near 400 to 3,000 K] could be represented to within acceptable engineering accuracy with the simplified equation C(sub p) (sup 0) = A/T + B, where A and B are fluid-dependent constants and T is temperature (K).
Analysis of space tug operating techniques (study 2.4). Volume 1: Executive summary
NASA Technical Reports Server (NTRS)
1972-01-01
The costs of tug refurbishment were studied, using existing cost estimating relationships, to establish the cost of maintaining the reusable third stage of the space transportation system. Refurbishment operations sheets which describe the actual tasks that are necessary to keep the equipment functioning properly were used along with refurbishment operations sheets which contain all of the pertinent descriptive information for each of the major vehicle areas. Tug refurbishment costs per mission are tabulated.
R2 & NE Block Group - 2010 Census; Housing and Population Summary
The TIGER/Line Files are shapefiles and related database files (.dbf) that are an extract of selected geographic and cartographic information from the U.S. Census Bureau's Master Address File / Topologically Integrated Geographic Encoding and Referencing (MAF/TIGER) Database (MTDB). The MTDB represents a seamless national file with no overlaps or gaps between parts, however, each TIGER/Line File is designed to stand alone as an independent data set, or they can be combined to cover the entire nation. Block Groups (BGs) are defined before tabulation block delineation and numbering, but are clusters of blocks within the same census tract that have the same first digit of their 4-digit census block number from the same decennial census. For example, Census 2000 tabulation blocks 3001, 3002, 3003,.., 3999 within Census 2000 tract 1210.02 are also within BG 3 within that census tract. Census 2000 BGs generally contained between 600 and 3,000 people, with an optimum size of 1,500 people. Most BGs were delineated by local participants in the Census Bureau's Participant Statistical Areas Program (PSAP). The Census Bureau delineated BGs only where the PSAP participant declined to delineate BGs or where the Census Bureau could not identify any local PSAP participant. A BG usually covers a contiguous area. Each census tract contains at least one BG, and BGs are uniquely numbered within census tract. Within the standard census geographic hierarchy, BGs never cross
Collins, Dannie L.; Flynn, Kathleen M.
1979-01-01
This report summarizes and makes available to other investigators the measured hydraulic data collected during a series of experiments designed to study the effect of patterned bed roughness on steady and unsteady open-channel flow. The patterned effect of the roughness was obtained by clear-cut mowing of designated areas of an otherwise fairly dense coverage of coastal Bermuda grass approximately 250 mm high. All experiments were conducted in the Flood Plain Simulation Facility during the period of October 7 through December 12, 1974. Data from 18 steady flow experiments and 10 unsteady flow experiments are summarized. Measured data included are ground-surface elevations, grass heights and densities, water-surface elevations and point velocities for all experiments. Additional tables of water-surface elevations and measured point velocities are included for the clear-cut areas for most experiments. One complete set of average water-surface elevations and one complete set of measured point velocities are tabulated for each steady flow experiment. Time series data, on a 2-minute time interval, are tabulated for both water-surface elevations and point velocities for each unsteady flow experiment. All data collected, including individual records of water-surface elevations for the steady flow experiments, have been stored on computer disk storage and can be retrieved using the computer programs listed in the attachment to this report. (Kosco-USGS)
Trogdon, Justin G; Shafer, Paul; Lindsay, Brianna; Coyne-Beasley, Tamera
2018-02-28
The objective of this study was to evaluate the impact of introduction of 9vHPV vaccine on HPV vaccination uptake (doses per capita) and initiation (≥1 doses), completion (≥3 doses) and compliance (≥3 doses within 12 months) by adolescents. We used a retrospective cohort analysis using North Carolina Immunization Registry (NCIR) data from January 2008 through October 2016. The sample included Vaccines for Children eligible adolescents aged 9 to 17 years in 2016, for whom the NCIR contains complete vaccination history. We applied an interrupted time series design to measure associations between ZIP Code Tabulation Area (ZCTA)-level HPV vaccination outcomes over time with the introduction of 9vHPV in North Carolina (NC) in July 2015. Each outcome displayed a linear upward trend over time with large seasonal spikes near August of each year, corresponding to the time when adolescents often receive other vaccines required for school entry. After accounting for these underlying trends, introduction of 9vHPV was not associated with a change in publicly funded HPV vaccination rates in NC. Our results indicate that 9vHPV substituted for 4vHPV in the first year after release in NC, but the release of 9vHPV was not associated with an overall change in HPV vaccination. Copyright © 2018 Elsevier Ltd. All rights reserved.
Yamanouchi, Satoshi; Ishii, Tadashi; Morino, Kazuma; Furukawa, Hajime; Hozawa, Atsushi; Ochi, Sae; Kushimoto, Shigeki
2014-12-01
When disasters that affect a wide area occur, external medical relief teams play a critical role in the affected areas by helping to alleviate the burden caused by surging numbers of individuals requiring health care. Despite this, no system has been established for managing deployed medical relief teams during the subacute phase following a disaster. After the Great East Japan Earthquake and tsunami, the Ishinomaki Medical Zone was the most severely-affected area. Approximately 6,000 people died or were missing, and the immediate evacuation of approximately 120,000 people to roughly 320 shelters was required. As many as 59 medical teams came to participate in relief activities. Daily coordination of activities and deployment locations became a significant burden to headquarters. The Area-based/Line-linking Support System (Area-Line System) was thus devised to resolve these issues for medical relief and coordinating activities. A retrospective analysis was performed to examine the effectiveness of the medical relief provided to evacuees using the Area-Line System with regards to the activities of the medical relief teams and the coordinating headquarters. The following were compared before and after establishment of the Area-Line System: (1) time required at the coordinating headquarters to collect and tabulate medical records from shelters visited; (2) time required at headquarters to determine deployment locations and activities of all medical relief teams; and (3) inter-area variation in number of patients per team. The time required to collect and tabulate medical records was reduced from approximately 300 to 70 minutes/day. The number of teams at headquarters required to sort through data was reduced from 60 to 14. The time required to determine deployment locations and activities of the medical relief teams was reduced from approximately 150 hours/month to approximately 40 hours/month. Immediately prior to establishment of the Area-Line System, the variation of the number of patients per team was highest. Variation among regions did not increase after establishment of the system. This descriptive analysis indicated that implementation of the Area-Line System, a systematic approach for long-term disaster medical relief across a wide area, can increase the efficiency of relief provision to disaster-stricken areas.
Trauma with Injury Severity Score of 75: Are These Unsurvivable Injuries?
Peng, Jin; Wheeler, Krista; Shi, Junxin; Groner, Jonathan Ira; Haley, Kathryn Jo; Xiang, Huiyun
2015-01-01
Trauma patients with an ISS=75 have been deliberately excluded from some trauma studies because they were assumed to have "unsurvivable injuries." This study aimed to assess the true mortality among patients with an ISS=75, and to examine the characteristics and primary diagnoses of these patients. Retrospective review of the 2006-2010 U.S. Nationwide Emergency Department Sample (NEDS) generated 2,815 patients with an ISS=75 for analysis, representing an estimated 13,569 patients in the country. Dispositions from the emergency department and hospital for these patients were tabulated by trauma center level. Survivors and non-survivors were compared using Pearson's chi-square test. Primary diagnosis codes of these patients were tabulated by mortality status. Overall, about 48.6% of patients with an ISS=75 were discharged alive, 25.8% died and 25.6% had unknown mortality status. The mortality risks of these patients did not vary significantly across different levels of trauma centers (15.6% vs. 13.0%, P = 0.16). Non-survivors were more likely than survivors to: be male (81.2% vs. 74.4%, P < 0.0001), be over 65 years (20.3% vs. 10.2%, P < 0.0001), be uninsured (33.8% vs. 19.1%), have at least one chronic condition (58.0% vs. 43.7%, P <0.0001), sustain life-threatening injuries (79.2% vs. 49.4%, P<0.0001), sustain penetrating injuries (42.0% vs. 25.9%, P<0.0001), and have injuries caused by motor vehicle crashes (32.9% vs. 21.1%, P<0.0001) or firearms (21.9% vs. 4.4%, P<0.0001). The most frequent diagnosis code was 862.8 (injury to multiple and unspecified intrathoracic organs, without mention of open wound into cavity). Our results revealed that at least half of patients with an ISS=75 survived, demonstrating that the rationale for excluding patients with an ISS=75 from analysis is not always justified. To avoid bias and inaccurate results, trauma researchers should examine the mortality status of patients with an ISS=75 before exclusion, and explicitly describe their method of generating ISS scores. PMID:26230931
DOT National Transportation Integrated Search
1976-03-01
The report gives tabulations of survey responses which were collected in Morgantown, West Virginia, as part of a study to assess the impact of the installation of the Personal Rapid Transit (PRT) System.
Hydrogen technology survey: Thermophysical properties
NASA Technical Reports Server (NTRS)
Mccarty, R. D.
1975-01-01
The thermodynamic functions, transport properties, and physical properties of both liquid and gaseous hydrogen are presented. The low temperature regime is emphasized. The tabulation of the properties of normal hydrogen in both Si and engineering units is given along with the tabulation of parahydrogen.
NASA Technical Reports Server (NTRS)
Watkins, Charles E; Durling, Barbara J
1956-01-01
This report presents tabulated values of certain definite integral that are involved in the calculation of near-field propeller noise when the chordwise forces are assumed to be either uniform or of a Dirac delta type. The tabulations are over a wide range of operating conditions and are useful for estimating propeller noise when either the concept of an effective radius or radial distributions of forces are considered. Use of the tabulations is illustrated by several examples of calculated results for some specific propellers.
NASA Technical Reports Server (NTRS)
Strahler, A. H.; Woodcock, C. E.; Logan, T. L.
1983-01-01
A timber inventory of the Eldorado National Forest, located in east-central California, provides an example of the use of a Geographic Information System (GIS) to stratify large areas of land for sampling and the collection of statistical data. The raster-based GIS format of the VICAR/IBIS software system allows simple and rapid tabulation of areas, and facilitates the selection of random locations for ground sampling. Algorithms that simplify the complex spatial pattern of raster-based information, and convert raster format data to strings of coordinate vectors, provide a link to conventional vector-based geographic information systems.
Blood analyses of wolf pups and their ecological and metabolic interpretation
Seal, U.S.; Mech, L.D.; Van Ballenberghe, V.
1975-01-01
Blood samples were obtained from 32 wolf (Canis lupus) pups live-trapped over a three-year period in northern Minnesota. The results of 21 laboratory analyses of hematology and blood chemistry are tabulated and analyzed in terms of study area, age, sex, and year of co11ection. Mean values are compared to those reported for dogs in the same age group. The numerous differences between dog and wolf pups are interpreted in terms of nutritional levels and dietary composition with the suggestion that the wolves are not achieving their full growth potential. Individual abnormal test results are tabulated and possible interpretations are suggested. Abnormal results were observed in 13 animals including 10 of 11 animals sampled in 1972. The results in the 1972 animals indicated a poorer nutrition. This preponderance of abnormal test results in pups from 1972 is correlated with ecological studies on this wolf population indicating decreased survival. The potential value of such long-term integrated field and laboratory studies for providing a more complete understanding of changes in the dynamics of natural populations in terms of the responses of individual animals is demonstrated.
RSRA sixth scale wind tunnel test. [of scale model of Sikorsky Whirlwind Helicopter
NASA Technical Reports Server (NTRS)
Flemming, R.; Ruddell, A.
1974-01-01
The sixth scale model of the Sikorsky/NASA/Army rotor systems research aircraft was tested in an 18-foot section of a large subsonic wind tunnel for the purpose of obtaining basic data in the areas of performance, stability, and body surface loads. The model was mounted in the tunnel on the struts arranged in tandem. Basic testing was limited to forward flight with angles of yaw from -20 to +20 degrees and angles of attack from -20 to +25 degrees. Tunnel test speeds were varied up to 172 knots (q = 96 psf). Test data were monitored through a high speed static data acquisition system, linked to a PDP-6 computer. This system provided immediate records of angle of attack, angle of yaw, six component force and moment data, and static and total pressure information. The wind tunnel model was constructed of aluminum structural members with aluminum, fiberglass, and wood skins. Tabulated force and moment data, flow visualization photographs, tabulated surface pressure data are presented for the basic helicopter and compound configurations. Limited discussions of the results of the test are included.
Tornado climatology of the contiguous United States
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ramsdell, J.V.; Andrews, G.L.
1986-05-01
The characteristics of tornadoes that were reported in the contiguous United States for the period from January 1, 1954, through December 31, 1983, have been computed from data in the National Severe Storms Forecast Center tornado data base. The characteristics summarized in this report include frequency and locations of tornadoes, and their lengths, widths, and areas. Tornado strike and intensity probabilities have been estimated on a regional basis, and these estimates have been used to compute wind speeds with 10/sup -5/, 10/sup -6/, and 10/sup -7/ yr/sup -1/ probabilities of occurrence. The 10/sup -7/ yr/sup -1/ wind speeds range frommore » below 200 mph in the western United States to about 330 mph in the vicinity of Kansas and Nebraska. The appendices contain extensive tabulations of tornado statistics. Variations of the characteristics within the contiguous United States are presented in the summaries. Separate tabulations are provided for the contiguous United States, for each state, for each 5/sup 0/ and 1/sup 0/ latitude and longitude box, and for the eastern and western United States.« less
Saltus, R.W.; Kulander, Christopher S.; Potter, Christopher J.
2002-01-01
We have digitized, modified, and analyzed seismic interpretation maps of 12 subsurface stratigraphic horizons spanning portions of the National Petroleum Reserve in Alaska (NPRA). These original maps were prepared by Tetra Tech, Inc., based on about 15,000 miles of seismic data collected from 1974 to 1981. We have also digitized interpreted faults and seismic velocities from Tetra Tech maps. The seismic surfaces were digitized as two-way travel time horizons and converted to depth using Tetra Tech seismic velocities. The depth surfaces were then modified by long-wavelength corrections based on recent USGS seismic re-interpretation along regional seismic lines. We have developed and executed an algorithm to identify and calculate statistics on the area, volume, height, and depth of closed structures based on these seismic horizons. These closure statistics are tabulated and have been used as input to oil and gas assessment calculations for the region. Directories accompanying this report contain basic digitized data, processed data, maps, tabulations of closure statistics, and software relating to this project.
Ray-tracing critical-angle transmission gratings for the X-ray Surveyor and Explorer-size missions
NASA Astrophysics Data System (ADS)
Günther, Hans M.; Bautz, Marshall W.; Heilmann, Ralf K.; Huenemoerder, David P.; Marshall, Herman L.; Nowak, Michael A.; Schulz, Norbert S.
2016-07-01
We study a critical angle transmission (CAT) grating spectrograph that delivers a spectral resolution significantly above any X-ray spectrograph ever own. This new technology will allow us to resolve kinematic components in absorption and emission lines of galactic and extragalactic matter down to unprecedented dispersion levels. We perform ray-trace simulations to characterize the performance of the spectrograph in the context of an X-ray Surveyor or Arcus like layout (two mission concepts currently under study). Our newly developed ray-trace code is a tool suite to simulate the performance of X-ray observatories. The simulator code is written in Python, because the use of a high-level scripting language allows modifications of the simulated instrument design in very few lines of code. This is especially important in the early phase of mission development, when the performances of different configurations are contrasted. To reduce the run-time and allow for simulations of a few million photons in a few minutes on a desktop computer, the simulator code uses tabulated input (from theoretical models or laboratory measurements of samples) for grating efficiencies and mirror reflectivities. We find that the grating facet alignment tolerances to maintain at least 90% of resolving power that the spectrometer has with perfect alignment are (i) translation parallel to the optical axis below 0.5 mm, (ii) rotation around the optical axis or the groove direction below a few arcminutes, and (iii) constancy of the grating period to 1:105. Translations along and rotations around the remaining axes can be significantly larger than this without impacting the performance.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chiang, Chih-Chieh; Lin, Hsin-Hon; Lin, Chang-Shiun
Abstract-Multiple-photon emitters, such as In-111 or Se-75, have enormous potential in the field of nuclear medicine imaging. For example, Se-75 can be used to investigate the bile acid malabsorption and measure the bile acid pool loss. The simulation system for emission tomography (SimSET) is a well-known Monte Carlo simulation (MCS) code in nuclear medicine for its high computational efficiency. However, current SimSET cannot simulate these isotopes due to the lack of modeling of complex decay scheme and the time-dependent decay process. To extend the versatility of SimSET for simulation of those multi-photon emission isotopes, a time-resolved multiple photon history generatormore » based on SimSET codes is developed in present study. For developing the time-resolved SimSET (trSimSET) with radionuclide decay process, the new MCS model introduce new features, including decay time information and photon time-of-flight information, into this new code. The half-life of energy states were tabulated from the Evaluated Nuclear Structure Data File (ENSDF) database. The MCS results indicate that the overall percent difference is less than 8.5% for all simulation trials as compared to GATE. To sum up, we demonstrated that time-resolved SimSET multiple photon history generator can have comparable accuracy with GATE and keeping better computational efficiency. The new MCS code is very useful to study the multi-photon imaging of novel isotopes that needs the simulation of lifetime and the time-of-fight measurements. (authors)« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
MacGregor, M.H.; Cullen, D.E.; Howerton, R.J.
1976-07-04
The bibliographic citations in the Experimental Cross Section Information Library (ECSIL) as of July 4, 1976 are tabulated. The tabulation has three arrangements: alphabetically by author, alphabetically by publication, and numerically by reference number.
7 CFR 900.308 - Tabulation of ballots.
Code of Federal Regulations, 2010 CFR
2010-01-01
... Conduct of Referenda To Determine Producer Approval of Milk Marketing Orders To Be Made Effective Pursuant to Agricultural Marketing Agreement Act of 1937, as Amended § 900.308 Tabulation of ballots. (a... Agriculture Regulations of the Department of Agriculture (Continued) AGRICULTURAL MARKETING SERVICE (Marketing...
An ethical paradox: the effect of unethical conduct on medical students' values.
Satterwhite, R C; Satterwhite, W M; Enarson, C
2000-12-01
To report the ethical development of medical students across four years of education at one medical school. A questionnaire was distributed to all four classes at the Wake Forest University School of Medicine during the Spring of 1996. Three hundred and three students provided demographic information as well as information concerning their ethical development both as current medical students and future interns. Results were analyzed using cross-tabulations, correlations, and analysis of variance. Results suggested that the observation of and participation in unethical conduct may have disparaging effects on medical students' codes of ethics with 35% of the total sample (24% of first years rising to 55% of fourth years) stating that derogatory comments made by residents/attendings, either in the patient's presence or absence, were "sometimes" or "often" appropriate. However, approximately 70% of the sample contended that their personal code of ethics had not changed since beginning medical school and would not change as a resident. Results may represent an internal struggle that detracts from the medical school experience, both as a person and as a doctor. Our goal as educators is to alter the educational environment so that acceptance of such behaviour is not considered part of becoming a physician.
NASA Technical Reports Server (NTRS)
Marvin, Joseph G.; Brown, James L.; Gnoffo, Peter A.
2013-01-01
A database compilation of hypersonic shock-wave/turbulent boundary layer experiments is provided. The experiments selected for the database are either 2D or axisymmetric, and include both compression corner and impinging type SWTBL interactions. The strength of the interactions range from attached to incipient separation to fully separated flows. The experiments were chosen based on criterion to ensure quality of the datasets, to be relevant to NASA's missions and to be useful for validation and uncertainty assessment of CFD Navier-Stokes predictive methods, both now and in the future. An emphasis on datasets selected was on surface pressures and surface heating throughout the interaction, but include some wall shear stress distributions and flowfield profiles. Included, for selected cases, are example CFD grids and setup information, along with surface pressure and wall heating results from simulations using current NASA real-gas Navier-Stokes codes by which future CFD investigators can compare and evaluate physics modeling improvements and validation and uncertainty assessments of future CFD code developments. The experimental database is presented tabulated in the Appendices describing each experiment. The database is also provided in computer-readable ASCII files located on a companion DVD.
A supercritical airfoil experiment
NASA Technical Reports Server (NTRS)
Mateer, G. G.; Seegmiller, H. L.; Hand, L. A.; Szodruck, J.
1994-01-01
The purpose of this investigation is to provide a comprehensive data base for the validation of numerical simulations. The objective of the present paper is to provide a tabulation of the experimental data. The data were obtained in the two-dimensional, transonic flowfield surrounding a supercritical airfoil. A variety of flows were studied in which the boundary layer at the trailing edge of the model was either attached or separated. Unsteady flows were avoided by controlling the Mach number and angle of attack. Surface pressures were measured on both the model and wind tunnel walls, and the flowfield surrounding the model was documented using a laser Doppler velocimeter (LDV). Although wall interference could not be completely eliminated, its effect was minimized by employing the following techniques. Sidewall boundary layers were reduced by aspiration, and upper and lower walls were contoured to accommodate the flow around the model and the boundary-layer growth on the tunnel walls. A data base with minimal interference from a tunnel with solid walls provides an ideal basis for evaluating the development of codes for the transonic speed range because the codes can include the wall boundary conditions more precisely than interference connections can be made to the data sets.
41 CFR 101-26.509 - Tabulating machine cards.
Code of Federal Regulations, 2010 CFR
2010-07-01
... 41 Public Contracts and Property Management 2 2010-07-01 2010-07-01 true Tabulating machine cards. 101-26.509 Section 101-26.509 Public Contracts and Property Management Federal Property Management Regulations System FEDERAL PROPERTY MANAGEMENT REGULATIONS SUPPLY AND PROCUREMENT 26-PROCUREMENT SOURCES AND...
Diamagnetic Corrections and Pascal's Constants
ERIC Educational Resources Information Center
Bain, Gordon A.; Berry, John F.
2008-01-01
Measured magnetic susceptibilities of paramagnetic substances must typically be corrected for their underlying diamagnetism. This correction is often accomplished by using tabulated values for the diamagnetism of atoms, ions, or whole molecules. These tabulated values can be problematic since many sources contain incomplete and conflicting data.…
Evidence of photosymbiosis in Palaeozoic tabulate corals.
Zapalski, Mikolaj K
2014-01-22
Coral reefs form the most diverse of all marine ecosystems on the Earth. Corals are among their main components and owe their bioconstructing abilities to a symbiosis with algae (Symbiodinium). The coral-algae symbiosis had been traced back to the Triassic (ca 240 Ma). Modern reef-building corals (Scleractinia) appeared after the Permian-Triassic crisis; in the Palaeozoic, some of the main reef constructors were extinct tabulate corals. The calcium carbonate secreted by extant photosymbiotic corals bears characteristic isotope (C and O) signatures. The analysis of tabulate corals belonging to four orders (Favositida, Heliolitida, Syringoporida and Auloporida) from Silurian to Permian strata of Europe and Africa shows these characteristic carbon and oxygen stable isotope signatures. The δ(18)O to δ(13)C ratios in recent photosymbiotic scleractinians are very similar to those of Palaeozoic tabulates, thus providing strong evidence of such symbioses as early as the Middle Silurian (ca 430 Ma). Corals in Palaeozoic reefs used the same cellular mechanisms for carbonate secretion as recent reefs, and thus contributed to reef formation.
Rainfall-runoff data from small watersheds in Colorado, October 1974 through September 1977
Cochran, Betty J.; Hodges, H.E.; Livingston, R.K.; Jarret, R.D.
1979-01-01
Rainfall-runoff data from small watersheds in Colorado are being collected and analyzed for the purpose of defining the flood characteristics of these and other similar areas. Data collected from October 1974 through September 1977 at a total of 18 urban stations, 10 Denver Federal Center stations, and 48 rural (or highway) stations are tabulated at 5-minute time intervals. Additional information presented includes station descriptions and methods of data collection and analysis. (Kosco-USGS)
The current status of the U.S. MTBE industry
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rose, G.M.
1995-12-31
This paper reviews the status of the MTBE industry from its beginnings as a result of the Clean Air Act Amendments and the need for the use of oxygenates in non-attainment areas. During 1990--93 three world scale merchant plants were constructed and in 1994 two more were brought on stream. The paper tabulates reasons why MTBE gained the lion`s share of the oxygenates market. Finally the paper discusses the problems that now plague the industry and their causes.
NittanySat Final Report for University Nanosatellite-5 Program
2009-10-12
Figures 9 through 12 and tabulated in Table 2. Figure 9 – 14-MHz BPF . Figure 10 – 21-MHz BPF . Page 13 Figure 11 – 28-MHz BPF ...Figure 12 – 50-MHz BPF . Table 2 - Narrow Band-pass Filter Parameters Frequency Band [MHz] Bandwidth Range [MHz] Insertion Loss [dB] Return Loss...surface area, and surface properties (e.g., absorptivity, emissivity) of the various components. In order to make predictions and guide design choices, an
NASA Technical Reports Server (NTRS)
1974-01-01
An analysis of low cost management approaches for the development of the Earth Observatory Satellite (EOS) is presented. The factors of the program which tend to increase costs are identified. The NASA/Industry interface is stressed to show how the interface can be improved to produce reduced program costs. Techniques and examples of cost reduction which can be applied to the EOS program are tabulated. Specific recommendations for actions to be taken to reduce costs in prescribed areas are submitted.
Statewide analysis of the drainage-area ratio method for 34 streamflow percentile ranges in Texas
Asquith, William H.; Roussel, Meghan C.; Vrabel, Joseph
2006-01-01
The drainage-area ratio method commonly is used to estimate streamflow for sites where no streamflow data are available using data from one or more nearby streamflow-gaging stations. The method is intuitive and straightforward to implement and is in widespread use by analysts and managers of surface-water resources. The method equates the ratio of streamflow at two stream locations to the ratio of the respective drainage areas. In practice, unity often is assumed as the exponent on the drainage-area ratio, and unity also is assumed as a multiplicative bias correction. These two assumptions are evaluated in this investigation through statewide analysis of daily mean streamflow in Texas. The investigation was made by the U.S. Geological Survey in cooperation with the Texas Commission on Environmental Quality. More than 7.8 million values of daily mean streamflow for 712 U.S. Geological Survey streamflow-gaging stations in Texas were analyzed. To account for the influence of streamflow probability on the drainage-area ratio method, 34 percentile ranges were considered. The 34 ranges are the 4 quartiles (0-25, 25-50, 50-75, and 75-100 percent), the 5 intervals of the lower tail of the streamflow distribution (0-1, 1-2, 2-3, 3-4, and 4-5 percent), the 20 quintiles of the 4 quartiles (0-5, 5-10, 10-15, 15-20, 20-25, 25-30, 30-35, 35-40, 40-45, 45-50, 50-55, 55-60, 60-65, 65-70, 70-75, 75-80, 80-85, 85-90, 90-95, and 95-100 percent), and the 5 intervals of the upper tail of the streamflow distribution (95-96, 96-97, 97-98, 98-99 and 99-100 percent). For each of the 253,116 (712X711/2) unique pairings of stations and for each of the 34 percentile ranges, the concurrent daily mean streamflow values available for the two stations provided for station-pair application of the drainage-area ratio method. For each station pair, specific statistical summarization (median, mean, and standard deviation) of both the exponent and bias-correction components of the drainage-area ratio method were computed. Statewide statistics (median, mean, and standard deviation) of the station-pair specific statistics subsequently were computed and are tabulated herein. A separate analysis considered conditioning station pairs to those stations within 100 miles of each other and with the absolute value of the logarithm (base-10) of the ratio of the drainage areas greater than or equal to 0.25. Statewide statistics of the conditional station-pair specific statistics were computed and are tabulated. The conditional analysis is preferable because of the anticipation that small separation distances reflect similar hydrologic conditions and the observation of large variation in exponent estimates for similar-sized drainage areas. The conditional analysis determined that the exponent is about 0.89 for streamflow percentiles from 0 to about 50 percent, is about 0.92 for percentiles from about 50 to about 65 percent, and is about 0.93 for percentiles from about 65 to about 85 percent. The exponent decreases rapidly to about 0.70 for percentiles nearing 100 percent. The computation of the bias-correction factor is sensitive to the range analysis interval (range of streamflow percentile); however, evidence suggests that in practice the drainage-area method can be considered unbiased. Finally, for general application, suggested values of the exponent are tabulated for 54 percentiles of daily mean streamflow in Texas; when these values are used, the bias correction is unity.
NASA Technical Reports Server (NTRS)
Stallcop, James R.; Partridge, Harry; Levin, Eugene
1991-01-01
N2(+) and O2(+) potential energy curves have been constructed by combining measured data with the results from electronic structure calculations. These potential curves have been employed to determine accurate charge exchange cross sections, transport cross sections, and collision integrals for ground state N(+)-N and O(+)-O interactions. The cross sections have been calculated from a semiclassical approximation to the scattering using a computer code that fits a spline curve through the discrete potential data and incorporates the proper long-range behavior of the interactions forces. The collision integrals are tabulated for a broad range of temperatures 250-100,000 K and are intended to reduce the uncertainty in the values of the transport properties of nonequilibrium air, particularly at high temperatures.
GRACKLE: a chemistry and cooling library for astrophysics
NASA Astrophysics Data System (ADS)
Smith, Britton D.; Bryan, Greg L.; Glover, Simon C. O.; Goldbaum, Nathan J.; Turk, Matthew J.; Regan, John; Wise, John H.; Schive, Hsi-Yu; Abel, Tom; Emerick, Andrew; O'Shea, Brian W.; Anninos, Peter; Hummels, Cameron B.; Khochfar, Sadegh
2017-04-01
We present the GRACKLE chemistry and cooling library for astrophysical simulations and models. GRACKLE provides a treatment of non-equilibrium primordial chemistry and cooling for H, D and He species, including H2 formation on dust grains; tabulated primordial and metal cooling; multiple ultraviolet background models; and support for radiation transfer and arbitrary heat sources. The library has an easily implementable interface for simulation codes written in C, C++ and FORTRAN as well as a PYTHON interface with added convenience functions for semi-analytical models. As an open-source project, GRACKLE provides a community resource for accessing and disseminating astrochemical data and numerical methods. We present the full details of the core functionality, the simulation and PYTHON interfaces, testing infrastructure, performance and range of applicability. GRACKLE is a fully open-source project and new contributions are welcome.
Conalogue, David Mc; Kinn, Sue; Mulligan, Jo-Ann; McNeil, Malcolm
2017-03-21
In recognition of the need for long-term planning for global health research, and to inform future global health research priorities, the United Kingdom Department for International Development (DfID) carried out a public consultation between May and June 2015. The consultation aimed to elicit views on the (1) the long-term future global health research priorities; (2) areas likely to be less important over time; (3) how to improve research uptake in low-income countries; and (4) how to build research capacity in low-income countries. An online consultation was used to survey a wide range of participants on global health research priorities. The qualitative data was analysed using a thematic analysis, with frequency of codes in responses tabulated to approximate relative importance of themes and sub-themes. The public consultation yielded 421 responses. The survey responses confirmed the growing importance of non-communicable disease as a global health research priority, being placed above infectious diseases. Participants felt that the key area for reducing funding prioritisation was infectious diseases. The involvement of policymakers and other key stakeholders was seen as critical to drive research uptake, as was collaboration and partnership. Several methods to build research capacity in low-income countries were described, including capacity building educational programmes, mentorship programmes and research institution collaboration and partnership. The outcomes from this consultation survey provide valuable insights into how DfID stakeholders prioritise research. The outcomes from this survey were reviewed alongside other elements of a wider DfID consultation process to help inform long-term research prioritisation of global health research. There are limitations in this approach; the opportunistic nature of the survey's dissemination means the findings presented may not be representative of the full range of stakeholders or views.
Modified COMS Plaques for {sup 125}I and {sup 103}Pd Iris Melanoma Brachytherapy
DOE Office of Scientific and Technical Information (OSTI.GOV)
Thomson, Rowan M., E-mail: rthomson@physics.carleton.c; Furutani, Keith M.; Pulido, Jose S.
2010-11-15
Purpose: Novel plaques are used to treat iris melanoma at the Mayo Clinic Rochester. The plaques are a modification of the Collaborative Ocular Melanoma Study (COMS) 22 mm plaque design with a gold alloy backing, outer lip, and silicone polymer insert. An inner lip surrounds a 10 mm diameter cutout region at the plaque center. Plaques span 360{sup o}, 270{sup o}, and 180{sup o} arcs. This article describes dosimetry for these plaques and others used in the treatment of anterior eye melanomas. Methods and Materials: The EGSnrc user-code BrachyDose is used to perform Monte Carlo simulations. Plaques and seeds aremore » fully modeled. Three-dimensional dose distributions for different plaque models, TG-43 calculations, and {sup 125}I (model 6711) and {sup 103}Pd (model 200) seeds are compared via depth-dose curves, tabulation of doses at points of interest, and isodose contours. Results: Doses at points of interest differ by up to 70% from TG-43 calculations. The inner lip reduces corneal doses. Matching plaque arc length to tumor extent reduces doses to eye regions outside the treatment area. Maintaining the same prescription dose, {sup 103}Pd offers lower doses to critical structures than {sup 125}I, with the exception of the sclera adjacent to the plaque. Conclusion: The Mayo Clinic plaques offer several advantages for anterior eye tumor treatments. Doses to regions outside the treatment area are significantly reduced. Doses differ considerably from TG-43 predictions, illustrating the importance of complete Monte Carlo simulations. Calculations take a few minutes on a single CPU, making BrachyDose sufficiently fast for routine clinical treatment planning.« less
NASA Technical Reports Server (NTRS)
Morain, S. A. (Principal Investigator); Williams, D. L.
1974-01-01
The author has identified the following significant results. Wheat area, yield, and production statistics as derived from satellite image analysis, combined with a weather model, are presented for a ten county area in southwest Kansas. The data (representing the 1972-73 crop year) are compared for accuracy against both the USDA August estimate and its final (official) tabulation. The area estimates from imagery for both dryland and irrigated winter wheat were within 5% of the official figures for the same area, and predated them by almost one year. Yield on dryland wheat was estimated by the Thompson weather model to within 0.1% of the observed yield. A combined irrigated and dryland wheat production estimate for the ten county area was completed in July, 1973 and was within 1% of the production reported by USDA in February, 1974.
Monitoring changes in landscape pattern: use of Ikonos and Quickbird images.
Alphan, Hakan; Çelik, Nil
2016-02-01
This paper aimed to analyze short-term changes in landscape pattern that primarily results from building development in the east coast of Mersin Province (Turkey). Three sites were selected. Ikonos (2003) and Quickbird (2009) images for these sites were classified, and land cover transformations were quantitatively analyzed using cross-tabulation of classification results. Changes in landscape structure were assessed by comparing the calculated values of area/edge and shape metrics for the earlier and later dates. Area/edge metrics included percentage of land and edge density, while shape metrics included perimeter-area ratio, fractal dimension, and related circumscribing circle (RCC) metrics. Orchards and buildings were dominating land cover classes. Variations in patch edge, size, and shapes were also analyzed and discussed. Degradation of prime agricultural areas due to building development and implications of such development on habitat fragmentation were highlighted.
Interaction geometry: an ecological perspective.
Rolfe A. Leary
1976-01-01
A new mathematical coordinate system results from a unique combination of two frameworks long used by ecologists: the phase plane and coaction cross-tabulation. The resulting construct combines the classifying power of the cross-tabulation and discriminating power of the phase plane. It may be used for analysis and synthesis.
Tabulated Neutron Emission Rates for Plutonium Oxide
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shores, Erik Frederick
This work tabulates neutron emission rates for 80 plutonium oxide samples as reported in the literature. Plutonium-238 and plutonium-239 oxides are included and such emission rates are useful for scaling tallies from Monte Carlo simulations and estimating dose rates for health physics applications.
An a priori study of different tabulation methods for turbulent pulverised coal combustion
NASA Astrophysics Data System (ADS)
Luo, Yujuan; Wen, Xu; Wang, Haiou; Luo, Kun; Jin, Hanhui; Fan, Jianren
2018-05-01
In many practical pulverised coal combustion systems, different oxidiser streams exist, e.g. the primary- and secondary-air streams in the power plant boilers, which makes the modelling of these systems challenging. In this work, three tabulation methods for modelling pulverised coal combustion are evaluated through an a priori study. Pulverised coal flames stabilised in a three-dimensional turbulent counterflow, consisting of different oxidiser streams, are simulated with detailed chemistry first. Then, the thermo-chemical quantities calculated with different tabulation methods are compared to those from detailed chemistry solutions. The comparison shows that the conventional two-stream flamelet model with a fixed oxidiser temperature cannot predict the flame temperature correctly. The conventional two-stream flamelet model is then modified to set the oxidiser temperature equal to the fuel temperature, both of which are varied in the flamelets. By this means, the variations of oxidiser temperature can be considered. It is found that this modified tabulation method performs very well on prediction of the flame temperature. The third tabulation method is an extended three-stream flamelet model that was initially proposed for gaseous combustion. The results show that the reference gaseous temperature profile can be overall reproduced by the extended three-stream flamelet model. Interestingly, it is found that the predictions of major species mass fractions are not sensitive to the oxidiser temperature boundary conditions for the flamelet equations in the a priori analyses.
NASA Astrophysics Data System (ADS)
Lin, Na
Pneumonia and asthma, two common Ambulatory Care Sensitive Conditions (ACSCs), were two top reasons for the admission of children to the hospitals and emergency rooms in the United States in 2011. Pneumonia and asthma are potentially preventable if the child's environment is properly managed. Underlying vulnerabilities such as low socioeconomic status (SES) and proximity to air pollution play an important role in ACSCs hospitalization. Pneumonia and asthma are two common reasons for hospitalizations among children and missed school days in Texas Coastal Bend Area. This thesis examines the relationships between neighborhood socioeconomic characteristics, meteorological conditions and children ACSCs hospitalization, including pneumonia and asthma among children age 0-17 in this area. Hospital discharge data from 2007 to 2009 based on Zip Code Tabulation Area (ZCTA) were examined along with American Community Survey (ACS) data, air pollution data from Environmental Protection Agency (EPA) and temperature data from National Climatic Data Center (NCDC). Hotspot and Local Moran's I analyses were applied to identify the concentrations of the illnesses. Two regressions (OLS and GWR) were applied to identify factors that contribute the most to ACSCs hospitalization. Pearson's correlation was calculated to examine the relationship between meteorological condition and child hospitalization for asthma and pneumonia. A human subject survey was conducted to examine the relationships between neighborhood environment and children asthma cases. The main finding was that children from families with health insurance, children from single father families and children from poor families were more likely to visit hospital for ACSCs and pneumonia care. "Hispanic families" and especially "Hispanic families with father but no mother" also contributed most to child hospitalization for ACSCs and pneumonia, suggesting that family preventative health care education is needed for Hispanic families and particularly Hispanic fathers. Air pollution and temperature analysis revealed that high concentrations of Ozone and Sulfur Dioxide likely cause pneumonia and asthma hospitalization of children. Combined with dramatic change in temperature, air pollution played an important role in the hospitalization of pneumonia and asthma plagued children in the coastal bend area. The human subject survey showed that the time children spent outdoors was highly positive correlated with asthma rate.
NASA Astrophysics Data System (ADS)
Bradshaw, S. J.
2009-07-01
Context: The effects of non-equilibrium processes on the ionisation state of strongly emitting elements in the solar corona can be extremely difficult to assess and yet they are critically important. For example, there is much interest in dynamic heating events localised in the solar corona because they are believed to be responsible for its high temperature and yet recent work has shown that the hottest (≥107 K) emission predicted to be associated with these events can be observationally elusive due to the difficulty of creating the highly ionised states from which the expected emission arises. This leads to the possibility of observing instruments missing such heating events entirely. Aims: The equations describing the evolution of the ionisaton state are a very stiff system of coupled, partial differential equations whose solution can be numerically challenging and time-consuming. Without access to specialised codes and significant computational resources it is extremely difficult to avoid the assumption of an equilibrium ionisation state even when it clearly cannot be justified. The aim of the current work is to develop a computational tool to allow straightforward calculation of the time-dependent ionisation state for a wide variety of physical circumstances. Methods: A numerical model comprising the system of time-dependent ionisation equations for a particular element and tabulated values of plasma temperature as a function of time is developed. The tabulated values can be the solutions of an analytical model, the output from a numerical code or a set of observational measurements. An efficient numerical method to solve the ionisation equations is implemented. Results: A suite of tests is designed and run to demonstrate that the code provides reliable and accurate solutions for a number of scenarios including equilibration of the ion population and rapid heating followed by thermal conductive cooling. It is found that the solver can evolve the ionisation state to recover exactly the equilibrium state found by an independent, steady-state solver for all temperatures, resolve the extremely small ionisation/recombination timescales associated with rapid temperature changes at high densities, and provide stable and accurate solutions for both dominant and minor ion population fractions. Rapid heating and cooling of low to moderate density plasma is characterised by significant non-equilibrium ionisation conditions. The effective ionisation temperatures are significantly lower than the electron temperature and the values found are in close agreement with the previous work of others. At the very highest densities included in the present study an assumption of equilibrium ionisation is found to be robust. Conclusions: The computational tool presented here provides a straightforward and reliable way to calculate ionisation states for a wide variety of physical circumstances. The numerical code gives results that are accurate and consistent with previous studies, has relatively undemanding computational requirements and is freely available from the author.
Study of short haul high-density V/STOL transportation systems. Volume 2: Appendices
NASA Technical Reports Server (NTRS)
Solomon, H. L.
1972-01-01
Essential supporting data to the short haul transportation study are presented. The specific appendices are arena characteristics, aerospace transportation analysis computer program, economics, model calibration, STOLport siting and services path selection, STOL schedule definition, tabulated California corridor results, and tabulated Midwest arena results.
Floods in the Wapsipinicon River Basin, Iowa
Schwob, Harlan H.
1971-01-01
Flood-profile sheets show profiles of actual flood occurrences and computed profiles of the 25- and 50-year floods at most locations. These sheets also contain tabulations of the flood discharges profiled. A low-water profile and tabulated discharge indicate the range in elevation and discharge along the streams.
NASA Technical Reports Server (NTRS)
Nelson, D. P.
1981-01-01
Tabulated aerodynamic data from coannular nozzle performance tests are given for test runs 26 through 37. The data include nozzle thrust coefficient parameters, nozzle discharge coefficients, and static pressure tap measurements.
BLS Machine-Readable Data and Tabulating Routines.
ERIC Educational Resources Information Center
DiFillipo, Tony
This report describes the machine-readable data and tabulating routines that the Bureau of Labor Statistics (BLS) is prepared to distribute. An introduction discusses the LABSTAT (Labor Statistics) database and the BLS policy on release of unpublished data. Descriptions summarizing data stored in 25 files follow this format: overview, data…
Research Trends with Cross Tabulation Search Engine
ERIC Educational Resources Information Center
Yin, Chengjiu; Hirokawa, Sachio; Yau, Jane Yin-Kim; Hashimoto, Kiyota; Tabata, Yoshiyuki; Nakatoh, Tetsuya
2013-01-01
To help researchers in building a knowledge foundation of their research fields which could be a time-consuming process, the authors have developed a Cross Tabulation Search Engine (CTSE). Its purpose is to assist researchers in 1) conducting research surveys, 2) efficiently and effectively retrieving information (such as important researchers,…
The influence of community and individual health literacy on self-reported health status.
Sentell, Tetine; Zhang, Wei; Davis, James; Baker, Kathleen Kromer; Braun, Kathryn L
2014-02-01
Individual health literacy is an established predictor of individual health outcomes. Community-level health literacy may also impact individual health, yet limited research has simultaneously considered the influence of individual and community health literacy on individual health. The study goal was to determine if community health literacy had an independent relationship with individual self-reported health beyond individual health literacy. We used data from the 2008 and 2010 Hawai'i Health Survey, a representative statewide telephone survey. Multilevel models predicted individual self-reported health by both individual and community health literacy, controlling for relevant individual-level (education, race/ethnicity, gender, poverty, insurance status, age, and marital status) and community-level variables (community poverty and community education). The sample included 11,779 individuals within 37 communities. Individual health literacy was defined by validated self-reported measurement. Communities were defined by zip code combinations. Community health literacy was defined as the percentage of individuals within a community reporting low health literacy. Census data by ZIP Code Tabulation Areas provided community-level variables. In descriptive results, 18.2 % self-reported low health literacy, and 14.7 % reported self-reported poor health. Community-level low health literacy ranged from 5.37 % to 35.99 %. In final, multilevel models, both individual (OR: 2.00; 95 % CI: 1.63-2.44) and community low health literacy (OR: 1.02; 95 % CI: 1.00-1.03) were significantly positively associated with self-reported poor health status. Each percentage increase of average low health literacy within a community was associated with an approximately 2 % increase in poor self-reported health for individuals in that community. Also associated with poorer health were lower educational attainment, older age, poverty, and non-White race. Both individual and community health literacy are significant, distinct correlates of individual general health status. Primary care providers and facilities should consider and address health literacy at both community and individual levels.
Examining the Association Between Temperature and Mental Health in California
NASA Astrophysics Data System (ADS)
Basu, R.; Gavin, L.; Pearson, D.; Malig, B. J.; Ebisu, K.
2016-12-01
Background: The association between temperature and morbidity from some specific causes has been well established. However, the association between temperature and mental health effects has not been examined closely, although those with mental illnesses may be susceptible to temperature. Methods: We obtained daily counts of emergency room visits and hospitalizations (ICD-9 codes) from the California Office of Statewide Health Planning and Development from 16 California climate zones from 2005 - 2013. Mean apparent temperature was determined by combining monitored temperature and humidity data from the US EPA, California Irrigation Management Information System, and the National Oceanic Atmospheric Administration and weighting monitor values by distance to zip code tabulation areas (ZCTA) and ZCTA populations in the same climate zone as each monitor. We used a two-stage hierarchical model to analyze this data, adjusted by the following independent variables: mean daily apparent temperature, holiday, day of the week, and a natural spline smoothing function of time. The regression was performed for both warm (5/1 - 10/31) and cold (11/1 - 4/30) seasons. Results were stratified by race/ethnicity and age group. Results: We observed an association between same-day mean apparent temperature and mental health outcomes during the warm and cold seasons. We also observed associations between temperature and suicide/self-injury and homicide/assault injury. A 10°F increase in mean apparent temperature was associated with a 4.98% [95% confidence interval, 3.73-6.23], 5.82% [4.34-7.30], and 7.43% [6.75-8.12], increase in mental health events, suicide, and homicide events during the warm season, respectively. Similar results were observed during the cold season. Effect modification by race/ethnic and age groups was observed for some outcomes for both seasons. Conclusions: Increase in mean apparent temperature was found to have same-day associations with several mental health outcomes and external injuries both self-inflicted and inflicted by others.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bambang, P.; Hardjono, M.; Silalahi, L.
1996-08-01
Tarakan basin is one of the basins in East Kalimantan having a complicated geological condition. Tectonic repetition developed in this area constructed various stratigraphic traps. Sedimentary development in log data shows continuous regression in Meliat and Tabul Formations (Middle Meocene), Santul Formation (Late Miocene), Tarakan Formation (Pliocene) and Bunyu Formation (Pleistocene), Supported by seismic data, stratigraphic sequence in the basin is obvious, especially in Sembakung-Bangkudulis area. The sequence boundaries, mainly {open_quotes}lowstand{close_quotes} distribution as good prospective trap, can be mapped by applying tract systems and studying wavelet extract as seismic expression character of a reservoir. Subtle changes in pattern of stratigraphicmore » sequences can become a hint of sedimentary environment and its lithology content, supporting both exploration and exploitation planning.« less
The 1981 current research on aviation weather (bibliography)
NASA Technical Reports Server (NTRS)
Daniel, J.; Frost, W.
1982-01-01
Current and ongoing research programs related to various areas of aviation meteorology are presented. Literature searches of major abstract publications, were conducted. Research project managers of various government agencies involved in aviation meteorology research provided a list of current research project titles and managers, supporting organizations, performing organizations, the principal investigators, and the objectives. These are tabulated under the headings of advanced meteorological instruments, forecasting, icing, lightning and atmospheric electricity; fog, visibility, and ceilings; low level wind shear, storm hazards/severe storms, turbulence, winds, and ozone and other meteorological parameters. This information was reviewed and assembled into a bibliography providing a current readily useable source of information in the area of aviation meteorology.
Sweetkind, Donald S.; Bova, Shiera C.; Langenheim, V.E.; Shumaker, Lauren E.; Scheirer, Daniel S.
2013-01-01
Stratigraphic information from 391 oil and gas exploration wells from Cuyama Valley, California, and surrounding areas are herein compiled in digital form from reports that were released originally in paper form. The Cuyama Basin is located within the southeasternmost part of the Coast Ranges and north of the western Transverse Ranges, west of the San Andreas fault. Knowledge of the location and elevation of stratigraphic tops of formations throughout the basin is a first step toward understanding depositional trends and the structural evolution of the basin through time, and helps in understanding the slip history and partitioning of slip on San Andreas and related faults.
Ningaloo Reef: Shallow Marine Habitats Mapped Using a Hyperspectral Sensor
Kobryn, Halina T.; Wouters, Kristin; Beckley, Lynnath E.; Heege, Thomas
2013-01-01
Research, monitoring and management of large marine protected areas require detailed and up-to-date habitat maps. Ningaloo Marine Park (including the Muiron Islands) in north-western Australia (stretching across three degrees of latitude) was mapped to 20 m depth using HyMap airborne hyperspectral imagery (125 bands) at 3.5 m resolution across the 762 km2 of reef environment between the shoreline and reef slope. The imagery was corrected for atmospheric, air-water interface and water column influences to retrieve bottom reflectance and bathymetry using the physics-based Modular Inversion and Processing System. Using field-validated, image-derived spectra from a representative range of cover types, the classification combined a semi-automated, pixel-based approach with fuzzy logic and derivative techniques. Five thematic classification levels for benthic cover (with probability maps) were generated with varying degrees of detail, ranging from a basic one with three classes (biotic, abiotic and mixed) to the most detailed with 46 classes. The latter consisted of all abiotic and biotic seabed components and hard coral growth forms in dominant or mixed states. The overall accuracy of mapping for the most detailed maps was 70% for the highest classification level. Macro-algal communities formed most of the benthic cover, while hard and soft corals represented only about 7% of the mapped area (58.6 km2). Dense tabulate coral was the largest coral mosaic type (37% of all corals) and the rest of the corals were a mix of tabulate, digitate, massive and soft corals. Our results show that for this shallow, fringing reef environment situated in the arid tropics, hyperspectral remote sensing techniques can offer an efficient and cost-effective approach to mapping and monitoring reef habitats over large, remote and inaccessible areas. PMID:23922921
NASA Technical Reports Server (NTRS)
Hopkins, R. H.; Davis, J. R.; Rohatgi, A.; Campbell, R. B.; Blais, P. D.; Rai-Choudhury, P.; Stapleton, R. E.; Mollenkopf, H. C.; Mccormick, J. R.
1980-01-01
Two major topics are treated: methods to measure and evaluate impurity effects in silicon and comprehensive tabulations of data derived during the study. Discussions of deep level spectroscopy, detailed dark I-V measurements, recombination lifetime determination, scanned laser photo-response, conventional solar cell I-V techniques, and descriptions of silicon chemical analysis are presented and discussed. The tabulated data include lists of impurity segregation coefficients, ingot impurity analyses and estimated concentrations, typical deep level impurity spectra, photoconductive and open circuit decay lifetimes for individual metal-doped ingots, and a complete tabulation of the cell I-V characteristics of nearly 200 ingots.
Impact of the hard-coded parameters on the hydrologic fluxes of the land surface model Noah-MP
NASA Astrophysics Data System (ADS)
Cuntz, Matthias; Mai, Juliane; Samaniego, Luis; Clark, Martyn; Wulfmeyer, Volker; Attinger, Sabine; Thober, Stephan
2016-04-01
Land surface models incorporate a large number of processes, described by physical, chemical and empirical equations. The process descriptions contain a number of parameters that can be soil or plant type dependent and are typically read from tabulated input files. Land surface models may have, however, process descriptions that contain fixed, hard-coded numbers in the computer code, which are not identified as model parameters. Here we searched for hard-coded parameters in the computer code of the land surface model Noah with multiple process options (Noah-MP) to assess the importance of the fixed values on restricting the model's agility during parameter estimation. We found 139 hard-coded values in all Noah-MP process options, which are mostly spatially constant values. This is in addition to the 71 standard parameters of Noah-MP, which mostly get distributed spatially by given vegetation and soil input maps. We performed a Sobol' global sensitivity analysis of Noah-MP to variations of the standard and hard-coded parameters for a specific set of process options. 42 standard parameters and 75 hard-coded parameters were active with the chosen process options. The sensitivities of the hydrologic output fluxes latent heat and total runoff as well as their component fluxes were evaluated. These sensitivities were evaluated at twelve catchments of the Eastern United States with very different hydro-meteorological regimes. Noah-MP's hydrologic output fluxes are sensitive to two thirds of its standard parameters. The most sensitive parameter is, however, a hard-coded value in the formulation of soil surface resistance for evaporation, which proved to be oversensitive in other land surface models as well. Surface runoff is sensitive to almost all hard-coded parameters of the snow processes and the meteorological inputs. These parameter sensitivities diminish in total runoff. Assessing these parameters in model calibration would require detailed snow observations or the calculation of hydrologic signatures of the runoff data. Latent heat and total runoff exhibit very similar sensitivities towards standard and hard-coded parameters in Noah-MP because of their tight coupling via the water balance. It should therefore be comparable to calibrate Noah-MP either against latent heat observations or against river runoff data. Latent heat and total runoff are sensitive to both, plant and soil parameters. Calibrating only a parameter sub-set of only soil parameters, for example, thus limits the ability to derive realistic model parameters. It is thus recommended to include the most sensitive hard-coded model parameters that were exposed in this study when calibrating Noah-MP.
The changing food outlet distributions and local contextual factors in the United States.
Chen, Hsin-Jen; Wang, Youfa
2014-01-16
Little is known about the dynamics of the food outlet distributions associated with local contextual factors in the U.S. This study examines the changes in food stores/services at the 5-digit Zip Code Tabulated Area (ZCTA5) level in the U.S., and assesses contextual factors associated with the changes. Data from 27,878 ZCTA5s in the contiguous United States without an extreme change in the number of 6 types of food stores/services (supermarkets, small-size grocery stores, convenience stores, fresh/specialty food markets, carry-out restaurants, and full-service restaurants) were used. ZCTA5s' contextual factors were from the 2000 Census. Numbers of food stores/services were derived from the Census Business Pattern databases. Linear regression models assessed contextual factors' influences (racial/ethnic compositions, poverty rate, urbanization level, and foreign-born population%) on 1-year changes in food stores/services during 2000-2001, adjusted for population size, total business change, and census regions. Small-size grocery stores and fresh/specialty food markets increased more and convenience stores decreased more in Hispanic-predominant than other areas. Among supermarket-free places, new supermarkets were less likely to be introduced into black-predominant than white-predominant areas (odds ratio (OR) = 0.52, 95% CI = 0.30-0.92). However, among areas without the following type of store at baseline, supermarket (OR = 0.48 (0.33-0.70)), small-size grocery stores (OR = 1.32 (1.08-1.62)), and fresh/specialty food markets (OR = 0.70 (0.53-0.92)) were less likely to be introduced into areas of low foreign-born population than into areas of high foreign-born population. Higher poverty rate was associated with a greater decrease in supermarket, a less decrease in small-size grocery stores, and a less increase in carry-out restaurants (all p for trends <0.001). Urban areas experienced more increases in full-service and carry-out restaurants than suburban areas. Local area characteristics affect 1-year changes in food environment in the U.S. Hispanic population was associated with more food stores/services capable of supplying fresh food items. Black-predominant and poverty-afflicted areas had a greater decrease in supermarkets. Full-service and carry-out restaurants increased more in urban than suburban areas. Foreign-born population density was associated with introduction of grocery stores and fresh/specialty food markets into the areas.
The changing food outlet distributions and local contextual factors in the United States
2014-01-01
Background Little is known about the dynamics of the food outlet distributions associated with local contextual factors in the U.S. This study examines the changes in food stores/services at the 5-digit Zip Code Tabulated Area (ZCTA5) level in the U.S., and assesses contextual factors associated with the changes. Methods Data from 27,878 ZCTA5s in the contiguous United States without an extreme change in the number of 6 types of food stores/services (supermarkets, small-size grocery stores, convenience stores, fresh/specialty food markets, carry-out restaurants, and full-service restaurants) were used. ZCTA5s’ contextual factors were from the 2000 Census. Numbers of food stores/services were derived from the Census Business Pattern databases. Linear regression models assessed contextual factors’ influences (racial/ethnic compositions, poverty rate, urbanization level, and foreign-born population%) on 1-year changes in food stores/services during 2000–2001, adjusted for population size, total business change, and census regions. Results Small-size grocery stores and fresh/specialty food markets increased more and convenience stores decreased more in Hispanic-predominant than other areas. Among supermarket-free places, new supermarkets were less likely to be introduced into black-predominant than white-predominant areas (odds ratio (OR) = 0.52, 95% CI = 0.30-0.92). However, among areas without the following type of store at baseline, supermarket (OR = 0.48 (0.33-0.70)), small-size grocery stores (OR = 1.32 (1.08-1.62)), and fresh/specialty food markets (OR = 0.70 (0.53-0.92)) were less likely to be introduced into areas of low foreign-born population than into areas of high foreign-born population. Higher poverty rate was associated with a greater decrease in supermarket, a less decrease in small-size grocery stores, and a less increase in carry-out restaurants (all p for trends <0.001). Urban areas experienced more increases in full-service and carry-out restaurants than suburban areas. Conclusions Local area characteristics affect 1-year changes in food environment in the U.S. Hispanic population was associated with more food stores/services capable of supplying fresh food items. Black-predominant and poverty-afflicted areas had a greater decrease in supermarkets. Full-service and carry-out restaurants increased more in urban than suburban areas. Foreign-born population density was associated with introduction of grocery stores and fresh/specialty food markets into the areas. PMID:24433323
Zapalski, Mikołaj K.; Berkowski, Błażej; Wrzołek, Tomasz
2016-01-01
Famennian tabulate corals were very rare worldwide, and their biodiversity was relatively low. Here we report a unique tabulate fauna from the mid- and late Famennian of the western part of the Holy Cross Mountains (Kowala and Ostrówka), Poland. We describe eight species (four of them new, namely ?Michelinia vinni sp. nov., Thamnoptychia mistiaeni sp. nov., Syringopora kowalensis sp. nov. and Syringopora hilarowiczi sp. nov.); the whole fauna consists of ten species (two others described in previous papers). These corals form two assemblages—the lower, mid-Famennian with Thamnoptychia and the upper, late Famennian with representatives of genera ?Michelinia, Favosites, Syringopora and ?Yavorskia. The Famennian tabulates from Kowala represent the richest Famennian assemblage appearing after the F/F crisis (these faunas appear some 10 Ma after the extinction event). Corals described here most probably inhabited deeper water settings, near the limit between euphotic and disphotic zones or slightly above. At generic level, these faunas show similarities to other Devonian and Carboniferous faunas, which might suggest their ancestry to at least several Carboniferous lineages. Tabulate faunas described here represent new recruits (the basin of the Holy Cross mountains was not a refuge during the F/F crisis) and have no direct evolutionary linkage to Frasnian faunas from Kowala. The colonization of the seafloor took place in two separate steps: first was monospecific assemblage of Thamnoptychia, and later came the diversified Favosites-Syringopora-Michelinia fauna. PMID:27007689
NASA Astrophysics Data System (ADS)
Zhang, Hongda; Han, Chao; Ye, Taohong; Ren, Zhuyin
2016-03-01
A method of chemistry tabulation combined with presumed probability density function (PDF) is applied to simulate piloted premixed jet burner flames with high Karlovitz number using large eddy simulation. Thermo-chemistry states are tabulated by the combination of auto-ignition and extended auto-ignition model. To evaluate the predictive capability of the proposed tabulation method to represent the thermo-chemistry states under the condition of different fresh gases temperature, a-priori study is conducted by performing idealised transient one-dimensional premixed flame simulations. Presumed PDF is used to involve the interaction of turbulence and flame with beta PDF to model the reaction progress variable distribution. Two presumed PDF models, Dirichlet distribution and independent beta distribution, respectively, are applied for representing the interaction between two mixture fractions that are associated with three inlet streams. Comparisons of statistical results show that two presumed PDF models for the two mixture fractions are both capable of predicting temperature and major species profiles, however, they are shown to have a significant effect on the predictions for intermediate species. An analysis of the thermo-chemical state-space representation of the sub-grid scale (SGS) combustion model is performed by comparing correlations between the carbon monoxide mass fraction and temperature. The SGS combustion model based on the proposed chemistry tabulation can reasonably capture the peak value and change trend of intermediate species. Aspects regarding model extensions to adequately predict the peak location of intermediate species are discussed.
Zapalski, Mikołaj K; Berkowski, Błażej; Wrzołek, Tomasz
2016-01-01
Famennian tabulate corals were very rare worldwide, and their biodiversity was relatively low. Here we report a unique tabulate fauna from the mid- and late Famennian of the western part of the Holy Cross Mountains (Kowala and Ostrówka), Poland. We describe eight species (four of them new, namely ?Michelinia vinni sp. nov., Thamnoptychia mistiaeni sp. nov., Syringopora kowalensis sp. nov. and Syringopora hilarowiczi sp. nov.); the whole fauna consists of ten species (two others described in previous papers). These corals form two assemblages-the lower, mid-Famennian with Thamnoptychia and the upper, late Famennian with representatives of genera ?Michelinia, Favosites, Syringopora and ?Yavorskia. The Famennian tabulates from Kowala represent the richest Famennian assemblage appearing after the F/F crisis (these faunas appear some 10 Ma after the extinction event). Corals described here most probably inhabited deeper water settings, near the limit between euphotic and disphotic zones or slightly above. At generic level, these faunas show similarities to other Devonian and Carboniferous faunas, which might suggest their ancestry to at least several Carboniferous lineages. Tabulate faunas described here represent new recruits (the basin of the Holy Cross mountains was not a refuge during the F/F crisis) and have no direct evolutionary linkage to Frasnian faunas from Kowala. The colonization of the seafloor took place in two separate steps: first was monospecific assemblage of Thamnoptychia, and later came the diversified Favosites-Syringopora-Michelinia fauna.
NASA Technical Reports Server (NTRS)
Goldberg, Robert K.; Carney, Kelly S.; DuBois, Paul; Hoffarth, Canio; Rajan, Subramaniam; Blankenhorn, Gunther
2015-01-01
Several key capabilities have been identified by the aerospace community as lacking in the material/models for composite materials currently available within commercial transient dynamic finite element codes such as LS-DYNA. Some of the specific desired features that have been identified include the incorporation of both plasticity and damage within the material model, the capability of using the material model to analyze the response of both three-dimensional solid elements and two dimensional shell elements, and the ability to simulate the response of composites composed with a variety of composite architectures, including laminates, weaves and braids. In addition, a need has been expressed to have a material model that utilizes tabulated experimentally based input to define the evolution of plasticity and damage as opposed to utilizing discrete input parameters (such as modulus and strength) and analytical functions based on curve fitting. To begin to address these needs, an orthotropic macroscopic plasticity based model suitable for implementation within LS-DYNA has been developed. Specifically, the Tsai-Wu composite failure model has been generalized and extended to a strain-hardening based orthotropic plasticity model with a non-associative flow rule. The coefficients in the yield function are determined based on tabulated stress-strain curves in the various normal and shear directions, along with selected off-axis curves. Incorporating rate dependence into the yield function is achieved by using a series of tabluated input curves, each at a different constant strain rate. The non-associative flow-rule is used to compute the evolution of the effective plastic strain. Systematic procedures have been developed to determine the values of the various coefficients in the yield function and the flow rule based on the tabulated input data. An algorithm based on the radial return method has been developed to facilitate the numerical implementation of the material model. The presented paper will present in detail the development of the orthotropic plasticity model and the procedures used to obtain the required material parameters. Methods in which a combination of actual testing and selective numerical testing can be combined to yield the appropriate input data for the model will be described. A specific laminated polymer matrix composite will be examined to demonstrate the application of the model.
Flood of September 2008 in Northwestern Indiana
Fowler, Kathleen K.; Kim, Moon H.; Menke, Chad D.; Arvin, Donald V.
2010-01-01
During September 12-15, 2008, rainfall ranging from 2 to more than 11 inches fell on northwestern Indiana. The rainfall resulted in extensive flooding on many streams within the Lake Michigan and Kankakee River Basins during September 12-18, causing two deaths, evacuation of hundreds of residents, and millions of dollars of damage to residences, businesses, and infrastructure. In all, six counties in northwestern Indiana were declared Federal disaster areas. U.S. Geological Survey (USGS) streamgages at four locations recorded new record peak streamflows as a result of the heavy rainfall. Peak-gage-height data, peak-streamflow data, annual exceedance probabilities, and recurrence intervals are tabulated in this report for 10 USGS streamgages in northwestern Indiana. Recurrence intervals of flood-peak streamflows were estimated to be greater than 100 years at six streamgages. Because flooding was particularly severe in the communities of Munster, Dyer, Hammond, Highland, Gary, Lake Station, Hobart, Schererville, Merrillville, Michiana Shores, and Portage, high-water-park data collected after the flood were tabulated for those communities. Flood peak inundation maps and water-surface profiles for selected streams were made in a geographic information system by combining high-water-mark data with the highest resolution digital elevation model data available.
A flamelet model for supersonic non-premixed combustion with pressure variation
NASA Astrophysics Data System (ADS)
Zhao, Guo-Yan; Sun, Ming-Bo; Wu, Jin-Shui; Wang, Hong-Bo
2015-08-01
A modified flamelet model is proposed for studying supersonic combustion with pressure variation considering that pressure is far from homogenous in a supersonic combustor. In this model, the flamelet database are tabulated at a reference pressure, while quantities at other pressure are obtained using a sixth-order polynomial in pressure. Attributed to merit of the modified model which compute coefficients for the expansion only. And they brought less requirements for memory and table lookup time, expensive cost is avoided. The performance of modified model is much better than the approach of using a flamelet model-based method with tabulation at different pressure values. Two types of hydrogen fueled scramjet combustors were introduced to validate the modified flamelet model. It was observed that the temperature is sensitive to the choice of model in combustion area, which in return will significantly affect the pressure. It was found that the results of modified model were in good agreement with the experimental data compared with the isobaric flamelet model, especially for temperature, whose value is more accurately predicted. It is concluded that the modified flamelet model was more effective for cases with a wide range of pressure variation.
The generation of arbitrary order, non-classical, Gauss-type quadrature for transport applications
DOE Office of Scientific and Technical Information (OSTI.GOV)
Spence, Peter J., E-mail: peter.spence@awe.co.uk
A method is presented, based upon the Stieltjes method (1884), for the determination of non-classical Gauss-type quadrature rules, and the associated sets of abscissae and weights. The method is then used to generate a number of quadrature sets, to arbitrary order, which are primarily aimed at deterministic transport calculations. The quadrature rules and sets detailed include arbitrary order reproductions of those presented by Abu-Shumays in [4,8] (known as the QR sets, but labelled QRA here), in addition to a number of new rules and associated sets; these are generated in a similar way, and we label them the QRS quadraturemore » sets. The method presented here shifts the inherent difficulty (encountered by Abu-Shumays) associated with solving the non-linear moment equations, particular to the required quadrature rule, to one of the determination of non-classical weight functions and the subsequent calculation of various associated inner products. Once a quadrature rule has been written in a standard form, with an associated weight function having been identified, the calculation of the required inner products is achieved using specific variable transformations, in addition to the use of rapid, highly accurate quadrature suited to this purpose. The associated non-classical Gauss quadrature sets can then be determined, and this can be done to any order very rapidly. In this paper, instead of listing weights and abscissae for the different quadrature sets detailed (of which there are a number), the MATLAB code written to generate them is included as Appendix D. The accuracy and efficacy (in a transport setting) of the quadrature sets presented is not tested in this paper (although the accuracy of the QRA quadrature sets has been studied in [12,13]), but comparisons to tabulated results listed in [8] are made. When comparisons are made with one of the azimuthal QRA sets detailed in [8], the inherent difficulty in the method of generation, used there, becomes apparent, with the highest order tabulated sets showing unexpected anomalies. Although not in an actual transport setting, the accuracy of the sets presented here is assessed to some extent, by using them to approximate integrals (over an octant of the unit sphere) of various high order spherical harmonics. When this is done, errors in the tabulated QRA sets present themselves at the highest tabulated orders, whilst combinations of the new QRS quadrature sets offer some improvements in accuracy over the original QRA sets. Finally, in order to offer a quick, visual understanding of the various quadrature sets presented, when combined to give product sets for the purposes of integrating functions confined to the surface of a sphere, three-dimensional representations of points located on an octant of the unit sphere (as in [8,12]) are shown.« less
NASA Technical Reports Server (NTRS)
Simmonds, A. L.; Miller, C. G., III; Nealy, J. E.
1976-01-01
Equilibrium thermodynamic properties for pure ammonia were generated for a range of temperature from 500 to 50,000 K and pressure from 0.01 to 40 MN/sq m and are presented in tabulated and graphical form. Properties include pressure, temperature, density, enthalpy, speed of sound, entropy, molecular-weight ratio, specific heat at constant pressure, specific heat at constant volume, isentropic exponent, and species mole fractions. These properties were calculated by the method which is based on minimization of the Gibbs free energy. The data presented herein are for an 18-species ammonia model. Heats of formation and spectroscopic constants used as input data are presented. Comparison of several thermodynamic properties calculated with the present program and a second computer code is performed for a range of pressure and for temperatures up to 30,000 K.
Polarizability tensor invariants of H2, HD, and D2
NASA Astrophysics Data System (ADS)
Raj, Ankit; Hamaguchi, Hiro-o.; Witek, Henryk A.
2018-03-01
We report an exhaustive compilation of wavelength-dependent matrix elements over the mean polarizability (α ¯ ) and polarizability anisotropy (γ) operators for the rovibrational states of the H2, HD, and D2 molecules together with an accompanying computer program for their evaluation. The matrix elements can be readily evaluated using the provided codes for rovibrational states with J = 0-15 and v = 0-4 and for any laser wavelengths in the interval 182.25-1320.6 nm corresponding to popular, commercially available lasers. The presented results substantially extend the scope of the data available in the literature, both in respect of the rovibrational transitions analyzed and the range of covered laser frequencies. The presented detailed tabulation of accurate polarizability tensor invariants is essential for successful realization of our main long-term goal: developing a universal standard for determining absolute Raman cross sections and absolute Raman intensities in experimental Rayleigh and Raman scattering studies of molecules.
M-DAS: System for multispectral data analysis. [in Saginaw Bay, Michigan
NASA Technical Reports Server (NTRS)
Johnson, R. H.
1975-01-01
M-DAS is a ground data processing system designed for analysis of multispectral data. M-DAS operates on multispectral data from LANDSAT, S-192, M2S and other sources in CCT form. Interactive training by operator-investigators using a variable cursor on a color display was used to derive optimum processing coefficients and data on cluster separability. An advanced multivariate normal-maximum likelihood processing algorithm was used to produce output in various formats: color-coded film images, geometrically corrected map overlays, moving displays of scene sections, coverage tabulations and categorized CCTs. The analysis procedure for M-DAS involves three phases: (1) screening and training, (2) analysis of training data to compute performance predictions and processing coefficients, and (3) processing of multichannel input data into categorized results. Typical M-DAS applications involve iteration between each of these phases. A series of photographs of the M-DAS display are used to illustrate M-DAS operation.
Preprocessor and postprocessor computer programs for a radial-flow finite-element model
Pucci, A.A.; Pope, D.A.
1987-01-01
Preprocessing and postprocessing computer programs that enhance the utility of the U.S. Geological Survey radial-flow model have been developed. The preprocessor program: (1) generates a triangular finite element mesh from minimal data input, (2) produces graphical displays and tabulations of data for the mesh , and (3) prepares an input data file to use with the radial-flow model. The postprocessor program is a version of the radial-flow model, which was modified to (1) produce graphical output for simulation and field results, (2) generate a statistic for comparing the simulation results with observed data, and (3) allow hydrologic properties to vary in the simulated region. Examples of the use of the processor programs for a hypothetical aquifer test are presented. Instructions for the data files, format instructions, and a listing of the preprocessor and postprocessor source codes are given in the appendixes. (Author 's abstract)
Capillary Pressure of a Liquid Between Uniform Spheres Arranged in a Square-Packed Layer
NASA Technical Reports Server (NTRS)
Alexader, J. Iwan D.; Slobozhanin, Lev A.; Collicott, Steven H.
2004-01-01
The capillary pressure in the pores defined by equidimensional close-packed spheres is analyzed numerically. In the absence of gravity the menisci shapes are constructed using Surface Evolver code. This permits calculation the free surface mean curvature and hence the capillary pressure. The dependences of capillary pressure on the liquid volume constructed here for a set of contact angles allow one to determine the evolution of basic capillary characteristics under quasi-static infiltration and drainage. The maximum pressure difference between liquid and gas required for a meniscus passing through a pore is calculated and compared with that for hexagonal packing and with approximate solution given by Mason and Morrow [l]. The lower and upper critical liquid volumes that determine the stability limits for the equilibrium capillary liquid in contact with square packed array of spheres are tabulated for a set of contact angles.
Testing Moderating Detection Systems with {sup 252}Cf-Based Reference Neutron Fields
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hertel, Nolan E.; Sweezy, Jeremy; Sauber, Jeremiah S.
Calibration measurements were carried out on a probe designed to measure ambient dose equivalent in accordance with ICRP Pub 60 recommendations. It consists of a cylindrical {sup 3}He proportional counter surrounded by a 25-cm-diameter spherical polyethylene moderator. Its neutron response is optimized for dose rate measurements of neutrons between thermal energies and 20 MeV. The instrument was used to measure the dose rate in four separate neutron fields: unmoderated {sup 252}Cf, D{sub 2}O-moderated {sup 252}Cf, polyethylene-moderated {sup 252}Cf, and WEP neutron howitzer with {sup 252}Cf at its center. Dose equivalent measurements were performed at source-detector centerline distances from 50 tomore » 200 cm. The ratio of air-scatter- and room-return-corrected ambient dose equivalent rates to ambient dose equivalent rates calculated with the code MCNP are tabulated.« less
Intersection crash prediction modeling with macro-level data from various geographic units.
Lee, Jaeyoung; Abdel-Aty, Mohamed; Cai, Qing
2017-05-01
There have been great efforts to develop traffic crash prediction models for various types of facilities. The crash models have played a key role to identify crash hotspots and evaluate safety countermeasures. In recent, many macro-level crash prediction models have been developed to incorporate highway safety considerations in the long-term transportation planning process. Although the numerous macro-level studies have found that a variety of demographic and socioeconomic zonal characteristics have substantial effects on traffic safety, few studies have attempted to coalesce micro-level with macro-level data from existing geographic units for estimating crash models. In this study, the authors have developed a series of intersection crash models for total, severe, pedestrian, and bicycle crashes with macro-level data for seven spatial units. The study revealed that the total, severe, and bicycle crash models with ZIP-code tabulation area data performs the best, and the pedestrian crash models with census tract-based data outperforms the competing models. Furthermore, it was uncovered that intersection crash models can be drastically improved by only including random-effects for macro-level entities. Besides, the intersection crash models are even further enhanced by including other macro-level variables. Lastly, the pedestrian and bicycle crash modeling results imply that several macro-level variables (e.g., population density, proportions of specific age group, commuters who walk, or commuters using bicycle, etc.) can be a good surrogate exposure for those crashes. Copyright © 2017 Elsevier Ltd. All rights reserved.
1987-10-01
Ranch Hand Enlisted Flyers ... .......... ... R-14 R-1O Cross Tabulation of Pleurisy - (Abnormal, Total) by- Exposure Index Category-by-Age Category-by...by-Pack-Year Category for Ranch Hand Enlisted Groundcrew ........... .. R-19 R-15 Cross Tabulation of Pleurisy - (Abnormal, Total) by- Exposure Index
Food Tabulator. DOT No. 211.582-010. Cafeteria Occupations. Coordinator's Guide. First Edition.
ERIC Educational Resources Information Center
East Texas State Univ., Commerce. Occupational Curriculum Lab.
This study guide, one of eight individualized units developed for students enrolled in cooperative part-time training and employed in a cafeteria, is composed of information about one specific occupation; this unit focuses on the duties of the food tabulator. Materials provided in this guide for coordinator use include a student progress chart; a…
Moon view period tabulations (with station masking) for Manned Space Flight Network stations, book 1
NASA Technical Reports Server (NTRS)
Gattie, M. M.; Williams, R. L.
1970-01-01
The times during which MSFN stations can view the moon are tabulated. Station view periods for each month are given. All times and dates refer to Greenwich Mean Time. AOS and LOS refer to the center of the moon at zero degrees elevation for moon rise and set, respectively.
Asres, Yihunie Hibstie; Mathuthu, Manny; Birhane, Marelgn Derso
2018-04-22
This study provides current evidence about cross-section production processes in the theoretical and experimental results of neutron induced reaction of uranium isotope on projectile energy range of 1-100 MeV in order to improve the reliability of nuclear stimulation. In such fission reactions of 235 U within nuclear reactors, much amount of energy would be released as a product that able to satisfy the needs of energy to the world wide without polluting processes as compared to other sources. The main objective of this work is to transform a related knowledge in the neutron-induced fission reactions on 235 U through describing, analyzing and interpreting the theoretical results of the cross sections obtained from computer code COMPLET by comparing with the experimental data obtained from EXFOR. The cross section value of 235 U(n,2n) 234 U, 235 U(n,3n) 233 U, 235 U(n,γ) 236 U, 235 U(n,f) are obtained using computer code COMPLET and the corresponding experimental values were browsed by EXFOR, IAEA. The theoretical results are compared with the experimental data taken from EXFOR Data Bank. Computer code COMPLET has been used for the analysis with the same set of input parameters and the graphs were plotted by the help of spreadsheet & Origin-8 software. The quantification of uncertainties stemming from both experimental data and computer code calculation plays a significant role in the final evaluated results. The calculated results for total cross sections were compared with the experimental data taken from EXFOR in the literature, and good agreement was found between the experimental and theoretical data. This comparison of the calculated data was analyzed and interpreted with tabulation and graphical descriptions, and the results were briefly discussed within the text of this research work. Copyright © 2018 The Authors. Published by Elsevier Ltd.. All rights reserved.
ENVIRONMENTAL RADIOACTIVITY AT ARGONNE NATIONAL LABORATORY. Report for the Year 1958
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sedlet, J.
1959-08-01
Data are tabulated on the radioactive content of samples of surface water, lake and stream bottom silt, soil, plants, and air filters from the environs of the Argonne National Laboratory. Results are compared with those for similar samples collected from the area from 1952 through 1958. Fission prcduct activity from nuclear detocations was found in most samples from all locations. Fall-out activity was greatest during the spring and fall, and was particularly noticeable in air, precipitation, and plant samples. (For preceding period see ANL-5934.) (C.H.)
NASA Technical Reports Server (NTRS)
Evans, S.; Lewis, H.; Williamsen, J.; Evans, H.; Bohl, W.; Parker, Nelson (Technical Monitor)
2002-01-01
Orbital debris impacts on the International Space Station occur frequently. To date, none of the impacting particles has been sufficiently large to penetrate manned pressurized volumes. We used the Manned Spacecraft Crew Survivability code to evaluate the risk to crew of penetrations of pressurized modules at two assembly stages: after Flight lJ, when the pressurized elements of Kibo, the Japanese Experiment Module, are present, and after Flight lE, when the European Columbus Module is present. Our code is a Monte Carlo simulation of impacts on the Station that considers several potential event types that could lead to crew loss. Among the statistics tabulated by the program is the probability of death of one or more crew members, expressed as the risk factor, R. This risk factor is dependent on details of crew operations during both ordinary circumstances and decompression emergencies, as well as on details of internal module configurations. We conducted trade studies considering these procedure and configuration details to determine the bounds on R at the 1J and 1E stages in the assembly sequence. Here we compare the R-factor bounds, and procedures and configurations that reduce R at these stages.
MAGI: many-component galaxy initializer
NASA Astrophysics Data System (ADS)
Miki, Yohei; Umemura, Masayuki
2018-04-01
Providing initial conditions is an essential procedure for numerical simulations of galaxies. The initial conditions for idealized individual galaxies in N-body simulations should resemble observed galaxies and be dynamically stable for time-scales much longer than their characteristic dynamical times. However, generating a galaxy model ab initio as a system in dynamical equilibrium is a difficult task, since a galaxy contains several components, including a bulge, disc, and halo. Moreover, it is desirable that the initial-condition generator be fast and easy to use. We have now developed an initial-condition generator for galactic N-body simulations that satisfies these requirements. The developed generator adopts a distribution-function-based method, and it supports various kinds of density models, including custom-tabulated inputs and the presence of more than one disc. We tested the dynamical stability of systems generated by our code, representing early- and late-type galaxies, with N = 2097 152 and 8388 608 particles, respectively, and we found that the model galaxies maintain their initial distributions for at least 1 Gyr. The execution times required to generate the two models were 8.5 and 221.7 seconds, respectively, which is negligible compared to typical execution times for N-body simulations. The code is provided as open-source software and is publicly and freely available at https://bitbucket.org/ymiki/magi.
Injury risks of EMS responders: evidence from the National Fire Fighter Near-Miss Reporting System
Taylor, Jennifer A; Davis, Andrea L; Barnes, Brittany; Lacovara, Alicia V; Patel, Reema
2015-01-01
Objectives We analysed near-miss and injury events reported to the National Fire Fighter Near-Miss Reporting System (NFFNMRS) to investigate the workplace hazards and safety concerns of Emergency Medical Services (EMS) responders in the USA. Methods We reviewed 769 ‘non-fire emergency event’ reports from the NFFNMRS using a mixed methods approach. We identified 185 emergency medical calls and analysed their narrative text fields. We assigned Mechanism of Near-Miss/Injury and Nature of Injury codes and then tabulated frequencies (quantitative). We coded major themes regarding work hazards and safety concerns reported by the EMS responders (qualitative). Results Of the 185 emergency medical calls, the most commonly identified Mechanisms of Near-Miss/Injury to EMS responders was Assaults, followed by Struck-by Motor Vehicle, and Motor Vehicle Collision. The most commonly identified weapon used in an assault was a firearm. We identified 5 major domains of workplace hazards and safety concerns: Assaults by Patients, Risks from Motor Vehicles, Personal Protective Equipment, Relationships between Emergency Responders, and Policies, Procedures and Practices. Conclusions Narrative text from the NFFNMRS is a rich source of data that can be analysed quantitatively and qualitatively to provide insight into near-misses and injuries sustained by EMS responders. Near-miss reporting systems are critical components for occupational hazard surveillance. PMID:26068510
Health claim evidence requirements in Japan.
Yamada, Kazuhiko; Sato-Mito, Natsuko; Nagata, Junichi; Umegaki, Keizo
2008-06-01
In the early 1980s the Japanese scientific academy defined a functional food as a food having a tertiary or physiologically active function. The current Japanese "Food with Health Claims" include 2 categories. For the first category, "Food with Nutrient Function Claims," the label may be freely used if a product satisfies the standard for the minimum and maximum levels per daily portion usually consumed. The second category is defined as "Food for Specified Health Uses" (FOSHU). FOSHU foods are those that contain dietary ingredients that have beneficial effects on the physiological functions of the human body, maintain and promote health, and improve health-related conditions. Health claims on these foods correspond to the category of "other" function claims of the Codex Alimentarius. However, claims of disease-risk reduction are not currently allowed under FOSHU with an exception for calcium and folic acid. Manufacturers can emphasize the characteristics of their products and promote sales by labeling or claims. Therefore, the labeling should be clear and correct and avoid any chance of misinterpretation. The labeling of health claims on foods should always be based on scientific evidence. Any manufacturer who applies to the government for approval under the FOSHU code for its product must tabulate both published available publications and internal reports on the effectiveness of the product and/or its ingredients and provide a summary of each available publication or report. The tabulation must include in vitro metabolic and biochemical studies, in vivo studies, and randomized controlled trials on Japanese people. The overall philosophy of the Ministry is to maintain and improve the health status of people and to prevent chronic noncommunicable diseases through an approach that involves a well-balanced diet as well as through the use of "health foods" including "Food with Health Claims."
NASA Technical Reports Server (NTRS)
Goldberg, Robert K.; Carney, Kelly S.; DuBois, Paul; Khaled, Bilal; Hoffarth, Canio; Rajan, Subramaniam; Blankenhorn, Gunther
2016-01-01
A material model which incorporates several key capabilities which have been identified by the aerospace community as lacking in state-of-the art composite impact models is under development. In particular, a next generation composite impact material model, jointly developed by the FAA and NASA, is being implemented into the commercial transient dynamic finite element code LS-DYNA. The material model, which incorporates plasticity, damage, and failure, utilizes experimentally based tabulated input to define the evolution of plasticity and damage and the initiation of failure as opposed to specifying discrete input parameters (such as modulus and strength). The plasticity portion of the orthotropic, three-dimensional, macroscopic composite constitutive model is based on an extension of the Tsai-Wu composite failure model into a generalized yield function with a non-associative flow rule. For the damage model, a strain equivalent formulation is utilized to allow for the uncoupling of the deformation and damage analyses. In the damage model, a semi-coupled approach is employed where the overall damage in a particular coordinate direction is assumed to be a multiplicative combination of the damage in that direction resulting from the applied loads in the various coordinate directions. Due to the fact that the plasticity and damage models are uncoupled, test procedures and methods to both characterize the damage model and to covert the material stress-strain curves from the true (damaged) stress space to the effective (undamaged) stress space have been developed. A methodology has been developed to input the experimentally determined composite failure surface in a tabulated manner. An analytical approach is then utilized to track how close the current stress state is to the failure surface.
Theoretical Performance of Liquid Hydrogen with Liquid Oxygen as a Rocket Propellant
NASA Technical Reports Server (NTRS)
Gordon, Sanford; McBride, Bonnie J.
1959-01-01
Theoretical rocket performance for both equilibrium and frozen composition during expansion was calculated for the propellant combination liquid hydrogen and liquid oxygen at four chamber pressures (60, 150, 300, and 600 lb/sq in. abs) and a wide range of pressure ratios (1 to 4000) and oxidant-fuel ratios (1.190 to 39.683). Data are given to estimate performance parameters at chamber pressures other than those for which data are tabulated. The parameters included are specific impulse, specific impulse in vacuum, combustion-chamber temperature, nozzle-exit temperature, molecular weight, molecular-weight derivatives, characteristic velocity, coefficient of thrust, ratio of nozzle-exit area to throat area, specific heat at constant pressure, isentropic exponent, viscosity, thermal conductivity, Mach number, and equilibrium gas compositions.
2010 Military Family Life Project (MFLP) - Couples: Tabulations of Responses
2013-08-31
interest income; dividends; child support/alimony; social security, welfare assistance; and net rent, trusts, and royalties from any other investments ...2010 Military Family Life Project: Couples Tabulations of Responses Additional copies of this report may be obtained from: Defense... RESPONSES Defense Manpower Data Center Human Resources Strategic Assessment Program 4800 Mark Center Drive, Suite 04E25-01, Alexandria, VA 22350
Network Prime-Time Violence Tabulations for 1975-76 Season.
ERIC Educational Resources Information Center
Klapper, Joseph T.
This is an annual report on violence in prime-time television. The tabulations, based on 13 weeks of monitoring prime-time programs on three networks, indicate a decline in violence by 24% and a decline in the rate per hour of dramatic violence to 1.9 incidents per hour since last season. The study also indicated that the introduction of the…
Twelve-year planetary ephemeris: 1995-2006
NASA Technical Reports Server (NTRS)
Espenak, Fred
1994-01-01
Accurate geocentric positions and physical ephemerides are tabulated for the Sun, Moon, Mercury, Venus, Mars, Jupiter, Saturn, Uranus, Neptune, and Pluto for the 12-year period 1995 through 2006. The frequency interval is 2 days for the Sun and classical planets. Uranus, Neptune, and Pluto are tabulated at 4-day intervals. Because of its rapid apparent motion, the Moon's ephemeris is given daily.
Selected yield tables for plantations and natural stands in Inland Northwest Forests
Albert R. Stage; David L. Renner; Roger C. Chapman
1988-01-01
Yields arrayed by site index and age have been tabulated for plantations of 500 trees per acre, with five thinning regimes, for Douglas-fir, grand fir, and western larch. Yields were also tabulated for naturally regenerated stands of the grand fir-cedar-hemlock ecosystem of the Inland Empire. All yields were estimated with the Prognosis Model for Stand Development,...
Tabulated dose uniformity ratio and minimum dose data: rectangular 60Co source plaques
DOE Office of Scientific and Technical Information (OSTI.GOV)
Galanter, L.
1971-01-01
The data tabulated herein extend to rectangular cobalt-60 plaques the information presented for square plaques in BNL 50145 (Revised). The user is referred to BNL 50145 (Revised) and to the other reports listed for a complete discussion of the parameters involved in data generation and for instructions on the use of these data in gamma irradiator design.
ERIC Educational Resources Information Center
National Center for Education Statistics (ED), Washington, DC.
This document reports preliminary tabulations of public elementary and secondary school revenues and current expenditures for Fiscal Year 1987 (School Year 1986-87). Data shows revenues by local, state, intermediate, and federal sources, and current expenditures by categories of instruction, support services, noninstructional services, and fixed…
Tables of X-ray absorption corrections and dispersion corrections: the new versus the old
NASA Astrophysics Data System (ADS)
Creagh, Dudley
1990-11-01
This paper compares the data on X-ray absorption coefficients calculated by Creagh and Hubbell and tabulated in International Tables for Crystallography, vol. C, ed. A.J.C. Wilson (1990) section 4.2.4 [1] with empirical (Saloman, Hubbell and Scofield, At. Data and Nucl. Data Tables 38 (1988) 1, [6]) and semi-empirical (Hubbell, McMaster, Kerr Del Grande and Mallett, in: International Tables for Crystallography, vol. IV, eds. Ibers and Hamilton (Kynoch, Birmingham, 1974) [2]) tabulations as well as the renormalized relativistic Dirac-Hartree-Fock calculations of Scofield [6]. It also makes comparisons of the real part of the dispersion correction ƒ‧(ω, 0) and tabulated in ref. [1], with theoretical data sets (Cromer and Liberman, J. Chem. Phys. 53 (1970) 1891, and Acta Crystallogr. A37 (1981) 267 [4,5]; Wang, Phys. Rev. A34 (1986) 636 [85]; Kissel, in: Workshop Report on New Dimensions in X-ray Scattering, CONF-870459 (Livermore, 1987) p. 9 [86]) and data collected using a variety of experimental techniques. In both cases the data tabulated in ref. [1] is shown to give improved self-consistency and agreement with experiment.
NASA Astrophysics Data System (ADS)
Manjanaik, N.; Parameshachari, B. D.; Hanumanthappa, S. N.; Banu, Reshma
2017-08-01
Intra prediction process of H.264 video coding standard used to code first frame i.e. Intra frame of video to obtain good coding efficiency compare to previous video coding standard series. More benefit of intra frame coding is to reduce spatial pixel redundancy with in current frame, reduces computational complexity and provides better rate distortion performance. To code Intra frame it use existing process Rate Distortion Optimization (RDO) method. This method increases computational complexity, increases in bit rate and reduces picture quality so it is difficult to implement in real time applications, so the many researcher has been developed fast mode decision algorithm for coding of intra frame. The previous work carried on Intra frame coding in H.264 standard using fast decision mode intra prediction algorithm based on different techniques was achieved increased in bit rate, degradation of picture quality(PSNR) for different quantization parameters. Many previous approaches of fast mode decision algorithms on intra frame coding achieved only reduction of computational complexity or it save encoding time and limitation was increase in bit rate with loss of quality of picture. In order to avoid increase in bit rate and loss of picture quality a better approach was developed. In this paper developed a better approach i.e. Gaussian pulse for Intra frame coding using diagonal down left intra prediction mode to achieve higher coding efficiency in terms of PSNR and bitrate. In proposed method Gaussian pulse is multiplied with each 4x4 frequency domain coefficients of 4x4 sub macro block of macro block of current frame before quantization process. Multiplication of Gaussian pulse for each 4x4 integer transformed coefficients at macro block levels scales the information of the coefficients in a reversible manner. The resulting signal would turn abstract. Frequency samples are abstract in a known and controllable manner without intermixing of coefficients, it avoids picture getting bad hit for higher values of quantization parameters. The proposed work was implemented using MATLAB and JM 18.6 reference software. The proposed work measure the performance parameters PSNR, bit rate and compression of intra frame of yuv video sequences in QCIF resolution under different values of quantization parameter with Gaussian value for diagonal down left intra prediction mode. The simulation results of proposed algorithm are tabulated and compared with previous algorithm i.e. Tian et al method. The proposed algorithm achieved reduced in bit rate averagely 30.98% and maintain consistent picture quality for QCIF sequences compared to previous algorithm i.e. Tian et al method.
Thomas, H.E.
1946-01-01
Ground-water investigations in Utah by the Geological Survey of the U.S. Department of the Interior have been in progress since 1935, in cooperation with the Utah State Engineer. This cooperative work includes (1) determination of the fluctuations of water level in most of the developed ground-water areas in the state, based upon measurements which are tabulated and published annually by the Geological Survey; and (2) detailed investigations of specific ground-water areas to determine source, movement, disposal, quantity and quality of the ground water, and to show the relation of present development to the maximum economic development of which those areas are capable. Such detailed investigations have been completed during the past decade for areas in Iron, Millard, Salt Lake, Tooele, and Weber Counties, and are referred to in discussion subsequently. Similar investigations are now in progress in other areas in Davis, Iron, and Weber Counties.
WASP (Write a Scientific Paper) using Excel - 2: Pivot tables.
Grech, Victor
2018-02-01
Data analysis at the descriptive stage and the eventual presentation of results requires the tabulation and summarisation of data. This exercise should always precede inferential statistics. Pivot tables and pivot charts are one of Excel's most powerful and underutilised features, with tabulation functions that immensely facilitate descriptive statistics. Pivot tables permit users to dynamically summarise and cross-tabulate data, create tables in several dimensions, offer a range of summary statistics and can be modified interactively with instant outputs. Large and detailed datasets are thereby easily manipulated making pivot tables arguably the best way to explore, summarise and present data from many different angles. This second paper in the WASP series in Early Human Development provides pointers for pivot table manipulation in Excel™. Copyright © 2018 Elsevier B.V. All rights reserved.
Expansion of Tabulated Scattering Matrices in Generalized Spherical Functions
NASA Technical Reports Server (NTRS)
Mishchenko, Michael I.; Geogdzhayev, Igor V.; Yang, Ping
2016-01-01
An efficient way to solve the vector radiative transfer equation for plane-parallel turbid media is to Fourier-decompose it in azimuth. This methodology is typically based on the analytical computation of the Fourier components of the phase matrix and is predicated on the knowledge of the coefficients appearing in the expansion of the normalized scattering matrix in generalized spherical functions. Quite often the expansion coefficients have to be determined from tabulated values of the scattering matrix obtained from measurements or calculated by solving the Maxwell equations. In such cases one needs an efficient and accurate computer procedure converting a tabulated scattering matrix into the corresponding set of expansion coefficients. This short communication summarizes the theoretical basis of this procedure and serves as the user guide to a simple public-domain FORTRAN program.
Code of Federal Regulations, 2010 CFR
2010-01-01
... localities pursuant to 13 U.S.C. 141(c). The determination of the Secretary will be published in the Federal... until after he or she receives the recommendation of the Director of the Census, together with the... Director of the Census analyzing the methodologies that may be used in making the tabulations of population...
ERIC Educational Resources Information Center
South Carolina Commission on Higher Education, Columbia.
This tabulation of academic degrees and other awards has been compiled by the South Carolina Commission on Higher Education primarily as an aid to institutions and to state-wide planning for postsecondary education. The data are tabulated from information furnished to the Commission and to the U. S. Office of Education on OE Form 2300 - 2.l…
Power law relationships for rain attenuation and reflectivity
NASA Technical Reports Server (NTRS)
Devasirvatham, D. M. J.; Hodge, D. B.
1978-01-01
The equivalent reflectivity, specific attenuation and volumetric backscatter cross section of rain are calculated and tabulated at a number of frequencies from 1 to 500 GHz using classical Mie theory. The first two parameters are shown to be closely approximated as functions of rain rate by the power law aR to the b power. The a's and b's are also tabulated and plotted for convenient reference.
Sprawl in European urban areas
NASA Astrophysics Data System (ADS)
Prastacos, Poulicos; Lagarias, Apostolos
2016-08-01
In this paper the 2006 edition of the Urban Atlas database is used to tabulate areas of low development density, usually referred to as "sprawl", for many European cities. The Urban Atlas database contains information on the land use distribution in the 305 largest European cities. Twenty different land use types are recognized, with six of them representing urban fabric. Urban fabric classes are residential areas differentiated by the density of development, which is measured by the sealing degree parameter that ranges from 0% to 100% (non-developed, fully developed). Analysis is performed on the distribution of the middle to low density areas defined as those with sealing degree less than 50%. Seven different country groups in which urban areas have similar sprawl characteristics are identified and some key characteristics of sprawl are discussed. Population of an urban area is another parameter considered in the analysis. Two spatial metrics, average patch size and mean distance to the nearest neighboring patch of the same class, are used to describe proximity/separation characteristics of sprawl in the urban areas of the seven groups.
Karpati, Adam M.; Perrin, Mary C.; Matte, Tom; Leighton, Jessica; Schwartz, Joel; Barr, R. Graham
2004-01-01
Pyrethroid pesticides were applied via ground spraying to residential neighborhoods in New York City during July–September 2000 to control mosquito vectors of West Nile virus (WNV). Case reports link pyrethroid exposure to asthma exacerbations, but population-level effects on asthma from large-scale mosquito control programs have not been assessed. We conducted this analysis to determine whether widespread urban pyrethroid pesticide use was associated with increased rates of emergency department (ED) visits for asthma. We recorded the dates and locations of pyrethroid spraying during the 2000 WNV season in New York City and tabulated all ED visits for asthma to public hospitals from October 1999 through November 2000 by date and ZIP code of patients’ residences. The association between pesticide application and asthma-related emergency visits was evaluated across date and ZIP code, adjusting for season, day of week, and daily temperature, precipitation, particulate, and ozone levels. There were 62,827 ED visits for asthma during the 14-month study period, across 162 ZIP codes. The number of asthma visits was similar in the 3-day periods before and after spraying (510 vs. 501, p = 0.78). In multivariate analyses, daily rates of asthma visits were not associated with pesticide spraying (rate ratio = 0.92; 95% confidence interval, 0.80–1.07). Secondary analyses among children and for chronic obstructive pulmonary disease yielded similar null results. This analysis shows that spraying pyrethroids for WNV control in New York City was not followed by population-level increases in public hospital ED visit rates for asthma. PMID:15289164
Noel, Jonathan K; Xuan, Ziming; Babor, Thomas F
2017-07-03
Beer marketing in the United States is controlled through self-regulation, whereby the beer industry has created a marketing code and enforces its use. We performed a thematic content analysis on beer ads broadcast during a U.S. college athletic event and determined which themes are associated with violations of a self-regulated alcohol marketing code. 289 beer ads broadcast during the U.S. NCAA Men's and Women's 1999-2008 basketball tournaments were assessed for the presence of 23 thematic content areas. Associations between themes and violations of the U.S. Beer Institute's Marketing and Advertising Code were determined using generalized linear models. Humor (61.3%), taste (61.0%), masculinity (49.2%), and enjoyment (36.5%) were the most prevalent content areas. Nine content areas (i.e., conformity, ethnicity, sensation seeking, sociability, romance, special occasions, text responsibility messages, tradition, and individuality) were positively associated with code violations (p < 0.001-0.042). There were significantly more content areas positively associated with code violations than content areas negatively associated with code violations (p < 0.001). Several thematic content areas were positively associated with code violations. The results can inform existing efforts to revise self-regulated alcohol marketing codes to ensure better protection of vulnerable populations. The use of several themes is concerning in relation to adolescent alcohol use and health disparities.
NASA Technical Reports Server (NTRS)
Hunt, J. L.; Souders, S. W.
1975-01-01
Normal- and oblique-shock flow parameters for air in thermochemical equilibrium are tabulated as a function of shock angle for altitudes ranging from 15.24 km to 91.44 km in increments of 7.62 km at selected hypersonic speeds. Post-shock parameters tabulated include flow-deflection angle, velocity, Mach number, compressibility factor, isentropic exponent, viscosity, Reynolds number, entropy difference, and static pressure, temperature, density, and enthalpy ratios across the shock. A procedure is presented for obtaining oblique-shock flow properties in equilibrium air on surfaces at various angles of attack, sweep, and dihedral by use of the two-dimensional tabulations. Plots of the flow parameters against flow-deflection angle are presented at altitudes of 30.48, 60.96, and 91.44 km for various stream velocities.
Experimental study of a generic high-speed civil transport: Tabulated data
NASA Technical Reports Server (NTRS)
Belton, Pamela S.; Campbell, Richard L.
1992-01-01
An experimental study of a generic high-speed civil transport was conducted in LaRC's 8-Foot Transonic Pressure Tunnel. The data base was obtained for the purpose of assessing the accuracy of various levels of computational analysis. Two models differing only in wing tip geometry were tested with and without flow-through nacelles. The baseline model has a curved or crescent wing tip shape while the second model has a more conventional straight wing tip shape. The study was conducted at Mach numbers from 0.30-1.19. Force data were obtained on both the straight and curved wing tip models. Only the curved wing tip model was instrumented for measuring pressures. Longitudinal and lateral-directional aerodynamic data are presented without analysis in tabulated form. Pressure coefficients for the curved wing tip model are also presented in tabulated form.
2012 Survey of Reserve Components Spouses (RCSS): Tabulations of Responses
2012-09-30
injury/medical problems Child care problems Other family/personal obligation Maternity / paternity leave Labor dispute Weather affected job School...Did you interact with the unit or Service point of contact? ............................. 262 36. How satisfied are you with the level of assistance...I did not interact with the unit or Service point of contact were tabulated separately, as responses to the constructed question Did you interact
2015 Workplace and Gender Relations Survey of Reserve Component Members: Tabulations of Responses
2016-03-17
Report 6. AUTHOR(S) DefenseResearch, Surveys,andStatistics Center (RSSC) 7. PERFORMING ORGANIZATION NAME(S) AND ADDRESS(ES) Defense Manpower Data Center...2015 WORKPLACE AND GENDER RELATIONS SURVEY OF RESERVE COMPONENT MEMBERS: TABULATIONS OF RESPONSES Defense Manpower Data Center Defense Research...Surveys, and Statistics Center 4800 Mark Center Drive, Suite 04E25-01, Alexandria, VA 22350-4000 ii DMDC Acknowledgments The Defense Manpower Data
Natural resources information system.
NASA Technical Reports Server (NTRS)
Leachtenauer, J. C.; Woll, A. M.
1972-01-01
A computer-based Natural Resources Information System was developed for the Bureaus of Indian Affairs and Land Management. The system stores, processes and displays data useful to the land manager in the decision making process. Emphasis is placed on the use of remote sensing as a data source. Data input consists of maps, imagery overlays, and on-site data. Maps and overlays are entered using a digitizer and stored as irregular polygons, lines and points. Processing functions include set intersection, union and difference and area, length and value computations. Data output consists of computer tabulations and overlays prepared on a drum plotter.
Hill, B.R.; Hill, J.R.; Nolan, K.M.
1988-01-01
Data were collected during a 4-yr study of sediment sources in four drainage basins tributary to Lake Tahoe, California-Nevada. The study areas include the Blackwood, General, Edgewood, and Logan House Creek basins. Data include changes in bank and bed positions at channel cross sections; results of stream-channel mapping; analyses of bank and bed material samples; tabulations of bed material point counts; measured rates of hillslope erosion; dimensions of gullies; suspended-sediment data collected during synoptic snowmelt sampling; and physiographic data for the four study basins. (USGS)
Hill, B.R.; Hill, J.R.; Nolan, K.M.
1990-01-01
Data were collected during a 5-year study of sediment sources in four drainage basins tributary to Lake Tahoe, California-Nevada. The study areas include the Blackwood Creek, General Creek, Edgewood Creek, and Logan House Creek basins. Data include changes in bank and bed positions at channel cross sections; results of stream-channel inventories; analyses of bank and bed material samples; tabulations of bed-material pebble counts; measured rates of hillslope erosion; dimensions of gullies; suspended-sediment data collected during synoptic snowmelt sampling; and physiographic data for the four study basins. (USGS)
NASA Technical Reports Server (NTRS)
Vostreys, R. W.
1981-01-01
Sounding rockets, artificial Earth satellites, and space probes launched betweeen January 1 and December 31, 1980 are listed. Data tabulated for the rocket launchings show launching site, instruments carried, date of launch, agency rocket identification, sponsoring country, experiment discipline, peak altitude, and the experimenter or institution responsible. Tables for satellites and space probes show COSPAR designation, spacecraft name, country, launch date, epoch date, orbit type, apoapsis, periapsis and inclination period. The functions and responsibilities of the World Data Center and the areas of scientific interest at the seven subcenters are defined. An alphabetical listing of experimenters using the sounding rockets is also provided.
The 25-MB sounding data and synoptic charts for NASA's AVE 2 pilot experiment
NASA Technical Reports Server (NTRS)
Scoggins, J. R.; Turner, R. E.
1975-01-01
Rawinsonde data were tabulated at 25-mb intervals from the surface to 25 mb for the 54 stations participating in the atmospheric variability experiment 2 Pilot Experiment which began at 12 Greenwich mean time on May 11 and ended at 12 Greenwich mean time on May 12, 1974. Soundings were made at 3 hour intervals. Methods of processing and data accuracy are discussed, and synoptic charts prepared from the data are presented. The area covered by the sounding stations is the eastern United States east of approximately 105 deg west longitude.
Satellite-aided evaluation of population exposure to air pollution
NASA Technical Reports Server (NTRS)
Todd, W. J.; George, A. J., Jr.; Bryant, N. A.
1979-01-01
The evaluation of population exposure to air pollution through the computer processing of Landsat digital land use data, along with total suspended particulate estimates and population data by census tracts, is demonstrated. Digital image processing was employed to analyze simultaneously data from Landsat MSS bands 4 through 7 in order to extract land use and land cover information. The three data sets were spatially registered in a digital format, compatible with integrated computer processing, and cross-tabulated. A map illustrating relative air quality by 2-sq km cells for the residential population in the Portland, Oregon area is obtained.
Spiegelman, M
1966-06-01
The plan calls for monographs on 16 specific topics and a review volume. The specific topics include a monograph on fertility, another on marriage and divorce, and one on social and economic factors in mortality, all of which are of obvious demographic interest. The core of the development of the monographs dealing with the disease topics, and accidents, homicide, and suicide, is a tabulation program involving specific and age-adjusted death rates based upon the deaths in the United States for the three year period 1959-61 and the 1960 census of population. The program called for a set of standard tabulations common to each monograph and also a set of special tabulations tailored to meet the particular needs for each topic as specified by its authors. This wealth of demographic detail regarding mortality in the United States during 1959-61 fills only one aspect of its health picture. To fill out the health picture, use is being made of the great volume of morbidity data out of the National Health Interview and Examination Surveys and a wide variety of other specialized surveys. In all of these, the findings are invariably described in terms of the demographic characteristics of age and sex, and frequently in terms of such social-economic variables as education, income, work status, and occupational class.A large part of the content of the monograph on social and economic factors in mortality will come out of available results of the 1960 census-death certificate matching study at the University of Chicago. One important feature of this monograph, not contained in the Chicago study, is a regression and correlation analysis of total mortality in 202 Standard Metropolitan Statistical Areas in relation to a large number of social, economic, and climatic factors. Among the features in the monograph on fertility are chapters on the medical and biological characteristics of births, fecundity and family planning, and an analysis of cohort fertility trends. The monograph on marriage and divorce will make use of the detailed tabulations from the 1960 census on marital status in relation to education, race, parentage, earnings and other classifications of individuals and of husband and wife in married couples. It will also contain a matching study of divorce records in certain states for a short period before the census of 1960 with the returns in the census. The project is supported by a grant from the United States Public Health Service (CH 00075; formerly RG 08262). Publication will be by the Harvard University Press.
NASA Technical Reports Server (NTRS)
Rogers, R. H. (Principal Investigator)
1975-01-01
The author has identified the following significant results. Preliminary results in Saginaw Bay show that processed LANDSAT data provides a synoptic view of turbidity and circulation patterns that no degree of ground monitoring can provide. Processed imagery was produced to show nine discrete categories of turbidity, as indicated by nine Secchi depths between 0.3 and 3.3 meters. Analysis of lakes near Madison, Wisconsin show that inland lake water can be categorized by LANDSAT as clear, tannin, algal, and red clay. LANDSAT's capability to inventory watershed land use was throughly demonstrated in the Ohio-Kentucky-Indiana regional planning area. Computer tabulations providing area covered by each of 16 land use categories were rapidly and economically produced for each of the 225 watersheds and nine counties.
NASA Technical Reports Server (NTRS)
Clinton, N. J. (Principal Investigator)
1980-01-01
Labeling errors made in the large area crop inventory experiment transition year estimates by Earth Observation Division image analysts are identified and quantified. The analysis was made from a subset of blind sites in six U.S. Great Plains states (Oklahoma, Kansas, Montana, Minnesota, North and South Dakota). The image interpretation basically was well done, resulting in a total omission error rate of 24 percent and a commission error rate of 4 percent. The largest amount of error was caused by factors beyond the control of the analysts who were following the interpretation procedures. The odd signatures, the largest error cause group, occurred mostly in areas of moisture abnormality. Multicrop labeling was tabulated showing the distribution of labeling for all crops.
Sweetkind, D.S.; Taylor, E.M.
2010-01-01
Downhole lithologic information and aquifer pumping test data are reported from 464 wells from a broad area of the northern part of the Coast Ranges in California. These data were originally published in paper form as numerous tables within three USGS Water-Supply Papers describing geology and groundwater conditions in Napa and Sonoma Valleys, the Santa Rosa and Petaluma Valley areas, and in the Russian River Valley and areas in Sonoma and Mendocino Counties, Calif. The well data are compiled in this report in digital form suitable for use in a digital mapping environment. These data, although mostly from relatively shallow water wells, provide important subsurface information that displays the disposition and facies transition of lithologic units throughout this broad area. Well lithologic data themselves and simple three-dimensional interpolation of those data show distinct spatial patterns that are linked to subsurface stratigraphy and structure and can be used to aid in the assessment of the groundwater resources.
Current and anticipated uses of thermal hydraulic codes in Korea
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kim, Kyung-Doo; Chang, Won-Pyo
1997-07-01
In Korea, the current uses of thermal hydraulic codes are categorized into 3 areas. The first application is in designing both nuclear fuel and NSSS. The codes have usually been introduced based on the technology transfer programs agreed between KAERI and the foreign vendors. Another area is in the supporting of the plant operations and licensing by the utility. The third category is research purposes. In this area assessments and some applications to the safety issue resolutions are major activities using the best estimate thermal hydraulic codes such as RELAP5/MOD3 and CATHARE2. Recently KEPCO plans to couple thermal hydraulic codesmore » with a neutronics code for the design of the evolutionary type reactor by 2004. KAERI also plans to develop its own best estimate thermal hydraulic code, however, application range is different from KEPCO developing code. Considering these activities, it is anticipated that use of the best estimate hydraulic analysis code developed in Korea may be possible in the area of safety evaluation within 10 years.« less
Fragment-to-Lead Medicinal Chemistry Publications in 2015.
Johnson, Christopher N; Erlanson, Daniel A; Murray, Christopher W; Rees, David C
2017-01-12
Fragment-based drug discovery (FBDD) is now well-established as a technology for generating new chemical leads and drugs. This Miniperspective provides a tabulated overview of the fragment-to-lead literature published in the year 2015, together with a commentary on trends observed across the FBDD field during this time. It is hoped that this tabulated summary will provide a useful point of reference for both FBDD practitioners and the wider medicinal chemistry community.
2017 Workplace and Gender Relations Survey of Reserve Component Members: Tabulations of Responses
2018-04-30
2017 Workplace and Gender Relations Survey of Reserve Component Members Tabulations of Responses Additional copies of this report may be...http://www.dtic.mil/dtic/order.html Ask for report by DTIC# OPA Report No. 2018-012 April 2018 2017 Workplace and Gender Relations Survey of...Alexandria, VA 22350-4000 2017 Workplace and Gender Relations Survey of Reserve Component Members ii OPA Acknowledgments The Office of People Analytics
UniGene Tabulator: a full parser for the UniGene format.
Lenzi, Luca; Frabetti, Flavia; Facchin, Federica; Casadei, Raffaella; Vitale, Lorenza; Canaider, Silvia; Carinci, Paolo; Zannotti, Maria; Strippoli, Pierluigi
2006-10-15
UniGene Tabulator 1.0 provides a solution for full parsing of UniGene flat file format; it implements a structured graphical representation of each data field present in UniGene following import into a common database managing system usable in a personal computer. This database includes related tables for sequence, protein similarity, sequence-tagged site (STS) and transcript map interval (TXMAP) data, plus a summary table where each record represents a UniGene cluster. UniGene Tabulator enables full local management of UniGene data, allowing parsing, querying, indexing, retrieving, exporting and analysis of UniGene data in a relational database form, usable on Macintosh (OS X 10.3.9 or later) and Windows (2000, with service pack 4, XP, with service pack 2 or later) operating systems-based computers. The current release, including both the FileMaker runtime applications, is freely available at http://apollo11.isto.unibo.it/software/
Occupational dermatoses reported in Brazil from 2007 to 2014*
Lise, Michelle Larissa Zini; Feijó, Fernando Ribas; Lise, Michael Laurence Zini; Lise, Claudia Ribeiro Zini; de Campos, Luis Carlos Elejalde
2018-01-01
Background Occupational diseases are very prevalent in the world, especially in developing countries. Occupational dermatoses are responsible for most of these cases. However, epidemiological studies are rare in Brazil. Objectives To verify the panorama of occupational skin diseases in Brazil describing frequencies of work-related dermatoses and their sociodemographic and occupational patterns. Methods We used retrospective data from the Notifiable Diseases Information System, (from 2007-2014) tabulated with the Tab program for Windows - TabWin12. We used intentional non-probability sampling and sequential selection, considering all notified occupational dermatoses. Results All cases of occupational dermatoses referred to in the period were analyzed (n = 4710). Males and the age group of 35-49 years were the most affected. The most affected body area was the upper limb (34.2%) and the hand (25.4%). The "causative agent" field in the forms was not filled in 69.4% of cases, with chrome as the most prevalent cause reported (11.8%). ICD-10 codes more prevalent were L23, L24, and L25, corresponding to 34.2% of the sample. In total, 29% of patients needed to take a sick leave. No cases evolved to death and there were 0.2% of total as permanent disability. Study limitations The amount of missing information for various items in the system draws attention. Conclusions Treatment of patients with occupational dermatitis include the identification and removal of the causative agent and specific treatment of the disease. Diagnosis delay in cases of occupational dermatoses brings social and financial consequences to the work and life of workers. PMID:29641693
Geospatial Information System Analysis of Healthcare Need and Telemedicine Delivery in California.
Kaufman, Taylor; Geraghty, Estella M; Dullet, Navjit; King, Jesse; Kissee, Jamie; Marcin, James P
2017-05-01
Geospatial Information Systems (GIS) superimpose data on geographical maps to provide visual representations of data by region. Few studies have used GIS data to investigate if telemedicine services are preferentially provided to communities of greatest need. This study compared the healthcare needs of communities with and without telemedicine services from a university-based telemedicine program. Originating sites for all telemedicine consultations between July 1996 and December 2013 were geocoded using ArcGIS software. ZIP Code Tabulation Areas (ZCTAs) were extracted from the 2010 U.S. Census Bureau's Topologically Integrated Geographic Encoding and Referencing file and assigned a community needs index (CNI) score to reflect the ZCTA community's healthcare needs based on evidence-based barriers to healthcare access. CNI scores were compared across communities with and without active telemedicine services. One hundred ninety-four originating telemedicine clinic sites in California were evaluated. The mean CNI score for ZCTAs with at least one telemedicine clinic was significantly higher (3.32 ± 0.84) than those without a telemedicine site (2.95 ± 0.99) and higher than the mean ZCTAs for all of California (2.99 ± 1.01). Of the 194 telemedicine clinics, 71.4% were located in communities with above average need and 33.2% were located in communities with very high needs. Originating sites receiving telemedicine services from a university-based telemedicine program were located in regions with significantly higher community healthcare needs. Leveraging a geospatial information system to understand community healthcare needs provides an opportunity for payers, hospitals, and patients to be strategic in the allocation of telemedicine services.
A Population-Based Assessment of Heartburn in Urban Black Americans
Friedenberg, Frank K.; Makipour, Kian; Palit, Amiya; Shah, Sweetang; Vanar, Vishwas; Richter, Joel E.
2013-01-01
Background Prevalence data for heartburn in the urban Black American community is lacking. In order to estimate prevalence for this community we analyzed data from an ongoing cohort study in progress at our hospital. Comprehensive interviews allowed for exploration of factors associated with heartburn. Methods Complex, stratified sampling design. Survey invitations are hand delivered to random blocks in a single zip code tabulation area. One member per eligible household is invited to complete a computer-based survey. Heartburn was defined as ≥ 3 days/week of symptoms as defined by the Montreal Definition and Classification of GERD. Scaling and weighting factors were utilized to estimate population-level prevalence. Multivariate logistic regression was used to identify independent predictor variables for heartburn. Results Enrolled 379 participants corresponding to a weighted sample size of 22,409 (20,888–23,930) citizens. Demographic characteristics of the sample closely matched those of the entire targeted population. Overall, the weighted prevalence of heartburn ≥ 3 times per week was 17.6% (16.4%–18.8%). Variables independently associated with heartburn were BMI, daily caloric and fat intake, diabetes mellitus (OR=2.95; 2.59–3.36), cigarette smoking, and alcohol consumption (OR=2.55; 2.25–2.89). Factors inversely associated included illicit drug use and increased physical activity. Waist: hip ratio showed no relationship. Conclusions The prevalence of heartburn ≥ 3 times per week is high in the Black American community. Adverse lifestyle behaviors showed particularly important associations. Our study needs to be replicated in other communities with similar demographics. PMID:23237330
Residential Segregation and the Availability of Primary Care Physicians
Gaskin, Darrell J; Dinwiddie, Gniesha Y; Chan, Kitty S; McCleary, Rachael R
2012-01-01
Objective To examine the association between residential segregation and geographic access to primary care physicians (PCPs) in metropolitan statistical areas (MSAs). Data Sources We combined zip code level data on primary care physicians from the 2006 American Medical Association master file with demographic, socioeconomic, and segregation measures from the 2000 U.S. Census. Our sample consisted of 15,465 zip codes located completely or partially in an MSA. Methods We defined PCP shortage areas as those zip codes with no PCP or a population to PCP ratio of >3,500. Using logistic regressions, we estimated the association between a zip code's odds of being a PCP shortage area and its minority composition and degree of segregation in its MSA. Principal Findings We found that odds of being a PCP shortage area were 67 percent higher for majority African American zip codes but 27 percent lower for majority Hispanic zip codes. The association varied with the degree of segregation. As the degree of segregation increased, the odds of being a PCP shortage area increased for majority African American zip codes; however, the converse was true for majority Hispanic and Asian zip codes. Conclusions Efforts to address PCP shortages should target African American communities especially in segregated MSAs. PMID:22524264
Recent Updates to the MELCOR 1.8.2 Code for ITER Applications
DOE Office of Scientific and Technical Information (OSTI.GOV)
Merrill, Brad J
This report documents recent changes made to the MELCOR 1.8.2 computer code for application to the International Thermonuclear Experimental Reactor (ITER), as required by ITER Task Agreement ITA 81-18. There are four areas of change documented by this report. The first area is the addition to this code of a model for transporting HTO. The second area is the updating of the material oxidation correlations to match those specified in the ITER Safety Analysis Data List (SADL). The third area replaces a modification to an aerosol tranpsort subroutine that specified the nominal aerosol density internally with one that now allowsmore » the user to specify this density through user input. The fourth area corrected an error that existed in an air condensation subroutine of previous versions of this modified MELCOR code. The appendices of this report contain FORTRAN listings of the coding for these modifications.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
NONE
1995-11-03
In March 1995, Affiliated Engineers SE, Inc. (AESE) was retained by the Mobile District U.S. Army Corps of Engineers to perform a Limited Energy Study for Holston Army Ammunition Plant, Kingsport, Tennessee. The field survey of existing conditions was completed in May 1995. The results of this field survey were subsequently tabulated and used to generate single line building drawings on Autocad. This report summarizes the results obtained from this field investigation and the analysis of various alternative Energy Conservation Opportunities (ECO`s). To develop the field data into various alternative ECO concepts or models, we utilized an `Excel` spreadsheet tomore » tabulate and compare energy consumption, installation and operating costs for various ECO`s. These ECO`s were then analyzed for suitability for the Energy Conservation Investment Program (ECIP) using the government`s software package called Life Cycle Cost in Design (LCCID). The Scope of Work developed by the U.S. Army Corps of Engineers gave the following tasks: (1) Perform a field survey to gather information on existing operating conditions and equipment at Hoiston Army Ammunition Plant, Area `A`. (2) Perform a field survey to gather information on existing boilers laid away at Volunteer Army Ammunition Plant in Chattanooga, Tennessee. (3) Provide a list of suggested ECO`s. (4) Analyze ECO`s using the LCCID program. (5) Perform savings to investment ratio (SIR) calculation. (6) Rank ECO`s per SIR`s. (7) Provide information on study assumptions and document equations used in calculations. (8) Perform Life Cycle Cost Analysis. (9) Perform Synergism Analysis. (10) Calculate Energy/Cost Ratios. (11) Calculate Benefit/Cost Ratios. (12) Provide documentation in the form of Project Development Brochures (PDB`s) and DD Form 139« less
Nichols, D.J.; Flores, R.M.
2006-01-01
A biostratigraphic (palynostratigraphic) zonation of Paleocene rocks was established in the northeastern Wind River Basin near Waltman, Natrona County, Wyoming, in 1978 and subsequently applied extensively by various workers throughout the Rocky Mountain region. Because the original study on which the zonation was based was proprietary, precise details about the locations of the two reference sections and the samples on which the zonation was based were not published and are no longer retrievable. Therefore, it is useful (although not required) to designate formally a new reference section for the Paleocene biozones. Accordingly, exposures of Paleocene and associated strata within and west of the Castle Gardens Petroglyph Site in Fremont County, Wyoming, in the east-central part of the Wind River Basin, were selected for this purpose. At this location, composite stratigraphic sections encompassing 740 m of strata were measured, described, and sampled. Productive samples yielded characteristic Maastrichtian palynomorphs from the lower part of the sampled interval and diagnostic species of the six palynological biozones zones widely known as P1 (lower Paleocene) through P6 (upper Paleocene), through an interval of about 580 m. The Paleocene biozones are present in the same consistent stratigraphic order in the Castle Gardens area as observed in the 1978 study and subsequent studies throughout the Rocky Mountain region. In accordance with the North American Stratigraphic Code, the historical background is presented; intent to establish the Castle Gardens reference section is declared; the category, rank, and formal names of biostratigraphic units within it are specified; and the features of the biozonation are described, including biozone boundaries, ages, and regional relations. Occurrences of biostratigraphically significant palynological species within each biozone in the reference section are tabulated, and presence of these and other species in correlative biozones are discussed. The new reference section in the Castle Gardens area replaces the original reference sections near Waltman and provides a better-documented foundation for Paleocene palynostratigraphy in the Rocky Mountain region.
Zhou, Yuhong; Bemanian, Amin; Beyer, Kirsten M M
2017-04-01
Background: Residential racial segregation is still neglected in contemporary examinations of racial health disparities, including studies of cancer. Even fewer studies examine the processes by which segregation occurs, such as through housing discrimination. This study aims to examine relationships among housing discrimination, segregation, and colorectal cancer survival in southeastern Wisconsin. Methods: Cancer incidence data were obtained from the Wisconsin Cancer Reporting System for two southeastern Wisconsin metropolitan areas. Two indices of mortgage discrimination were derived from Home Mortgage Disclosure Act data, and a measure of segregation (the location quotient) was calculated from U.S. census data; all predictors were specified at the ZIP Code Tabulation Area level. Cox proportional hazards regression was used to examine associations between mortgage discrimination, segregation, and colorectal cancer survival in southeastern Wisconsin. Results: For all-cause mortality, racial bias in mortgage lending was significantly associated with a greater hazard rate among blacks [HR = 1.37; 95% confidence interval (CI), 1.06-1.76] and among black women (HR = 1.53; 95% CI, 1.06-2.21), but not black men in sex-specific models. No associations were identified for redlining or the location quotient. Additional work is needed to determine whether these findings can be replicated in other geographical settings. Conclusions: Our findings indicate that black women in particular experience poorer colorectal cancer survival in neighborhoods characterized by racial bias in mortgage lending, a measure of institutional racism. These findings are in line with previous studies of breast cancer survival. Impact: Housing discrimination and institutional racism may be important targets for policy change to reduce health disparities, including cancer disparities. Cancer Epidemiol Biomarkers Prev; 26(4); 561-8. ©2017 AACR See all the articles in this CEBP Focus section, "Geospatial Approaches to Cancer Control and Population Sciences." ©2017 American Association for Cancer Research.
Evans, S; Lewis, H; Williamsen, J; Evans, H; Bohl, W
2004-01-01
Orbital debris impacts on the International Space Station occur frequently. To date, none of the impacting particles has been large enough to penetrate manned pressurized volumes. We used the Manned Spacecraft Crew Survivability code to evaluate the risk to crew of penetrations of pressurized modules at two assembly stages: after Flight 1J, when the pressurized elements of Kibo, the Japanese Experiment Module, are present, and after Flight 1E, when the European Columbus Module is present. Our code is a Monte-Carlo simulation of impacts on the Station that considers several potential event types that could lead to crew loss. Among the statistics tabulated by the program is the probability of death of one or more crew members in the event of a penetration, expressed as the risk factor, R. This risk factor is dependent on details of crew operations during both ordinary circumstances and decompression emergencies, as well as on details of internal module configurations. We conducted trade studies considering these procedure and configuration details to determine the bounds on R at the 1J and 1E stages in the assembly sequence. Here we compare the R-factor bounds, and procedures could that reduce R at these stages. Published by Elsevier Ltd on behalf of COSPAR.
NASA Technical Reports Server (NTRS)
Evans, S.; Lewis, H.; Williamsen, J.; Evans, H.; Bohl, W.
2004-01-01
Orbital debris impacts on the International Space Station occur frequently. To date, none of the impacting particles has been large enough to penetrate manned pressurized volumes. We used the Manned Spacecraft Crew Survivability code to evaluate the risk to crew of penetrations of pressurized modules at two assembly stages: after Flight 1J, when the pressurized elements of Kibo, the Japanese Experiment Module, are present, and after Flight 1E, when the European Columbus Module is present. Our code is a Monte-Carlo simulation of impacts on the Station that considers several potential event types that could lead to crew loss. Among the statistics tabulated by the program is the probability of death of one or more crew members in the event of a penetration, expressed as the risk factor, R. This risk factor is dependent on details of crew operations during both ordinary circumstances and decompression emergencies, as well as on details of internal module configurations. We conducted trade studies considering these procedure and configuration details to determine the bounds on R at the 1J and 1E stages in the assembly sequence. Here we compare the R-factor bounds, and procedures could that reduce R at these stages. Published by Elsevier Ltd on behalf of COSPAR.
An ethical paradox: the effect of unethical conduct on medical students' values
Satterwhite, R.; Satterwhite, W.; Enarson, C.
2000-01-01
Objective—To report the ethical development of medical students across four years of education at one medical school. Design and setting—A questionnaire was distributed to all four classes at the Wake Forest University School of Medicine during the Spring of 1996. Participants—Three hundred and three students provided demographic information as well as information concerning their ethical development both as current medical students and future interns. Main measurements—Results were analyzed using cross-tabulations, correlations, and analysis of variance. Results—Results suggested that the observation of and participation in unethical conduct1 may have disparaging effects on medical students' codes of ethics with 35% of the total sample (24% of first years rising to 55% of fourth years) stating that derogatory comments made by residents/attendings, either in the patient's presence or absence, were "sometimes" or "often" appropriate. However, approximately 70% of the sample contended that their personal code of ethics had not changed since beginning medical school and would not change as a resident. Conclusions—Results may represent an internal struggle that detracts from the medical school experience, both as a person and as a doctor. Our goal as educators is to alter the educational environment so that acceptance of such behaviour is not considered part of becoming a physician. Key Words: Ethics • ethical development • paradox • medical students • derogatory comments PMID:11270946
Algorithm for Lossless Compression of Calibrated Hyperspectral Imagery
NASA Technical Reports Server (NTRS)
Kiely, Aaron B.; Klimesh, Matthew A.
2010-01-01
A two-stage predictive method was developed for lossless compression of calibrated hyperspectral imagery. The first prediction stage uses a conventional linear predictor intended to exploit spatial and/or spectral dependencies in the data. The compressor tabulates counts of the past values of the difference between this initial prediction and the actual sample value. To form the ultimate predicted value, in the second stage, these counts are combined with an adaptively updated weight function intended to capture information about data regularities introduced by the calibration process. Finally, prediction residuals are losslessly encoded using adaptive arithmetic coding. Algorithms of this type are commonly tested on a readily available collection of images from the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) hyperspectral imager. On the standard calibrated AVIRIS hyperspectral images that are most widely used for compression benchmarking, the new compressor provides more than 0.5 bits/sample improvement over the previous best compression results. The algorithm has been implemented in Mathematica. The compression algorithm was demonstrated as beneficial on 12-bit calibrated AVIRIS images.
Shuttle cryogenic supply system optimization study. Volume 5A-1: Users manual for math models
NASA Technical Reports Server (NTRS)
1973-01-01
The Integrated Math Model for Cryogenic Systems is a flexible, broadly applicable systems parametric analysis tool. The program will effectively accommodate systems of considerable complexity involving large numbers of performance dependent variables such as are found in the individual and integrated cryogen systems. Basically, the program logic structure pursues an orderly progression path through any given system in much the same fashion as is employed for manual systems analysis. The system configuration schematic is converted to an alpha-numeric formatted configuration data table input starting with the cryogen consumer and identifying all components, such as lines, fittings, and valves, each in its proper order and ending with the cryogen supply source assembly. Then, for each of the constituent component assemblies, such as gas generators, turbo machinery, heat exchangers, and accumulators, the performance requirements are assembled in input data tabulations. Systems operating constraints and duty cycle definitions are further added as input data coded to the configuration operating sequence.
GPU accelerated simulations of three-dimensional flow of power-law fluids in a driven cube
NASA Astrophysics Data System (ADS)
Jin, K.; Vanka, S. P.; Agarwal, R. K.; Thomas, B. G.
2017-01-01
Newtonian fluid flow in two- and three-dimensional cavities with a moving wall has been studied extensively in a number of previous works. However, relatively a fewer number of studies have considered the motion of non-Newtonian fluids such as shear thinning and shear thickening power law fluids. In this paper, we have simulated the three-dimensional, non-Newtonian flow of a power law fluid in a cubic cavity driven by shear from the top wall. We have used an in-house developed fractional step code, implemented on a Graphics Processor Unit. Three Reynolds numbers have been studied with power law index set to 0.5, 1.0 and 1.5. The flow patterns, viscosity distributions and velocity profiles are presented for Reynolds numbers of 100, 400 and 1000. All three Reynolds numbers are found to yield steady state flows. Tabulated values of velocity are given for the nine cases studied, including the Newtonian cases.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sehmel, G.A.
1989-05-01
Thermodynamic data for aqueous species and solids that contain cyanide and antimony were tabulated from several commonly accepted, published sources of thermodynamic data and recent journal article. The review does not include gases or organic complexes of either antimony or cyanide, nor does the review include the sulfur compounds of cyanide. The basic thermodynamic data, ..delta..G/sub f,298//sup o/, ..delta..H/sub f,298//sup o/, and S/sub f//sup o/ values, were chosen to represent each solid phase and aqueous species for which data were available in the appropriate standard state. From these data the equilibrium constants (log K/sub r,298//sup o/) and enthalpies of reactionmore » (..delta..H/sub r,298//sup o/) at 298 K (25/degree/C) were calculated for reactions involving the formation of these aqueous species and solids from the basic components. 34 refs., 14 tabs.« less
Ringe, Stefan; Oberhofer, Harald; Hille, Christoph; Matera, Sebastian; Reuter, Karsten
2016-08-09
The size-modified Poisson-Boltzmann (MPB) equation is an efficient implicit solvation model which also captures electrolytic solvent effects. It combines an account of the dielectric solvent response with a mean-field description of solvated finite-sized ions. We present a general solution scheme for the MPB equation based on a fast function-space-oriented Newton method and a Green's function preconditioned iterative linear solver. In contrast to popular multigrid solvers, this approach allows us to fully exploit specialized integration grids and optimized integration schemes. We describe a corresponding numerically efficient implementation for the full-potential density-functional theory (DFT) code FHI-aims. We show that together with an additional Stern layer correction the DFT+MPB approach can describe the mean activity coefficient of a KCl aqueous solution over a wide range of concentrations. The high sensitivity of the calculated activity coefficient on the employed ionic parameters thereby suggests to use extensively tabulated experimental activity coefficients of salt solutions for a systematic parametrization protocol.
NASA Astrophysics Data System (ADS)
Parviainen, Hannu
2015-10-01
PyLDTk automates the calculation of custom stellar limb darkening (LD) profiles and model-specific limb darkening coefficients (LDC) using the library of PHOENIX-generated specific intensity spectra by Husser et al. (2013). It facilitates exoplanet transit light curve modeling, especially transmission spectroscopy where the modeling is carried out for custom narrow passbands. PyLDTk construct model-specific priors on the limb darkening coefficients prior to the transit light curve modeling. It can also be directly integrated into the log posterior computation of any pre-existing transit modeling code with minimal modifications to constrain the LD model parameter space directly by the LD profile, allowing for the marginalization over the whole parameter space that can explain the profile without the need to approximate this constraint by a prior distribution. This is useful when using a high-order limb darkening model where the coefficients are often correlated, and the priors estimated from the tabulated values usually fail to include these correlations.
NORAD LOOK ANGLES AND PIO SATELLITE PACKAGE
NASA Technical Reports Server (NTRS)
ANONYMOUS
1994-01-01
This program package consists of two programs. First is the NORAD Look Angles Program, which computes satellite look angles (azimuth, elevation, and range) as well as the subsatellite points (latitude, longitude, and height). The second program in this package is the PIO Satellite Program, which computes sighting directions, visibility times, and the maximum elevation angle attained during each pass of an earth-orbiting satellite. Computations take into consideration the observing location and the effect of the earth's shadow on the satellite visibility. Input consists of a magnetic tape prepared by the NORAD Look Angles Program and punched cards containing reference Julian date, right ascension, declination, mean sidereal time at zero hours universal time of the reference date, and daily changes of these quantities. Output consists of a tabulated listing of the satellite's rise and set times, direction, and the maximum elevation angle visible from each observing location. This program has been implemented on the GE 635. The program Assembler code can easily be replaced by FORTRAN statements.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ryadovikov, V. N., E-mail: riadovikov@ihep.ru; Aleev, A. N.; Ardashev, E. N.
2016-03-15
The results of an analysis of data from the SERP-E-184 experiment devoted to studying mechanisms of the production of charmed particles in proton–nucleus interactions at 70 GeV and their decays are presented. The data in question were obtained upon irradiating the SVD-2 active target consisting of carbon, silicon, and lead plates with a beam of 70-GeV protons. A detailed simulation on the basis of the FRITIOF7.02 and GEANT3.21 code packages made it possible to optimize event-selection criteria and to calculate the detection efficiency for Λ{sub c}{sup +} baryons. After selecting a signal from the threebody decay of a Λ{sub c}{supmore » +} baryon, the inclusive cross section for its production at near-threshold energies, its lifetime, and the parameter of the A dependence of the cross section were found. The Λ{sub c}{sup +} -baryon yields are tabulated along with data from other experiments and theoretical predictions.« less
NASA Technical Reports Server (NTRS)
Nelson, D. P.; Morris, P. M.
1980-01-01
The component detail design drawings of the one sixth scale model of the variable cycle engine testbed demonstrator exhaust syatem tested are presented. Also provided are the basic acoustic and aerodynamic data acquired during the experimental model tests. The model drawings, an index to the acoustic data, an index to the aerodynamic data, tabulated and graphical acoustic data, and the tabulated aerodynamic data and graphs are discussed.
NASA Technical Reports Server (NTRS)
Omidvar, K.
1977-01-01
The branching ratios in hydrogen-like atoms due to the electric-dipole transitions are tabulated for the initial principal and azimuthal quantum numbers n prime l prime, and final principal and azimuthal quantum numbers n l. Average values with respect to l prime are given. The branching ratios not tabulated, including the initial states n prime yields infinity l prime corresponding to the threshold of the continuum, could be obtained by extrapolation.
Lamb Shift of n = 1 and n = 2 States of Hydrogen-like Atoms, 1 ≤ Z ≤ 110
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yerokhin, V. A.; Shabaev, V. M.
2015-09-15
Theoretical energy levels of the n = 1 and n = 2 states of hydrogen-like atoms with the nuclear charge numbers 1 ≤ Z ≤ 110 are tabulated. The tabulation is based on ab initio quantum electrodynamics calculations performed to all orders in the nuclear binding strength parameter Zα, where α is the fine structure constant. Theoretical errors due to various effects are critically examined and estimated.
A risk adjustment approach to estimating the burden of skin disease in the United States.
Lim, Henry W; Collins, Scott A B; Resneck, Jack S; Bolognia, Jean; Hodge, Julie A; Rohrer, Thomas A; Van Beek, Marta J; Margolis, David J; Sober, Arthur J; Weinstock, Martin A; Nerenz, David R; Begolka, Wendy Smith; Moyano, Jose V
2018-01-01
Direct insurance claims tabulation and risk adjustment statistical methods can be used to estimate health care costs associated with various diseases. In this third manuscript derived from the new national Burden of Skin Disease Report from the American Academy of Dermatology, a risk adjustment method that was based on modeling the average annual costs of individuals with or without specific diseases, and specifically tailored for 24 skin disease categories, was used to estimate the economic burden of skin disease. The results were compared with the claims tabulation method used in the first 2 parts of this project. The risk adjustment method estimated the direct health care costs of skin diseases to be $46 billion in 2013, approximately $15 billion less than estimates using claims tabulation. For individual skin diseases, the risk adjustment cost estimates ranged from 11% to 297% of those obtained using claims tabulation for the 10 most costly skin disease categories. Although either method may be used for purposes of estimating the costs of skin disease, the choice of method will affect the end result. These findings serve as an important reference for future discussions about the method chosen in health care payment models to estimate both the cost of skin disease and the potential cost impact of care changes. Copyright © 2017 American Academy of Dermatology, Inc. Published by Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Spiriti, Justin; Zuckerman, Daniel M.
2015-12-01
Traditional coarse-graining based on a reduced number of interaction sites often entails a significant sacrifice of chemical accuracy. As an alternative, we present a method for simulating large systems composed of interacting macromolecules using an energy tabulation strategy previously devised for small rigid molecules or molecular fragments [S. Lettieri and D. M. Zuckerman, J. Comput. Chem. 33, 268-275 (2012); J. Spiriti and D. M. Zuckerman, J. Chem. Theory Comput. 10, 5161-5177 (2014)]. We treat proteins as rigid and construct distance and orientation-dependent tables of the interaction energy between them. Arbitrarily detailed interactions may be incorporated into the tables, but as a proof-of-principle, we tabulate a simple α-carbon Gō-like model for interactions between dimeric subunits of the hepatitis B viral capsid. This model is significantly more structurally realistic than previous models used in capsid assembly studies. We are able to increase the speed of Monte Carlo simulations by a factor of up to 6700 compared to simulations without tables, with only minimal further loss in accuracy. To obtain further enhancement of sampling, we combine tabulation with the weighted ensemble (WE) method, in which multiple parallel simulations are occasionally replicated or pruned in order to sample targeted regions of a reaction coordinate space. In the initial study reported here, WE is able to yield pathways of the final ˜25% of the assembly process.
Enforcement authority and vegetation change at Kumbhalgarh wildlife sanctuary, Rajasthan, India.
Robbins, Paul F; Chhangani, Anil K; Rice, Jennifer; Trigosa, Erika; Mohnot, S M
2007-09-01
Land cover change in protected areas is often associated with human use, especially illicit extraction, but the direction and spatial distribution of such effects and their drivers are poorly understood. We analyze and explain the spatial distribution of vegetation change at the Kumbhalgarh Wildlife Sanctuary in the Aravalli range of Rajasthan, India using remotely sensed data and observation of conservation institutions. Two satellite images are examined in time series over the 13 years following the founding of the sanctuary through a cross-tabulation technique of dominant classes of vegetation density. The resulting change trajectories are compared for their relative distance to high-traffic forest entrance points for local users. The results show 28% of the study area undergoing change, though in multiple trajectories, with both increasing and decreasing density of vegetation in discrete patches. Areas of change are shown to be closer to entrance points than areas experiencing no change. The patchiness of change results from complex issues in local enforcement authority for middle and lower-level officials in Forest Department bureaucracy, leading to further questions about the efficacy and impact of use restrictions in Protected Areas.
The historical biogeography of Mammalia
Springer, Mark S.; Meredith, Robert W.; Janecka, Jan E.; Murphy, William J.
2011-01-01
Palaeobiogeographic reconstructions are underpinned by phylogenies, divergence times and ancestral area reconstructions, which together yield ancestral area chronograms that provide a basis for proposing and testing hypotheses of dispersal and vicariance. Methods for area coding include multi-state coding with a single character, binary coding with multiple characters and string coding. Ancestral reconstruction methods are divided into parsimony versus Bayesian/likelihood approaches. We compared nine methods for reconstructing ancestral areas for placental mammals. Ambiguous reconstructions were a problem for all methods. Important differences resulted from coding areas based on the geographical ranges of extant species versus the geographical provenance of the oldest fossil for each lineage. Africa and South America were reconstructed as the ancestral areas for Afrotheria and Xenarthra, respectively. Most methods reconstructed Eurasia as the ancestral area for Boreoeutheria, Euarchontoglires and Laurasiatheria. The coincidence of molecular dates for the separation of Afrotheria and Xenarthra at approximately 100 Ma with the plate tectonic sundering of Africa and South America hints at the importance of vicariance in the early history of Placentalia. Dispersal has also been important including the origins of Madagascar's endemic mammal fauna. Further studies will benefit from increased taxon sampling and the application of new ancestral area reconstruction methods. PMID:21807730
An equivalent dissipation rate model for capturing history effects in non-premixed flames
Kundu, Prithwish; Echekki, Tarek; Pei, Yuanjiang; ...
2016-11-11
The effects of strain rate history on turbulent flames have been studied in the. past decades with 1D counter flow diffusion flame (CFDF) configurations subjected to oscillating strain rates. In this work, these unsteady effects are studied for complex hydrocarbon fuel surrogates at engine relevant conditions with unsteady strain rates experienced by flamelets in a typical spray flame. Tabulated combustion models are based on a steady scalar dissipation rate (SDR) assumption and hence cannot capture these unsteady strain effects; even though they can capture the unsteady chemistry. In this work, 1D CFDF with varying strain rates are simulated using twomore » different modeling approaches: steady SDR assumption and unsteady flamelet model. Comparative studies show that the history effects due to unsteady SDR are directly proportional to the temporal gradient of the SDR. A new equivalent SDR model based on the history of a flamelet is proposed. An averaging procedure is constructed such that the most recent histories are given higher weights. This equivalent SDR is then used with the steady SDR assumption in 1D flamelets. Results show a good agreement between tabulated flamelet solution and the unsteady flamelet results. This equivalent SDR concept is further implemented and compared against 3D spray flames (Engine Combustion Network Spray A). Tabulated models based on steady SDR assumption under-predict autoignition and flame lift-off when compared with an unsteady Representative Interactive Flamelet (RIF) model. However, equivalent SDR model coupled with the tabulated model predicted autoignition and flame lift-off very close to those reported by the RIF model. This model is further validated for a range of injection pressures for Spray A flames. As a result, the new modeling framework now enables tabulated models with significantly lower computational cost to account for unsteady history effects.« less
An equivalent dissipation rate model for capturing history effects in non-premixed flames
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kundu, Prithwish; Echekki, Tarek; Pei, Yuanjiang
The effects of strain rate history on turbulent flames have been studied in the. past decades with 1D counter flow diffusion flame (CFDF) configurations subjected to oscillating strain rates. In this work, these unsteady effects are studied for complex hydrocarbon fuel surrogates at engine relevant conditions with unsteady strain rates experienced by flamelets in a typical spray flame. Tabulated combustion models are based on a steady scalar dissipation rate (SDR) assumption and hence cannot capture these unsteady strain effects; even though they can capture the unsteady chemistry. In this work, 1D CFDF with varying strain rates are simulated using twomore » different modeling approaches: steady SDR assumption and unsteady flamelet model. Comparative studies show that the history effects due to unsteady SDR are directly proportional to the temporal gradient of the SDR. A new equivalent SDR model based on the history of a flamelet is proposed. An averaging procedure is constructed such that the most recent histories are given higher weights. This equivalent SDR is then used with the steady SDR assumption in 1D flamelets. Results show a good agreement between tabulated flamelet solution and the unsteady flamelet results. This equivalent SDR concept is further implemented and compared against 3D spray flames (Engine Combustion Network Spray A). Tabulated models based on steady SDR assumption under-predict autoignition and flame lift-off when compared with an unsteady Representative Interactive Flamelet (RIF) model. However, equivalent SDR model coupled with the tabulated model predicted autoignition and flame lift-off very close to those reported by the RIF model. This model is further validated for a range of injection pressures for Spray A flames. As a result, the new modeling framework now enables tabulated models with significantly lower computational cost to account for unsteady history effects.« less
Nonstationary Deformation of an Elastic Layer with Mixed Boundary Conditions
NASA Astrophysics Data System (ADS)
Kubenko, V. D.
2016-11-01
The analytic solution to the plane problem for an elastic layer under a nonstationary surface load is found for mixed boundary conditions: normal stress and tangential displacement are specified on one side of the layer (fourth boundary-value problem of elasticity) and tangential stress and normal displacement are specified on the other side of the layer (second boundary-value problem of elasticity). The Laplace and Fourier integral transforms are applied. The inverse Laplace and Fourier transforms are found exactly using tabulated formulas and convolution theorems for various nonstationary loads. Explicit analytical expressions for stresses and displacements are derived. Loads applied to a constant surface area and to a surface area varying in a prescribed manner are considered. Computations demonstrate the dependence of the normal stress on time and spatial coordinates. Features of wave processes are analyzed
Processing and Probability Analysis of Pulsed Terahertz NDE of Corrosion under Shuttle Tile Data
NASA Technical Reports Server (NTRS)
Anastasi, Robert F.; Madaras, Eric I.; Seebo, Jeffrey P.; Ely, Thomas M.
2009-01-01
This paper examines data processing and probability analysis of pulsed terahertz NDE scans of corrosion defects under a Shuttle tile. Pulsed terahertz data collected from an aluminum plate with fabricated corrosion defects and covered with a Shuttle tile is presented. The corrosion defects imaged were fabricated by electrochemically etching areas of various diameter and depth in the plate. In this work, the aluminum plate echo signal is located in the terahertz time-of-flight data and a threshold is applied to produce a binary image of sample features. Feature location and area are examined and identified as corrosion through comparison with the known defect layout. The results are tabulated with hit, miss, or false call information for a probability of detection analysis that is used to identify an optimal processing threshold.
1988-07-01
and SEdward 11oke (We tat) 13a. TYPE OF REPORT 13b. TIME COVERED 14. DATE OF REPORT (Year, Month, Day) 15. PAGE COUNT Interim FROM 86/10 TO 87/06...TABULATIONS OF ENLISTED MARKETS 1. INTRODUCTION The Aray Communications Objectives Measurement System survey has been designed to provide timely information to...1987. During that time 6774 youth, ages 16 through 24, com- pleted the 30 minute ACOMS youth interview. A similar volume is also available for the
Real gas flow fields about three dimensional configurations
NASA Technical Reports Server (NTRS)
Balakrishnan, A.; Lombard, C. K.; Davy, W. C.
1983-01-01
Real gas, inviscid supersonic flow fields over a three-dimensional configuration are determined using a factored implicit algorithm. Air in chemical equilibrium is considered and its local thermodynamic properties are computed by an equilibrium composition method. Numerical solutions are presented for both real and ideal gases at three different Mach numbers and at two different altitudes. Selected results are illustrated by contour plots and are also tabulated for future reference. Results obtained compare well with existing tabulated numerical solutions and hence validate the solution technique.
Oceanic tide maps and spherical harmonic coefficients from Geosat altimetry
NASA Technical Reports Server (NTRS)
Cartwright, D. E.; Ray, R. D.; Sanchez, B. V.
1991-01-01
Maps and tables for the global ocean tides, 69 degree N to 68 degree S, derived from two years of Geosat altimetry are presented. Global maps of local and Greenwich admittance of the (altimetric) ocean tide, and maps of amplitude and Greenwich phase lag of the ocean tide are shown for M(sub 2), S(sub 2), N(sub 2), O(sub 1), and K(sub 1). Larger scale maps of amplitude and phases are also shown for regional areas of special interest. Spherical harmonic coefficients of the ocean tide through degree and order 8 are tabulated for the six major constituents.
DOE Office of Scientific and Technical Information (OSTI.GOV)
NONE
1995-12-31
In June 1995, Affiliated Engineers SE, Inc. (AESE) was retained by the Mobile District U.S. Army Corps of Engineers to perform a Limited Energy Study for Holston Army Ammunition Plant, Kingsport, Tennessee. The field survey of existing conditions was completed in July 1995. The results of this field survey were subsequently tabulated and used to generate single line process flow diagrams on Autocad. A subsequent one day field survey was conducted in August 1995. This report summarizes the results obtained from field investigation and the analysis of various alternative Energy Conservation Opportunities (ECO`s).
Radioactive waste disposal in the marine environment
NASA Astrophysics Data System (ADS)
Anderson, D. R.
In order to find the optimal solution to waste disposal problems, it is necessary to make comparisons between disposal media. It has become obvious to many within the scientific community that the single medium approach leads to over protection of one medium at the expense of the others. Cross media comparisons are being conducted in the Department of Energy ocean disposal programs for several radioactive wastes. Investigations in three areas address model development, comparisons of laboratory tests with field results and predictions, and research needs in marine disposal of radioactive waste. Tabulated data are included on composition of liquid high level waste and concentration of some natural radionuclides in the sea.
The 2008 Circum-Arctic Resource Appraisal
Moore, Thomas E.; Gautier, Donald L.
2017-11-15
Professional Paper 1824 comprises 30 chapters by various U.S. Geological Survey authors, including introduction and methodology chapters, which together provide documentation of the geological basis and methodology of the 2008 Circum-Arctic Resource Appraisal, results of which were first released in August 2008. Twenty-eight chapters summarize the petroleum geology and resource potential of individual, geologically defined provinces north of the Arctic Circle, including those of northern Alaska, northern Canada, east and west Greenland, and most of Arctic Russia, as well as certain offshore areas of the north Atlantic Basin and the Polar Sea. Appendixes tabulate the input and output information used during the assessment.
Berke, Ethan M; Shi, Xun
2009-04-29
Travel time is an important metric of geographic access to health care. We compared strategies of estimating travel times when only subject ZIP code data were available. Using simulated data from New Hampshire and Arizona, we estimated travel times to nearest cancer centers by using: 1) geometric centroid of ZIP code polygons as origins, 2) population centroids as origin, 3) service area rings around each cancer center, assigning subjects to rings by assuming they are evenly distributed within their ZIP code, 4) service area rings around each center, assuming the subjects follow the population distribution within the ZIP code. We used travel times based on street addresses as true values to validate estimates. Population-based methods have smaller errors than geometry-based methods. Within categories (geometry or population), centroid and service area methods have similar errors. Errors are smaller in urban areas than in rural areas. Population-based methods are superior to the geometry-based methods, with the population centroid method appearing to be the best choice for estimating travel time. Estimates in rural areas are less reliable.
, fax, or phone. Use this area to send us your question. Your name: Your E-Mail address: Your Phone Number (voice): Please include area code. Your Fax Number: Please include area code. Send Your Question
NASA Technical Reports Server (NTRS)
Sader, S. A.; Joyce, A. T.
1984-01-01
The relationship between forest clearing, biophysical factors (e.g, ecological zones, slope gradient, soils), and transportation network in Costa Rica was analyzed. The location of forested areas at four reference datas (1940, 1950, 1961, and 1977) as derived from aerial photography and LANDSAT MSS data was digitilized and entered into a geographically-referenced data base. Ecological zones as protrayed by the Holdridge Life Zone Ecology System, and the location of roads and railways were also digitized from maps of the entire country as input to the data base. Information on slope gradient and soils was digitized from maps of a 21,000 square kilometer area. The total area of forest cleared over four decades are related to biophysical factors was analyzed within the data base and deforestation rates and trends were tabulated. The relatiohship between forest clearing and ecological zone and the influence of topography, sils, and transportation network are presented and discussed.
Fundamental limits of scintillation detector timing precision
NASA Astrophysics Data System (ADS)
Derenzo, Stephen E.; Choong, Woon-Seng; Moses, William W.
2014-07-01
In this paper we review the primary factors that affect the timing precision of a scintillation detector. Monte Carlo calculations were performed to explore the dependence of the timing precision on the number of photoelectrons, the scintillator decay and rise times, the depth of interaction uncertainty, the time dispersion of the optical photons (modeled as an exponential decay), the photodetector rise time and transit time jitter, the leading-edge trigger level, and electronic noise. The Monte Carlo code was used to estimate the practical limits on the timing precision for an energy deposition of 511 keV in 3 mm × 3 mm × 30 mm Lu2SiO5:Ce and LaBr3:Ce crystals. The calculated timing precisions are consistent with the best experimental literature values. We then calculated the timing precision for 820 cases that sampled scintillator rise times from 0 to 1.0 ns, photon dispersion times from 0 to 0.2 ns, photodetector time jitters from 0 to 0.5 ns fwhm, and A from 10 to 10 000 photoelectrons per ns decay time. Since the timing precision R was found to depend on A-1/2 more than any other factor, we tabulated the parameter B, where R = BA-1/2. An empirical analytical formula was found that fit the tabulated values of B with an rms deviation of 2.2% of the value of B. The theoretical lower bound of the timing precision was calculated for the example of 0.5 ns rise time, 0.1 ns photon dispersion, and 0.2 ns fwhm photodetector time jitter. The lower bound was at most 15% lower than leading-edge timing discrimination for A from 10 to 10 000 photoelectrons ns-1. A timing precision of 8 ps fwhm should be possible for an energy deposition of 511 keV using currently available photodetectors if a theoretically possible scintillator were developed that could produce 10 000 photoelectrons ns-1.
Fundamental Limits of Scintillation Detector Timing Precision
Derenzo, Stephen E.; Choong, Woon-Seng; Moses, William W.
2014-01-01
In this paper we review the primary factors that affect the timing precision of a scintillation detector. Monte Carlo calculations were performed to explore the dependence of the timing precision on the number of photoelectrons, the scintillator decay and rise times, the depth of interaction uncertainty, the time dispersion of the optical photons (modeled as an exponential decay), the photodetector rise time and transit time jitter, the leading-edge trigger level, and electronic noise. The Monte Carlo code was used to estimate the practical limits on the timing precision for an energy deposition of 511 keV in 3 mm × 3 mm × 30 mm Lu2SiO5:Ce and LaBr3:Ce crystals. The calculated timing precisions are consistent with the best experimental literature values. We then calculated the timing precision for 820 cases that sampled scintillator rise times from 0 to 1.0 ns, photon dispersion times from 0 to 0.2 ns, photodetector time jitters from 0 to 0.5 ns fwhm, and A from 10 to 10,000 photoelectrons per ns decay time. Since the timing precision R was found to depend on A−1/2 more than any other factor, we tabulated the parameter B, where R = BA−1/2. An empirical analytical formula was found that fit the tabulated values of B with an rms deviation of 2.2% of the value of B. The theoretical lower bound of the timing precision was calculated for the example of 0.5 ns rise time, 0.1 ns photon dispersion, and 0.2 ns fwhm photodetector time jitter. The lower bound was at most 15% lower than leading-edge timing discrimination for A from 10 to 10,000 photoelectrons/ns. A timing precision of 8 ps fwhm should be possible for an energy deposition of 511 keV using currently available photodetectors if a theoretically possible scintillator were developed that could produce 10,000 photoelectrons/ns. PMID:24874216
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rechard, Robert P.
This report presents a concise history in tabular form of events leading up to site identification in 1978, site selection in 1987, subsequent characterization, and ongoing analysis through 2009 of the performance of a repository for spent nuclear fuel and high - level radioactive waste at Yucca Mountain in southern Nevada. The tabulated events generally occurred in five periods: (1) commitment to mined geologic disposal and identification of sites; (2) site selection and analysis, based on regional geologic characterization through literature and analogous data; (3) feasibility analysis demonstrating calculation procedures and importance of system components, based on rough measures ofmore » performance using surface exploration, waste process knowledge, and general laboratory experiments; (4) suitability analysis demonstrating viability of disposal system, based on environment - specific laboratory experiments, in - situ experiments, and underground disposal system characterization; and (5) compliance analysis, based on completed site - specific characterization . The current sixth period beyond 2010 represents a new effort to set waste management policy in the United States. Because the relationship is important to understanding the evolution of the Yucca Mountain Project , the tabulation also shows the interaction between the policy realm and technical realm using four broad categories of events : (a) Regulatory requirements and related federal policy in laws and court decisions, (c) Presidential and agency directives, (c) technical milestones of implementing institutions, and (d) critiques of the Yucca Mountain Project and pertinent national and world events related to nuclear energy and radioactive waste. Preface The historical progression of technical milestones for the Yucca Mountain Project was originally developed for 10 journal articles in a special issue of Reliability Engineering System Safety on the performance assessment for the Yucca Mountain license application [1-10]. The listing of mile stones, a distinct feature of those articles, has been collected and tabulated here. A lthough a brief description is presented here (based on the summaries in the 10 journal articles), the emphasis remains on the tabulation because of its usefulness in providing a comprehensive but concise history of the Yucca Mountain Project. The tabulation presented here is more elaborate than originally presented in that many of the interactions that occurred between the technical realm and policy realm can be depicted in separate columns. The usefulness of the milestones table is due in part to L.A. Connolly, for editorial and reference support, and S.K. Best, Raytheon, and L. Mays, Sandia National Laboratories (SNL), for illustration support. Reviewers P.N. Swift, SNL, and K. Gupta, University of Oklahoma, helped improve the discussion. The historical perspective presented is that of the author and is not necessarily held by reviewers, Sandia National Laboratories , and the US Department of Energy. As a historic perspective, the author is reporting on the work of others; however, any interpretative error s of the documentation are those of the author alone. The characterization and modeling of the Yucca Mountain disposal system required numerous participants with expertise in many areas of science and technology, as evident from the extensive reference list. Their diligent efforts are generally acknowledged here and through the many references to their impressive work, but the 10 journal articles acknowledge by name many of the numerous participants that contributed to the Yucca Mountain Project .« less
Least-Squares Neutron Spectral Adjustment with STAYSL PNNL
NASA Astrophysics Data System (ADS)
Greenwood, L. R.; Johnson, C. D.
2016-02-01
The STAYSL PNNL computer code, a descendant of the STAY'SL code [1], performs neutron spectral adjustment of a starting neutron spectrum, applying a least squares method to determine adjustments based on saturated activation rates, neutron cross sections from evaluated nuclear data libraries, and all associated covariances. STAYSL PNNL is provided as part of a comprehensive suite of programs [2], where additional tools in the suite are used for assembling a set of nuclear data libraries and determining all required corrections to the measured data to determine saturated activation rates. Neutron cross section and covariance data are taken from the International Reactor Dosimetry File (IRDF-2002) [3], which was sponsored by the International Atomic Energy Agency (IAEA), though work is planned to update to data from the IAEA's International Reactor Dosimetry and Fusion File (IRDFF) [4]. The nuclear data and associated covariances are extracted from IRDF-2002 using the third-party NJOY99 computer code [5]. The NJpp translation code converts the extracted data into a library data array format suitable for use as input to STAYSL PNNL. The software suite also includes three utilities to calculate corrections to measured activation rates. Neutron self-shielding corrections are calculated as a function of neutron energy with the SHIELD code and are applied to the group cross sections prior to spectral adjustment, thus making the corrections independent of the neutron spectrum. The SigPhi Calculator is a Microsoft Excel spreadsheet used for calculating saturated activation rates from raw gamma activities by applying corrections for gamma self-absorption, neutron burn-up, and the irradiation history. Gamma self-absorption and neutron burn-up corrections are calculated (iteratively in the case of the burn-up) within the SigPhi Calculator spreadsheet. The irradiation history corrections are calculated using the BCF computer code and are inserted into the SigPhi Calculator workbook for use in correcting the measured activities. Output from the SigPhi Calculator is automatically produced, and consists of a portion of the STAYSL PNNL input file data that is required to run the spectral adjustment calculations. Within STAYSL PNNL, the least-squares process is performed in one step, without iteration, and provides rapid results on PC platforms. STAYSL PNNL creates multiple output files with tabulated results, data suitable for plotting, and data formatted for use in subsequent radiation damage calculations using the SPECTER computer code (which is not included in the STAYSL PNNL suite). All components of the software suite have undergone extensive testing and validation prior to release and test cases are provided with the package.
47 CFR 52.19 - Area code relief.
Code of Federal Regulations, 2013 CFR
2013-10-01
... Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) COMMON CARRIER SERVICES (CONTINUED) NUMBERING... commissions may perform any or all functions related to initiation and development of area code relief plans... paragraph (b)(2) of this section. For the purposes of this paragraph, initiation and development of area...
47 CFR 52.19 - Area code relief.
Code of Federal Regulations, 2012 CFR
2012-10-01
... Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) COMMON CARRIER SERVICES (CONTINUED) NUMBERING... commissions may perform any or all functions related to initiation and development of area code relief plans... paragraph (b)(2) of this section. For the purposes of this paragraph, initiation and development of area...
47 CFR 52.19 - Area code relief.
Code of Federal Regulations, 2014 CFR
2014-10-01
... Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) COMMON CARRIER SERVICES (CONTINUED) NUMBERING... commissions may perform any or all functions related to initiation and development of area code relief plans... paragraph (b)(2) of this section. For the purposes of this paragraph, initiation and development of area...
NASA Technical Reports Server (NTRS)
Graves, M. E.; King, R. L.; Brown, S. C.
1973-01-01
Extreme values, median values, and nine percentile values are tabulated for eight meteorological variables at Cape Kennedy, Florida and at Vandenberg Air Force Base, California. The variables are temperature, relative humidity, station pressure, water vapor pressure, water vapor mixing ratio, density, and enthalpy. For each month eight hours are tabulated, namely, 0100, 0400, 0700, 1000, 1300, 1600, 1900, and 2200 local time. These statistics are intended for general use for the space shuttle design trade-off analysis and are not to be used for specific design values.
Exact Doppler broadening of tabulated cross sections. [SIGMA 1 kernel broadening method
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cullen, D.E.; Weisbin, C.R.
1976-07-01
The SIGMA1 kernel broadening method is presented to Doppler broaden to any required accuracy a cross section that is described by a table of values and linear-linear interpolation in energy-cross section between tabulated values. The method is demonstrated to have no temperature or energy limitations and to be equally applicable to neutron or charged-particle cross sections. The method is qualitatively and quantitatively compared to contemporary approximate methods of Doppler broadening with particular emphasis on the effect of each approximation introduced.
VizieR Online Data Catalog: Radiative forces for stellar envelopes (Seaton, 1997)
NASA Astrophysics Data System (ADS)
Seaton, M. J.; Yan, Y.; Mihalas, D.; Pradhan, A. K.
2000-02-01
(1) Primary data files, stages.zz These files give data for the calculation of radiative accelerations, GRAD, for elements with nuclear charge zz. Data are available for zz=06, 07, 08, 10, 11, 12, 13, 14, 16, 18, 20, 24, 25, 26 and 28. Calculations are made using data from the Opacity Project (see papers SYMP and IXZ). The data are given for each ionisation stage, j. They are tabulated on a mesh of (T, Ne, CHI) where T is temperature, Ne electron density and CHI is abundance multiplier. The files include data for ionisation fractions, for each (T, Ne). The file contents are described in the paper ACC and as comments in the code add.f (2) Code add.f This reads a file stages.zz and creates a file acc.zz giving radiative accelerations averaged over ionisation stages. The code prompts for names of input and output files. The code, as provided, gives equal weights (as defined in the paper ACC) to all stages. Th weights are set in SUBROUTINE WEIGHTS, which could be changed to give any weights preferred by the user. The dependence of diffusion coefficients on ionisation stage is given by a function ZET, which is defined in SUBROUTINE ZETA. The expressions used for ZET are as given in the paper. The user can change that subroutine if other expressions are preferred. The output file contains values, ZETBAR, of ZET, averaged over ionisation stages. (3) Files acc.zz Radiative accelerations computed using add.f as provided. The user will need to run the code add.f only if it is required to change the subroutines WEIGHTS or ZETA. The contents of the files acc.zz are described in the paper ACC and in comments contained in the code add.f. (4) Code accfit.f This code gives gives radiative accelerations, and some related data, for a stellar model. Methods used to interpolate data to the values of (T, RHO) for the stellar model are based on those used in the code opfit.for (see the paper OPF). The executable file accfit.com runs accfit.f. It uses a list of files given in accfit.files (see that file for further description). The mesh used for the abundance-multiplier CHI on the output file will generally be finer than that used in the input files acc.zz. The mesh to be used is specified on a file chi.dat. For a test run, the stellar model used is given in the file 10000_4.2 (Teff=10000 K, LOG10(g)=4.2) The output file from that test run is acc100004.2. The contents of the output file are described in the paper ACC and as comments in the code accfit.f. (5) The code diff.f This code reads the output file (e.g. acc1000004.2) created by accfit.f. For any specified depth point in the model and value of CHI, it gives values of radiative accelerations, the quantity ZETBAR required for calculation of diffusion coefficients, and Rosseland-mean opacities. The code prompts for input data. It creates a file recording all data calculated. The code diff.f is intended for incorporation, as a set of subroutines, in codes for diffusion calculations. (1 data file).
Racial differences in the household composition of elders by age, gender, and area of residence.
Coward, R T; Lee, G R; Netzer, J K; Cutler, S J; Danigelis, N L
1996-01-01
Using census data and an innovative technique for describing the composition of households from the perspective of elders, this research provides a more detailed description of race differences in living arrangements of older persons than has previously existed. In addition, cross tabulations of race with other factors known to influence household composition (gender, age, and area of residence) are examined. While white women are more likely than black women to live alone, the reverse is true among men. Whites are more likely than blacks to live in married-couple-only households, and blacks are more likely than whites to live in multigenerational households, although these differences decrease with advancing age. Blacks are also more likely than whites to live with collateral kin or non-kin only, although these categories comprise small proportions of the population. Implications of these differences for caregiving and quality of life among older persons are discussed.
RSRA sixth scale wind tunnel test. Tabulated balance data, volume 2
NASA Technical Reports Server (NTRS)
Ruddell, A.; Flemming, R.
1974-01-01
Summaries are presented of all the force and moment data acquired during the RSRA Sixth Scale Wind Tunnel Test. These data include and supplement the data presented in curve form in previous reports. Each summary includes the model configuration, wing and empennage incidences and deflections, and recorded balance data. The first group of data in each summary presents the force and moment data in full scale parametric form, the dynamic pressure and velocity in the test section, and the powered nacelle fan speed. The second and third groups of data are the balance data in nondimensional coefficient form. The wind axis coefficient data corresponds to the parametric data divided by the wing area for forces and divided by the product of the wing area and wing span or mean aerodynamic chord for moments. The stability axis data resolves the wind axis data with respect to the angle of yaw.
A new species of Pentadinium from Eastern Anatolia, Turkey, Pentadinium galileoi
Sancay, R.H.; Bati, Z.; Edwards, L.E.; Ertug, K.I.
2007-01-01
The new gonyaulacoid dinoflagellate Pentadinium galileoi Sancay et al., sp. nov. from the Oligocene-Lower Miocene sediments of Eastern Anatolia has been identified. It is spherical, chordate with prominant discoidal cingulum and distally furcate apical, sulcal, and antapical processes. It has a type P(3???) archeopyle, and periarcheopyle is larger than endoarcheopyle. Tabulation is distinct and it has a formula of 1pr, 3-4',6???,6c,6???', 1p,1??????, Wall structure is granular in intraplate areas, and it has gonal spines at apex, posterior intercalary, and antapex. Processes at gonal positions may be on a variably developed antapical 'skirt' formed by extended septa (2-3 ??m), minute bifurcate spines intergonally and along cingulum at plate intersections. Periphragm and endophragm appressed except at cingular area and at septa. The size of the cysts ranges between 73-89??m with an avarage of 84??m.
Long-term Exposure to PM2.5 and Mortality Among Older Adults in the Southeastern US.
Wang, Yan; Shi, Liuhua; Lee, Mihye; Liu, Pengfei; Di, Qian; Zanobetti, Antonella; Schwartz, Joel D
2017-03-01
Little is known about what factors modify the effect of long-term exposure to PM2.5 on mortality, in part because in most previous studies certain groups such as rural residents and individuals with lower socioeconomic status (SES) are under-represented. We studied 13.1 million Medicare beneficiaries (age ≥65) residing in seven southeastern US states during 2000-2013 with 95 million person-years of follow-up. We predicted annual average of PM2.5 in each zip code tabulation area (ZCTA) using a hybrid spatiotemporal model. We fit Cox proportional hazards models to estimate the association between long-term PM2.5 and mortality. We tested effect modification by individual-level covariates (race, sex, eligibility for both Medicare and Medicaid, and medical history), neighborhood-level covariates (urbanicity, percentage below poverty level, lower education, median income, and median home value), mean summer temperature, and mass fraction of 11 PM2.5 components. The hazard ratio (HR) for death was 1.021 (95% confidence interval: 1.019, 1.022) per 1 μg m increase in annual PM2.5. The HR decreased with age. It was higher among males, non-whites, dual-eligible individuals, and beneficiaries with previous hospital admissions. It was higher in neighborhoods with lower SES or higher urbanicity. The HR increased with mean summer temperature. The risk associated with PM2.5 increased with relative concentration of elemental carbon, vanadium, copper, calcium, and iron and decreased with nitrate, organic carbon, and sulfate. Associations between long-term PM2.5 exposure and death were modified by individual-level, neighborhood-level variables, temperature, and chemical compositions.
The development and implementation of a hospital safety and sanitation task force.
Weinstein, S A; Whelan, M
1985-01-01
We have found that the above activities are facilitated by utilizing a computerized filing program. The computer formal provides quick, concise reports and statistical data analysis. The formation catalogued can be easily analysis. The information catalogued can be easily categorized and analyzed. The time invested in organizational meetings concerning committee membership, guideline review and creating an efficient incident coding scheme was essential to the success of the project. Problems seem to arise when several different departments share a common space for different activities. For example, medication room sanitation throughout the hospital was suboptimal probably related to the fact that Nursing, Pharmacy and General Stores all utilized the area. No individual department felt it was their sole responsibility to ensure appropriate sanitation. Through the efforts of the task force, we were able to facilitate cooperation among the departments with each doing their share. As a result, medication rooms are now cleaner, neater and therefore safer. Additional incidents have been corrected since the original tabulation of this report. A priority goal is to facilitate corrective action in a more timely manner. It is critical that surveys should not be intended to be "white glove inspections" with criticisms placed on specific departments. This can be accomplished by informing department managers of the intent of the surveys and by allowing staff to participate in all activities. In addition, the support and cooperation of administration is a major factor in the potential success of a program of this nature. As a result of the interdisciplinary team approach, the general safety environment for patients, visitors, employees and volunteers has significantly been improved at our institution.
Tomer, M D; Boomer, K M B; Porter, S A; Gelder, B K; James, D E; McLellan, E
2015-05-01
A watershed's riparian corridor presents opportunities to stabilize streambanks, intercept runoff, and influence shallow groundwater with riparian buffers. This paper presents a system to classify these riparian opportunities and apply them toward riparian management planning in hydrologic unit code 12 watersheds. In two headwater watersheds from each of three landform regions found in Iowa and Illinois, high-resolution (3-m grid) digital elevation models were analyzed to identify spatial distributions of surface runoff contributions and zones with shallow water tables (SWTs) (within 1.5 m of the channel elevation) along the riparian corridors. Results were tabulated, and a cross classification was applied. Classes of buffers include those primarily placed to (i) trap runoff and sediment, (ii) influence shallow groundwater, (iii) address both runoff and shallow groundwater, and (iv) maintain/improve stream bank stability. Riparian buffers occupying about 2.5% of these six watersheds could effectively intercept runoff contributions from 81 to 94% of the watersheds' contributing areas. However, extents of riparian zones where a narrow buffer (<10 m wide) would adequately intercept runoff but where >25 m width of buffer vegetation could root to a SWT varied according to landform region ( < 0.10). Yet, these wide-SWT riparian zones were widespread and occupied 23 to 53% of the lengths of stream banks among the six watersheds. The wide-SWT setting provides opportunities to reduce dissolved nutrients (particularly NO-N) carried via groundwater. This riparian classification and mapping system is part of a ArcGIS toolbox and could provide a consistent basis to identify riparian management opportunities in Midwestern headwater catchments wherever high-resolution elevation data are available. Copyright © by the American Society of Agronomy, Crop Science Society of America, and Soil Science Society of America, Inc.
Bagcivan, Gulcan; Dionne-Odom, J Nicholas; Frost, Jennifer; Plunkett, Margaret; Stephens, Lisa A; Bishop, Peggy; Taylor, Richard A; Li, Zhongze; Tucker, Rodney; Bakitas, Marie
2018-01-01
Early outpatient palliative care consultations are recommended by clinical oncology guidelines globally. Despite these recommendations, it is unclear which components should be included in these encounters. Describe the evaluation and treatment recommendations made in early outpatient palliative care consultations. Outpatient palliative care consultation chart notes were qualitatively coded and frequencies tabulated. Outpatient palliative care consultations were automatically triggered as part of an early versus delayed randomized controlled trial (November 2010 to April 2013) for patients newly diagnosed with advanced cancer living in the rural Northeastern US. In all, 142 patients (early = 70; delayed = 72) had outpatient palliative care consultations. The top areas addressed in these consultations were general evaluations-marital/partner status (81.7%), spirituality/emotional well-being (80.3%), and caregiver/family support (79.6%); symptoms-mood (81.7%), pain (73.9%), and cognitive/mental status (68.3%); general treatment recommendations-counseling (39.4%), maintaining current medications (34.5%), and initiating new medication (23.9%); and symptom-specific treatment recommendations-pain (22.5%), constipation (12.7%), depression (12.0%), advanced directive completion (43.0%), identifying a surrogate (21.8%), and discussing illness trajectory (21.1%). Compared to the early group, providers were more likely to evaluate general pain ( p = 0.035) and hospice awareness ( p = 0.005) and discuss/recommend hospice ( p = 0.002) in delayed group participants. Outpatient palliative care consultations for newly diagnosed advanced cancer patients can address patients' needs and provide recommendations on issues that might not otherwise be addressed early in the disease course. Future prospective studies should ascertain the value of early outpatient palliative care consultations that are automatically triggered based on diagnosis or documented symptom indicators versus reliance on oncologist referral.
Crowson, Matthew G; Schulz, Kristine; Parham, Kourosh; Vambutas, Andrea; Witsell, David; Lee, Walter T; Shin, Jennifer J; Pynnonen, Melissa A; Nguyen-Huynh, Anh; Ryan, Sheila E; Langman, Alan
2016-07-01
(1) Integrate practice-based patient encounters using the Dartmouth Atlas Medicare database to understand practice treatments for Ménière's disease (MD). (2) Describe differences in the practice patterns between academic and community providers for MD. Practice-based research database review. CHEER (Creating Healthcare Excellence through Education and Research) network academic and community providers. MD patient data were identified with ICD-9 and CPT codes. Demographics, unique visits, and procedures per patient were tabulated. The Dartmouth Atlas of Health Care was used to reference regional health care utilization. Statistical analysis included 1-way analyses of variance, bivariate linear regression, and Student's t tests, with significance set at P < .05. A total of 2071 unique patients with MD were identified from 8 academic and 10 community otolaryngology-head and neck surgery provider centers nationally. Average age was 56.5 years; 63.9% were female; and 91.4% self-reported white ethnicity. There was an average of 3.2 visits per patient. Western providers had the highest average visits per patient. Midwest providers had the highest average procedures per patient. Community providers had more visits per site and per patient than did academic providers. Academic providers had significantly more operative procedures per site (P = .0002) when compared with community providers. Health care service areas with higher total Medicare reimbursements per enrollee did not report significantly more operative procedures being performed. This is the first practice-based clinical research database study to describe MD practice patterns. We demonstrate that academic otolaryngology-head and neck surgery providers perform significantly more operative procedures than do community providers for MD, and we validate these data with an independent Medicare spending database. © American Academy of Otolaryngology—Head and Neck Surgery Foundation 2015.
Population-based assessment of heartburn in urban Black Americans.
Friedenberg, F K; Makipour, K; Palit, A; Shah, S; Vanar, V; Richter, J E
2013-08-01
Prevalence data for heartburn in the urban Black American community is lacking. In order to estimate prevalence for this community, we analyzed data from an ongoing cohort study in progress at our hospital. Comprehensive interviews allowed for exploration of factors associated with heartburn. Complex, stratified sampling design was the method used. Survey invitations are hand-delivered to random blocks in a single zip code tabulation area. One member per eligible household is invited to complete a computer-based survey. Heartburn was defined as ≥ 3 days/week of symptoms as defined by the Montreal Definition and Classification of gastroesophageal reflux disease. Scaling and weighting factors were utilized to estimate population level prevalence. Multivariate logistic regression was used to identify independent predictor variables for heartburn. Enrolled 379 participants corresponding to a weighted sample size of 22,409 (20,888-23,930) citizens. Demographic characteristics of the sample closely matched those of the entire targeted population. Overall, the weighted prevalence of heartburn ≥ 3 times per week was 17.6% (16.4-18.8%). Variables independently associated with heartburn were body mass index, daily caloric and fat intake, diabetes mellitus (odds ratio = 2.95; 2.59-3.36), cigarette smoking, and alcohol consumption (odds ratio = 2.55; 2.25-2.89). Factors inversely associated included illicit drug use and increased physical activity. Waist : hip ratio showed no relationship. The prevalence of heartburn ≥ 3 times per week is high in the Black American community. Adverse lifestyle behaviors showed particularly important associations. Our study needs to be replicated in other communities with similar demographics. © 2012 Wiley Periodicals, Inc. and the International Society for Diseases of the Esophagus.
Wang, Frederick; Koltz, Peter F; Sbitany, Hani
2014-11-01
The American College of Surgeons National Surgical Quality Improvement Program database was implemented to longitudinally track surgical 30-day surgical outcomes and complications. The authors analyze the program-reported outcomes for immediate breast reconstruction from 2007 to 2011, to assess whether longitudinal data collection has improved national outcomes and to highlight areas in need of continued improvement. The authors reviewed the database from 2007 to 2011 and identified encounters for immediate breast reconstruction using Current Procedural Terminology codes for prosthetic and autologous reconstruction. Demographics and comorbidities were tabulated for all patients. Postoperative complications analyzed included surgical-site infection, wound dehiscence, implant or flap loss, pulmonary embolism, and respiratory infections. A total of 15,978 patients underwent mastectomy and immediate reconstruction. Fewer smokers underwent immediate reconstruction over time (p=0.126), whereas more obese patients (p=0.001) and American Society of Anesthesiologists class 3 and 4 patients (p<0.001) underwent surgery. An overall increase in superficial surgical-site infection was noted, from 1.7 percent to 2.3 percent (p=0.214). Wound dehiscence (p=0.036) increased over time, whereas implant loss (p=0.015) and flap loss (p=0.012) decreased over time. Mean operative times increased over the analyzed years, as did all complications for prosthetic and autologous reconstruction. The American College of Surgeons National Surgical Quality Improvement Program data set has shown an increase in complications for immediate breast reconstruction over time, because of a longitudinally higher number of comorbid patients and longer operative times. This knowledge allows plastic surgeons the unique opportunity to improve patient selection criteria and efficiency. Therapeutic, III.
Santos, José; Monteagudo, Ángel
2017-03-27
The canonical code, although prevailing in complex genomes, is not universal. It was shown the canonical genetic code superior robustness compared to random codes, but it is not clearly determined how it evolved towards its current form. The error minimization theory considers the minimization of point mutation adverse effect as the main selection factor in the evolution of the code. We have used simulated evolution in a computer to search for optimized codes, which helps to obtain information about the optimization level of the canonical code in its evolution. A genetic algorithm searches for efficient codes in a fitness landscape that corresponds with the adaptability of possible hypothetical genetic codes. The lower the effects of errors or mutations in the codon bases of a hypothetical code, the more efficient or optimal is that code. The inclusion of the fitness sharing technique in the evolutionary algorithm allows the extent to which the canonical genetic code is in an area corresponding to a deep local minimum to be easily determined, even in the high dimensional spaces considered. The analyses show that the canonical code is not in a deep local minimum and that the fitness landscape is not a multimodal fitness landscape with deep and separated peaks. Moreover, the canonical code is clearly far away from the areas of higher fitness in the landscape. Given the non-presence of deep local minima in the landscape, although the code could evolve and different forces could shape its structure, the fitness landscape nature considered in the error minimization theory does not explain why the canonical code ended its evolution in a location which is not an area of a localized deep minimum of the huge fitness landscape.
Federal Register 2010, 2011, 2012, 2013, 2014
2010-04-16
... for Residential Construction in High Wind Areas. ICC 700: National Green Building Standard. The... Codes and Standards that are comprehensive, coordinated, and necessary to regulate the built environment... International Codes and Standards consist of the following: ICC Codes International Building Code. International...
Monrose, Erica; Ledergerber, Jessica; Acheampong, Derrick; Jandorf, Lina
2017-09-21
To assess participants' reasons for seeking cancer screening information at community health fairs and what they do with the information they receive. Mixed quantitative and qualitative approach was used. Community health fairs are organized in underserved New York City neighbourhoods. From June 14, 2016 to August 26, 2016, cancer prevention tables providing information about various cancer screenings were established at 12 local community health fairs in New York City. In-person and follow up telephone surveys assessing interest in the cancer prevention table, personal cancer screening adherence rates, information-sharing behaviours and demographic variables have been taken into account. Statistical analyses were performed using IBM SPSS 22.0: frequencies, descriptive, cross tabulations. All qualitative data was coded by theme so that it could be analysed through SPSS. For example, Were you interested in a specific cancer? may be coded as 2 for yes , breast cancer . One hundred and sixteen patrons participated in the initial survey. Of those, 88 (78%) agreed to give their contact information for the follow-up survey and 60 follow-up surveys were completed (68%). Of those who reported reading the material, 45% shared the information; 15% subsequently spoke to a provider about cancer screenings and 40% intended to speak to a provider. Participants disseminated information without prompting; suggesting the reach of these fairs extends beyond the people who visit our table. Future studies should look at whether patrons would share information at higher rates when they are explicitly encouraged to share the information.
Van Wave, Timothy W; Decker, Michael
2003-04-01
Development of a method using marketing research data to assess food purchase behavior and consequent nutrient availability for purposes of nutrition surveillance, evaluation of intervention effects, and epidemiologic studies of diet-health relationships. Data collected on household food purchases accrued over a 13-week period were selected by using Universal Product Code numbers and household characteristics from a marketing research database. Universal Product Code numbers for 39,408 dairy product purchases were linked to a standard reference for food composition to estimate the nutrient content of foods purchased over time. Two thousand one hundred sixty-one households located in Victoria, Texas, and surrounding communities who were active members of a frequent shopper program. Demographic characteristics of sample households and the nutrient content of their dairy product purchases were analyzed using frequency distribution, cross tabulation, analysis of variance, and t test procedures. A method for using marketing research data was successfully used to estimate household purchases of specific foods and their nutrient content from a marketing database containing hundreds of thousands of records. Distribution of dairy product purchases and their concomitant nutrients between Hispanic and non-Hispanic households were significant (P<.01, P<.001, respectively) and sustained over time. Purchase records from large, nationally representative panels of shoppers, such as those maintained by major market research companies, might be used to accomplish detailed longitudinal epidemiologic studies or surveillance of national food- and nutrient-purchasing patterns within and between countries and segments of their respective populations.
[Suicide trends in Colombia, 1985-2002].
Cendales, Ricardo; Vanegas, Claudia; Fierro, Marco; Córdoba, Rodrigo; Olarte, Ana
2007-10-01
To report trends in mortality from suicide in Colombia from 1985 to 2002 by sex, age group, and method, and determine the number of Years of Potential Life Lost (YPLL) to suicide during this period. Age- and sex-specific and age-adjusted crude mortality rates were calculated based on mortality and population information available from the official database of the Department of National Statistics Administration, Colombia. YPLL were estimated and adjusted for societal impact, age, and poor quality of mortality records. The results were tabulated according to codes X600-X849 and Y870 from the International Statistical Classification of Disease and Related Health Problems, 10th revision (ICD-10), and codes E950-E959 from the 9th revision (ICD-9). Suicide rates have been climbing in Colombia since 1998, particularly among young adults and males. The highest rates among males were in the age groups 20-29 years of age and over 70 years of age, and rates increased over time. Among females, the highest rates were recorded for the group 10-19 years of age. The YPLL rose in proportion with the increase in suicides, from 0.81% in 1981 to 2.20% in 2002. Among males, the most common methods used were firearms and explosives, hanging, and poison, with a relative increase in hanging; whereas among females, poison was most common. A rising trend in suicide rates in Colombia was confirmed, especially among the productive segment of the population, which has resulted in a marked increase in YPLL.
1990 National Water Quality Laboratory Services Catalog
Pritt, Jeffrey; Jones, Berwyn E.
1989-01-01
PREFACE This catalog provides information about analytical services available from the National Water Quality Laboratory (NWQL) to support programs of the Water Resources Division of the U.S. Geological Survey. To assist personnel in the selection of analytical services, the catalog lists cost, sample volume, applicable concentration range, detection level, precision of analysis, and preservation techniques for samples to be submitted for analysis. Prices for services reflect operationa1 costs, the complexity of each analytical procedure, and the costs to ensure analytical quality control. The catalog consists of five parts. Part 1 is a glossary of terminology; Part 2 lists the bottles, containers, solutions, and other materials that are available through the NWQL; Part 3 describes the field processing of samples to be submitted for analysis; Part 4 describes analytical services that are available; and Part 5 contains indices of analytical methodology and Chemical Abstract Services (CAS) numbers. Nomenclature used in the catalog is consistent with WATSTORE and STORET. The user is provided with laboratory codes and schedules that consist of groupings of parameters which are measured together in the NWQL. In cases where more than one analytical range is offered for a single element or compound, different laboratory codes are given. Book 5 of the series 'Techniques of Water Resources Investigations of the U.S. Geological Survey' should be consulted for more information about the analytical procedures included in the tabulations. This catalog supersedes U.S. Geological Survey Open-File Report 86-232 '1986-87-88 National Water Quality Laboratory Services Catalog', October 1985.
Water resources of the Cook Inlet Basin, Alaska
Freethey, Geoffrey W.; Scully, David R.
1980-01-01
Ground-water and surface-water systems of Cook Inlet basin, Alaska, are analyzed. Geologic and topographic features that control the movement and regional availability of ground water are explained and illustrated. Five aquifer systems beneath the most populous areas are described. Estimates of ground-water yield were determined for the region by using ground-water data for the populated areas and by extrapolating known subsurface conditions and interpreting subsurface conditions from surficial features in the other areas. Area maps of generalized geology, Quaternary sediment thickness, and general availability of ground water are shown. Surface-water resources are summarized by describing how basin characteristics affect the discharge in streams. Seasonal trend of streamflow for three types of streams is described. Regression equations for 4 streamflow characteristics (annual, monthly minimum, and maximum discharge) were obtained by using gaging station streamflow characteristics and 10 basin characteristics. In the 24 regression equations presented, drainage area is the most significant basin characteristic, but 5 others are used. Maps of mean annual unit runoff and minimum unit yield for 7 consecutive days with a recurrence interval of 10 years are shown. Historic discharge data at gaging stations is tabulated and representative low-flow and flood-flow frequency curves are shown. (USGS)
Development and application of structural dynamics analysis capabilities
NASA Technical Reports Server (NTRS)
Heinemann, Klaus W.; Hozaki, Shig
1994-01-01
Extensive research activities were performed in the area of multidisciplinary modeling and simulation of aerospace vehicles that are relevant to NASA Dryden Flight Research Facility. The efforts involved theoretical development, computer coding, and debugging of the STARS code. New solution procedures were developed in such areas as structures, CFD, and graphics, among others. Furthermore, systems-oriented codes were developed for rendering the code truly multidisciplinary and rather automated in nature. Also, work was performed in pre- and post-processing of engineering analysis data.
Channel fading for mobile satellite communications using spread spectrum signaling and TDRSS
NASA Technical Reports Server (NTRS)
Jenkins, Jeffrey D.; Fan, Yiping; Osborne, William P.
1995-01-01
This paper will present some preliminary results from a propagation experiment which employed NASA's TDRSS and an 8 MHz chip rate spread spectrum signal. Channel fade statistics were measured and analyzed in 21 representative geographical locations covering urban/suburban, open plain, and forested areas. Cumulative distribution Functions (CDF's) of 12 individual locations are presented and classified based on location. Representative CDF's from each of these three types of terrain are summarized. These results are discussed, and the fade depths exceeded 10 percent of the time in three types of environments are tabulated. The spread spectrum fade statistics for tree-lined roads are compared with the Empirical Roadside Shadowing Model.
Sullivan, Shannon M; Tsiplova, Kate; Ungar, Wendy J
2016-10-01
Economic evaluations conducted in children have unique features compared to adults. Important developments in pediatric economic evaluation in recent years include new options for valuing health states for cost-utility analysis (CUA) and shifting child health priorities. The Pediatric Economic Database Evaluation (PEDE) project includes a comprehensive database of pediatric health economic evaluations published since 1980. The objective of this scoping review was to identify trends over time in the use of CUA and other analytic techniques, and the therapeutic areas chosen for study. Areas covered: Medical and grey literature were searched, key characteristics were extracted, frequencies were tabulated and cross-tabulations were performed. Differences between early (1980 and 1999) and late (2000 and 2014) periods were assessed using a chi-squared statistic. Of the 2,776 pediatric economic evaluations published between 1980 and 2014, substantially more were cost-effectiveness analyses (CEAs) and CUAs than cost benefit analyses and cost minimization analyses (63.9 and 24.9% versus 7.6 and 3.6%, respectively). This pattern was consistent regardless of the type of intervention, disease or age group studied. A trend toward higher proportions of CUAs and CEAs was evident in the later period (X 2 p < 0.0001). Other significant trends included a higher proportion of studies of preventive interventions (X 2 p < 0.0001), and more studies in children aged 1 to 12 years and fewer in perinates in the later period (X 2 p < 0.0001). Overall the most common disease class studied was infectious diseases (29.2%). Expert commentary: Pediatric economic evaluation continues to grow in volume and methodologic complexity. While CUAs have increased, whether their quality has improved remains unknown. Although most studies are in infectious disease, the volume of publications may not align with emerging child health priorities such as adolescent health, injury, developmental disabilities, mental health, and the use of personalized medicine. Increasing economic evaluations in these areas will enhance pediatric decision-making.
,
1988-01-01
This standard provides codes to be used for the identification of aquifer names and geologic units in the United States, the Caribbean and other outlying areas. Outlying areas include Puerto Rico, the Virgin Islands, American Samoa, the Midway Islands, Trust Territories of the Pacific Islands, and miscellaneous Pacific Islands. Each code identifies an aquifer or rock-stratigraphic unit and its age designation. The codes provide a standardized base for use by organizations in the storage, retrieval, and exchange of ground-water data; the indexing and inventory of ground-water data and information; the cataloging of ground-water data acquisition activities; and a variety of other applications.
,
1985-01-01
This standard provides codes to be used for the identification of aquifer names and geologic units in the United States, the Caribbean and other outlying areas. Outlying areas include Puerto Rico, the Virgin Islands, American Samoa, the Midway Islands, Trust Territories of the Pacific Islands, and miscellaneous Pacific Islands. Each code identifies an aquifer or rock-stratigraphic unit and its age designation. The codes provide a standardized base for use by organizations in the storage, retrieval, and exchange of ground-water data; the indexing and inventory of ground-water data and information; the cataloging of ground-water data acquisition activities; and a variety of other applications.
Gray, Kathleen; Martin-Sanchez, Fernando
2014-01-01
Background Research continues to present tenuous suggestions that social media is well suited to enhance management of chronic disease and improve health outcomes. Various studies have presented qualitative reports of health outcomes from social media use and have examined discourse and communication themes occurring through different social media. However, there is an absence of published studies examining and unpacking the underlying therapeutic mechanisms driving social media’s effects. Objective This paper presents a qualitative analysis thoroughly describing what social media therapeutically affords people living with chronic pain who are self-managing their condition. From this therapeutic affordance perspective, we aim to formulate a preliminary conceptual model aimed at better understanding "how" social media can influence patient outcomes. Methods In total, 218 people with chronic pain (PWCP) completed an online survey, investigating patient-reported outcomes (PROs) from social media use. Supplementary to quantitative data collected, participants were also given the opportunity to provide further open commentary regarding their use of social media as part of chronic pain management; 68/218 unique users (31.2%) chose to provide these free-text responses. Through thematic content analysis, 117 free-text responses regarding 10 types of social media were coded. Quotes were extracted and tabulated based on therapeutic affordances that we had previously identified. Inductive analysis was then performed to code defining language and emergent themes central to describing each affordance. Three investigators examined the responses, developed the coding scheme, and applied the coding to the data. Results We extracted 155 quotes from 117 free-text responses. The largest source of quotes came from social network site users (78/155, 50.3%). Analysis of component language used to describe the aforementioned affordances and emergent themes resulted in a final revision and renaming of therapeutic affordances: "exploration" (52/155, 33.5% of quotes), "connection" (50/155, 32.3% of quotes), "narration" (33/155, 21.3% of quotes), "adaptation" (13/155, 8.4% of quotes), and "self-presentation" (7/155, 4.5% of quotes). Of the most described affordances, "exploration" was based on a propensity for participants to explain their social media use for information seeking purposes. "Connection" placed greater emphasis on interaction, highlighting themes of "exchanging information" and "mitigating isolation". Responses regarding "narration" highlighted the value of shared experiences and the emotionally cathartic role this plays. Conclusions Much of the efficacy of social media may be explicable via a closer examination of therapeutic affordances. Particular areas that warrant attention include social media’s ability to filter and guide people to useful information, connect individuals, and share experiences. Further research into a variety of chronic conditions is warranted. Coupled with the results of the present study, a greater theoretical basis detailing how social media may foster health outcomes may lead to an improved evidence base for conducting research and may inform recommendations for social media use in chronic disease management. PMID:25533453
Merolli, Mark; Gray, Kathleen; Martin-Sanchez, Fernando
2014-12-22
Research continues to present tenuous suggestions that social media is well suited to enhance management of chronic disease and improve health outcomes. Various studies have presented qualitative reports of health outcomes from social media use and have examined discourse and communication themes occurring through different social media. However, there is an absence of published studies examining and unpacking the underlying therapeutic mechanisms driving social media's effects. This paper presents a qualitative analysis thoroughly describing what social media therapeutically affords people living with chronic pain who are self-managing their condition. From this therapeutic affordance perspective, we aim to formulate a preliminary conceptual model aimed at better understanding "how" social media can influence patient outcomes. In total, 218 people with chronic pain (PWCP) completed an online survey, investigating patient-reported outcomes (PROs) from social media use. Supplementary to quantitative data collected, participants were also given the opportunity to provide further open commentary regarding their use of social media as part of chronic pain management; 68/218 unique users (31.2%) chose to provide these free-text responses. Through thematic content analysis, 117 free-text responses regarding 10 types of social media were coded. Quotes were extracted and tabulated based on therapeutic affordances that we had previously identified. Inductive analysis was then performed to code defining language and emergent themes central to describing each affordance. Three investigators examined the responses, developed the coding scheme, and applied the coding to the data. We extracted 155 quotes from 117 free-text responses. The largest source of quotes came from social network site users (78/155, 50.3%). Analysis of component language used to describe the aforementioned affordances and emergent themes resulted in a final revision and renaming of therapeutic affordances: "exploration" (52/155, 33.5% of quotes), "connection" (50/155, 32.3% of quotes), "narration" (33/155, 21.3% of quotes), "adaptation" (13/155, 8.4% of quotes), and "self-presentation" (7/155, 4.5% of quotes). Of the most described affordances, "exploration" was based on a propensity for participants to explain their social media use for information seeking purposes. "Connection" placed greater emphasis on interaction, highlighting themes of "exchanging information" and "mitigating isolation". Responses regarding "narration" highlighted the value of shared experiences and the emotionally cathartic role this plays. Much of the efficacy of social media may be explicable via a closer examination of therapeutic affordances. Particular areas that warrant attention include social media's ability to filter and guide people to useful information, connect individuals, and share experiences. Further research into a variety of chronic conditions is warranted. Coupled with the results of the present study, a greater theoretical basis detailing how social media may foster health outcomes may lead to an improved evidence base for conducting research and may inform recommendations for social media use in chronic disease management.
Heymann, R; Weitmann, K; Weiss, S; Thierfelder, D; Flessa, S; Hoffmann, W
2009-07-01
This study examines and compares the frequency of home visits by general practitioners in regions with a lower population density and regions with a higher population density. The discussion centres on the hypothesis whether the number of home visits in rural and remote areas with a low population density is, in fact, higher than in urbanised areas with a higher population density. The average age of the population has been considered in both cases. The communities of Mecklenburg West-Pomerania were aggregated into postal code regions. The analysis is based on these postal code regions. The average frequency of home visits per 100 inhabitants/km2 has been calculated via a bivariate, linear regression model with the population density and the average age for the postal code region as independent variables. The results are based on billing data of the year 2006 as provided by the Association of Statutory Health Insurance Physicians of Mecklenburg-Western Pomerania. In a second step a variable which clustered the postal codes of urbanised areas was added to a multivariate model. The hypothesis of a negative correlation between the frequency of home visits and the population density of the areas examined cannot be confirmed for Mecklenburg-Western Pomerania. Following the dichotomisation of the postal code regions into sparsely and densely populated areas, only the very sparsely populated postal code regions (less than 100 inhabitants/km2) show a tendency towards a higher frequency of home visits. Overall, the frequency of home visits in sparsely populated postal code regions is 28.9% higher than in the densely populated postal code regions (more than 100 inhabitants/km2), although the number of general practitioners is approximately the same in both groups. In part this association seems to be confirmed by a positive correlation between the average age in the individual postal code regions and the number of home visits carried out in the area. As calculated on the basis of the data at hand, only the very sparsely populated areas with a still gradually decreasing population show a tendency towards a higher frequency of home visits. According to the data of 2006, the number of home visits remains high in sparsely populated areas. It may increase in the near future as the number of general practitioners in these areas will gradually decrease while the number of immobile and older inhabitants will increase.
NASA's AVE 7 experiment: 25-mb sounding data
NASA Technical Reports Server (NTRS)
Davis, J. G.; Fuelberg, H. E.; Turner, R. E.
1978-01-01
The AVE 7 Experiment is described and tabulated rawinsonde data at 25 mb internals from the surface to 25 mb for the 24 stations participating in the experiment are presented. Soundings were taken between 0000GMT May 2 and 1200 GMT May 3, 1978. The methods of data processing and the accuracy are briefly discussed. Selected synoptic charts prepared from the data are presented as well as an example of contact data. A tabulation of adverse weather events that occured during the AVE 7 period, including freezing temperature, snow, tornadoes, damaging winds, and flooding, is presented.
Program to Produce Tabulated Data Set Describing NSWC Burn Model for Hydrodynamic Computations
1990-09-11
helpful insights of Dr. Raafat Guirguis of the Naval Surface Warfare Center on how the NSWC Burn Model works, and Drs. Schittke and Feisler of...R. Guirguis ) 1 R13 (P. Miller ) 1 R13 (K. Kin) 2 R13 (C. Coffey) 1 R13 (H. Sandusky) 1 R13 (D. Tasker) 1 R13 (E. Lanar) 1 R13 (J. Forbes) 1 R13 (R...NAVSWC TR 90-364 AD-A238 710 PROGRAM TO PRODUCE TABULATED DATA SET DESCRIBING NSWC BURN MODEL FOR HYDRODYNAMIC COMPUTATIONS BY LEWIS C. HUDSON III
Radiometric performance of the Voyager cameras
NASA Technical Reports Server (NTRS)
Danielson, G. E.; Kupferman, P. N.; Johnson, T. V.; Soderblom, L. A.
1981-01-01
The Voyager Imaging Experiment provided high-quality data of Jupiter and the Galilean satellites with the two flyby trajectories in March and July of 1979. Moderately accurate radiometric measurements have been made using these data. This paper evaluates the radiometric results and describes the inflight and ground geometric and radiometric correction factors. The radiometric quantities of intensity I and geometric albedo I/F are derived, and scaling factors for each of the filters are tabulated for correcting the 'calibrated' data from the Image Processing Laboratory at JPL. In addition, the key characteristics of both Voyager I and Voyager 2 cameras are tabulated.
Digital computer program for nuclear reactor design water properties (LWBR Development Program)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lynn, L.L.
1967-07-01
An edit program MO899 for the tabulation of thermodynamic and transport properties of liquid and vapor water, frequently used in design calculations for pressurized water nuclear reactors, is described. The data tabulated are obtained from a FORTRAN IV subroutine named HOH. Values of enthalpy, specific volume, viscosity, and thermal conductivity are given for the following ranges: pressure from one bar (14.5 psia) to 175 bars (2538 psia) and temperature from as much as 320 deg C (608 deg F) below saturation up to 500 deg C (932 deg F) above saturation. (NSA 21: 38472)
Evaluation of the flame propagation within an SI engine using flame imaging and LES
NASA Astrophysics Data System (ADS)
He, Chao; Kuenne, Guido; Yildar, Esra; van Oijen, Jeroen; di Mare, Francesca; Sadiki, Amsini; Ding, Carl-Philipp; Baum, Elias; Peterson, Brian; Böhm, Benjamin; Janicka, Johannes
2017-11-01
This work shows experiments and simulations of the fired operation of a spark ignition engine with port-fuelled injection. The test rig considered is an optically accessible single cylinder engine specifically designed at TU Darmstadt for the detailed investigation of in-cylinder processes and model validation. The engine was operated under lean conditions using iso-octane as a substitute for gasoline. Experiments have been conducted to provide a sound database of the combustion process. A planar flame imaging technique has been applied within the swirl- and tumble-planes to provide statistical information on the combustion process to complement a pressure-based comparison between simulation and experiments. This data is then analysed and used to assess the large eddy simulation performed within this work. For the simulation, the engine code KIVA has been extended by the dynamically thickened flame model combined with chemistry reduction by means of pressure dependent tabulation. Sixty cycles have been simulated to perform a statistical evaluation. Based on a detailed comparison with the experimental data, a systematic study has been conducted to obtain insight into the most crucial modelling uncertainties.
Standard Model thermodynamics across the electroweak crossover
DOE Office of Scientific and Technical Information (OSTI.GOV)
Laine, M.; Meyer, M., E-mail: laine@itp.unibe.ch, E-mail: meyer@itp.unibe.ch
Even though the Standard Model with a Higgs mass m{sub H} = 125GeV possesses no bulk phase transition, its thermodynamics still experiences a 'soft point' at temperatures around T = 160GeV, with a deviation from ideal gas thermodynamics. Such a deviation may have an effect on precision computations of weakly interacting dark matter relic abundances if their mass is in the few TeV range, or on leptogenesis scenarios operating in this temperature range. By making use of results from lattice simulations based on a dimensionally reduced effective field theory, we estimate the relevant thermodynamic functions across the crossover. The resultsmore » are tabulated in a numerical form permitting for their insertion as a background equation of state into cosmological particle production/decoupling codes. We find that Higgs dynamics induces a non-trivial 'structure' visible e.g. in the heat capacity, but that in general the largest radiative corrections originate from QCD effects, reducing the energy density by a couple of percent from the free value even at T > 160GeV.« less
VizieR Online Data Catalog: Granulation model for 508 KIC stars (Cranmer+, 2014)
NASA Astrophysics Data System (ADS)
Cranmer, S. R.; Bastien, F. A.; Stassun, K. G.; Saar, S. H.
2016-01-01
A goal of this work is to find self-consistent and accurate ways to predict the properties of stellar light-curve variability, and to use this variability to calibrate against other methods of determining their fundamental parameters. Thus, it may be possible to develop the analysis of granular flicker measurements in a way that augments the results of asteroseismology and improves the accuracy of, e.g., stellar mass and radius measurements. To assist in this process, we provide tabulated data for 508 stars with photometric light curves measured by the Kepler mission, which also includes their derived masses and predicted values of the turbulent Mach number (Ma), the root-mean-square (rms) granulation intensity amplitude σ, and the flicker amplitude F8. These data are also hosted, with updates as needed, on the first author's Web site (http://www.cfa.harvard.edu/~scranmer/). With the data is a short code written in the Interactive Data Language (IDL) that reads the data and reproduces two of the three panels of Figure4 in the paper. (3 data files).
NASA Technical Reports Server (NTRS)
Bogart, D. D.; Shook, D. F.; Fieno, D.
1973-01-01
Integral tests of evaluated ENDF/B high-energy cross sections have been made by comparing measured and calculated neutron leakage flux spectra from spheres of various materials. An Am-Be (alpha,n) source was used to provide fast neutrons at the center of the test spheres of Be, CH2, Pb, Nb, Mo, Ta, and W. The absolute leakage flux spectra were measured in the energy range 0.5 to 12 MeV using a calibrated NE213 liquid scintillator neutron spectrometer. Absolute calculations of the spectra were made using version 3 ENDF/B cross sections and an S sub n discrete ordinates multigroup transport code. Generally excellent agreement was obtained for Be, CH2, Pb, and Mo, and good agreement was observed for Nb although discrepancies were observed for some energy ranges. Poor comparative results, obtained for Ta and W, are attributed to unsatisfactory nonelastic cross sections. The experimental sphere leakage flux spectra are tabulated and serve as possible benchmarks for these elements against which reevaluated cross sections may be tested.
Critical Care Coding for Neurologists.
Nuwer, Marc R; Vespa, Paul M
2015-10-01
Accurate coding is an important function of neurologic practice. This contribution to Continuum is part of an ongoing series that presents helpful coding information along with examples related to the issue topic. Tips for diagnosis coding, Evaluation and Management coding, procedure coding, or a combination are presented, depending on which is most applicable to the subject area of the issue.
Coding of Neuroinfectious Diseases.
Barkley, Gregory L
2015-12-01
Accurate coding is an important function of neurologic practice. This contribution to Continuum is part of an ongoing series that presents helpful coding information along with examples related to the issue topic. Tips for diagnosis coding, Evaluation and Management coding, procedure coding, or a combination are presented, depending on which is most applicable to the subject area of the issue.
Diagnostic Coding for Epilepsy.
Williams, Korwyn; Nuwer, Marc R; Buchhalter, Jeffrey R
2016-02-01
Accurate coding is an important function of neurologic practice. This contribution to Continuum is part of an ongoing series that presents helpful coding information along with examples related to the issue topic. Tips for diagnosis coding, Evaluation and Management coding, procedure coding, or a combination are presented, depending on which is most applicable to the subject area of the issue.
48 CFR 52.204-18 - Commercial and Government Entity Code Maintenance.
Code of Federal Regulations, 2014 CFR
2014-10-01
... assigned to entities located in the United States or its outlying areas by the Defense Logistics Agency... NATO Support Agency (NSPA) to entities located outside the United States and its outlying areas that... code on the contract. (c) Contractors located in the United States or its outlying areas that are not...
Yancey, Antronette K; Cole, Brian L; Brown, Rochelle; Williams, Jerome D; Hillier, Amy; Kline, Randolph S; Ashe, Marice; Grier, Sonya A; Backman, Desiree; McCarthy, William J
2009-03-01
Commercial marketing is a critical but understudied element of the sociocultural environment influencing Americans' food and beverage preferences and purchases. This marketing also likely influences the utilization of goods and services related to physical activity and sedentary behavior. A growing literature documents the targeting of racial/ethnic and income groups in commercial advertisements in magazines, on billboards, and on television that may contribute to sociodemographic disparities in obesity and chronic disease risk and protective behaviors. This article examines whether African Americans, Latinos, and people living in low-income neighborhoods are disproportionately exposed to advertisements for high-calorie, low nutrient-dense foods and beverages and for sedentary entertainment and transportation and are relatively underexposed to advertising for nutritious foods and beverages and goods and services promoting physical activities. Outdoor advertising density and content were compared in zip code areas selected to offer contrasts by area income and ethnicity in four cities: Los Angeles, Austin, New York City, and Philadelphia. Large variations were observed in the amount, type, and value of advertising in the selected zip code areas. Living in an upper-income neighborhood, regardless of its residents' predominant ethnicity, is generally protective against exposure to most types of obesity-promoting outdoor advertising (food, fast food, sugary beverages, sedentary entertainment, and transportation). The density of advertising varied by zip code area race/ethnicity, with African American zip code areas having the highest advertising densities, Latino zip code areas having slightly lower densities, and white zip code areas having the lowest densities. The potential health and economic implications of differential exposure to obesity-related advertising are substantial. Although substantive legal questions remain about the government's ability to regulate advertising, the success of limiting tobacco advertising offers lessons for reducing the marketing contribution to the obesigenicity of urban environments.
Motion-adaptive model-assisted compatible coding with spatiotemporal scalability
NASA Astrophysics Data System (ADS)
Lee, JaeBeom; Eleftheriadis, Alexandros
1997-01-01
We introduce the concept of motion adaptive spatio-temporal model-assisted compatible (MA-STMAC) coding, a technique to selectively encode areas of different importance to the human eye in terms of space and time in moving images with the consideration of object motion. PRevious STMAC was proposed base don the fact that human 'eye contact' and 'lip synchronization' are very important in person-to-person communication. Several areas including the eyes and lips need different types of quality, since different areas have different perceptual significance to human observers. The approach provides a better rate-distortion tradeoff than conventional image coding techniques base don MPEG-1, MPEG- 2, H.261, as well as H.263. STMAC coding is applied on top of an encoder, taking full advantage of its core design. Model motion tracking in our previous STMAC approach was not automatic. The proposed MA-STMAC coding considers the motion of the human face within the STMAC concept using automatic area detection. Experimental results are given using ITU-T H.263, addressing very low bit-rate compression.
A standard-driven approach for electronic submission to pharmaceutical regulatory authorities.
Lin, Ching-Heng; Chou, Hsin-I; Yang, Ueng-Cheng
2018-03-01
Using standards is not only useful for data interchange during the process of a clinical trial, but also useful for analyzing data in a review process. Any step, which speeds up approval of new drugs, may benefit patients. As a result, adopting standards for regulatory submission becomes mandatory in some countries. However, preparing standard-compliant documents, such as annotated case report form (aCRF), needs a great deal of knowledge and experience. The process is complex and labor-intensive. Therefore, there is a need to use information technology to facilitate this process. Instead of standardizing data after the completion of a clinical trial, this study proposed a standard-driven approach. This approach was achieved by implementing a computer-assisted "standard-driven pipeline (SDP)" in an existing clinical data management system. SDP used CDISC standards to drive all processes of a clinical trial, such as the design, data acquisition, tabulation, etc. RESULTS: A completed phase I/II trial was used to prove the concept and to evaluate the effects of this approach. By using the CDISC-compliant question library, aCRFs were generated automatically when the eCRFs were completed. For comparison purpose, the data collection process was simulated and the collected data was transformed by the SDP. This new approach reduced the missing data fields from sixty-two to eight and the controlled term mismatch field reduced from eight to zero during data tabulation. This standard-driven approach accelerated CRF annotation and assured data tabulation integrity. The benefits of this approach include an improvement in the use of standards during the clinical trial and a reduction in missing and unexpected data during tabulation. The standard-driven approach is an advanced design idea that can be used for future clinical information system development. Copyright © 2018 Elsevier Inc. All rights reserved.
Thermodynamic data for fifty reference elements
NASA Technical Reports Server (NTRS)
Mcbride, Bonnie J.; Gordon, Sanford; Reno, Martin A.
1993-01-01
This report is a compilation of thermodynamic functions of 50 elements in their standard reference state. The functions are C(sub p)(sup 0), (H(sup 0)(T) - H(sup 0)(0)), S(sup 0)(T), and -(G(sup 0)(T) - H(sup 0)(O)) for the elements Ag, Al, Ar, B, Ba, Be, Br2, C, Ca, Cd, Cl2, Co, Cr, Cs, Cu, F2, Fe, Ge, H2, He, Hg, I2, K, Kr, Li, Mg, Mn, Mo, N2, Na, Nb, Ne, Ni, O2, P, Pb, Rb, S, Si, Sn, Sr, Ta, Th, Ti, U, V, W, Xe, Zn, and Zr. Deuterium D2 and electron gas e(sup -) are also included. The data are tabulated as functions of temperature as well as given in the form of least-squares coefficients for two functional forms for C(sub p)(sup 0) with integration constants for enthalpy and entropy. One functional form for C(sub p)(sup 0) is a fourth-order polynomial and the other has two additional terms, one with T(exp -1) and the other with T(exp -2). The gases Ar, D2, e(sup -), H2, He, Kr, N2, Ne, O2, and Xe are tabulated for temperatures from 100 to 20,000 K. The remaining gases Cl2 and F2 are tabulated from 100 to 6000 K and 1000 to 6000 K. The second functional form for C(sub p)(sup 0) has an additional interval from 6000 to 20,000 K for the gases tabulated to 20,000 K. The fits are constrained so that the match at the common temperature endpoints. The temperature ranges for the condensed species vary with range of the data, phase changes, and shapes of the C(sub p)(sup 0) curves.
Employment and residential characteristics in relation to automated external defibrillator locations
Griffis, Heather M.; Band, Roger A; Ruther, Matthew; Harhay, Michael; Asch, David A.; Hershey, John C.; Hill, Shawndra; Nadkarni, Lindsay; Kilaru, Austin; Branas, Charles C.; Shofer, Frances; Nichol, Graham; Becker, Lance B.; Merchant, Raina M.
2015-01-01
Background Survival from out-of-hospital cardiac arrest (OHCA) is generally poor and varies by geography. Variability in automated external defibrillator (AED) locations may be a contributing factor. To inform optimal placement of AEDs, we investigated AED access in a major US city relative to demographic and employment characteristics. Methods and Results This was a retrospective analysis of a Philadelphia AED registry (2,559 total AEDs). The 2010 US Census and the Local Employment Dynamics (LED) database by ZIP code was used. AED access was calculated as the weighted areal percentage of each ZIP code covered by a 400 meter radius around each AED. Of 47 ZIP codes, only 9%(4) were high AED service areas. In 26%(12) of ZIP codes, less than 35% of the area was covered by AED service areas. Higher AED access ZIP codes were more likely to have a moderately populated residential area (p=0.032), higher median household income (p=0.006), and higher paying jobs (p=008). Conclusions The locations of AEDs vary across specific ZIP codes; select residential and employment characteristics explain some variation. Further work on evaluating OHCA locations, AED use and availability, and OHCA outcomes could inform AED placement policies. Optimizing the placement of AEDs through this work may help to increase survival. PMID:26856232
Code of Federal Regulations, 2012 CFR
2012-10-01
... Administration § 52.7 Definitions. As used in this subpart: (a) Area code or numbering plan area (NPA). The term “area code or numbering plan area” refers to the first three digits (NXX) of a ten-digit telephone... “central office code” refers to the second three digits (NXX) of a ten-digit telephone number in the form...
Code of Federal Regulations, 2011 CFR
2011-10-01
... Administration § 52.7 Definitions. As used in this subpart: (a) Area code or numbering plan area (NPA). The term “area code or numbering plan area” refers to the first three digits (NXX) of a ten-digit telephone... “central office code” refers to the second three digits (NXX) of a ten-digit telephone number in the form...
Code of Federal Regulations, 2013 CFR
2013-10-01
... Administration § 52.7 Definitions. As used in this subpart: (a) Area code or numbering plan area (NPA). The term “area code or numbering plan area” refers to the first three digits (NXX) of a ten-digit telephone... “central office code” refers to the second three digits (NXX) of a ten-digit telephone number in the form...
Code of Federal Regulations, 2014 CFR
2014-10-01
... Administration § 52.7 Definitions. As used in this subpart: (a) Area code or numbering plan area (NPA). The term “area code or numbering plan area” refers to the first three digits (NXX) of a ten-digit telephone... “central office code” refers to the second three digits (NXX) of a ten-digit telephone number in the form...
Planning for Psychological Operations A Proposal
1997-03-01
of military PSYOP and calls for redefining an area of operations that has changed little over the years. Additionally, the establishment of an...19b. TELEPHONE NUMBER International Area Code Area Code Telephone Number 703767-9007 DSN 427-9007 Standard Form 298 (Rev. 8-98) Prescribed by ANSI Std...aspect of military PSYOP and calls for redefining an area of operations that has changed little over the years. Additionally, the establishment of an
National Underground Mines Inventory
1983-10-01
system is well designed to minimize water accumulation on the drift levels. In many areas, sufficient water has accumulated to make the use of boots a...four characters designate Field office. 17-18 State Code Pic 99 FIPS code for state in which minets located. 19-21 County Code Plc 999 FIPS code for... Designate a general product class based onSIC code. 28-29 Nine Type Plc 99 Natal/Nonmetal mine type code. Based on subunit operations code and canvass code
O' Lenick, Cassandra R; Chang, Howard H; Kramer, Michael R; Winquist, Andrea; Mulholland, James A; Friberg, Mariel D; Sarnat, Stefanie Ebelt
2017-04-05
Ground-level ozone is a potent airway irritant and a determinant of respiratory morbidity. Susceptibility to the health effects of ambient ozone may be influenced by both intrinsic and extrinsic factors, such as neighborhood socioeconomic status (SES). Questions remain regarding the manner and extent that factors such as SES influence ozone-related health effects, particularly across different study areas. Using a 2-stage modeling approach we evaluated neighborhood SES as a modifier of ozone-related pediatric respiratory morbidity in Atlanta, Dallas, & St. Louis. We acquired multi-year data on emergency department (ED) visits among 5-18 year olds with a primary diagnosis of respiratory disease in each city. Daily concentrations of 8-h maximum ambient ozone were estimated for all ZIP Code Tabulation Areas (ZCTA) in each city by fusing observed concentration data from available network monitors with simulations from an emissions-based chemical transport model. In the first stage, we used conditional logistic regression to estimate ZCTA-specific odds ratios (OR) between ozone and respiratory ED visits, controlling for temporal trends and meteorology. In the second stage, we combined ZCTA-level estimates in a Bayesian hierarchical model to assess overall associations and effect modification by neighborhood SES considering categorical and continuous SES indicators (e.g., ZCTA-specific levels of poverty). We estimated ORs and 95% posterior intervals (PI) for a 25 ppb increase in ozone. The hierarchical model combined effect estimates from 179 ZCTAs in Atlanta, 205 ZCTAs in Dallas, and 151 ZCTAs in St. Louis. The strongest overall association of ozone and pediatric respiratory disease was in Atlanta (OR = 1.08, 95% PI: 1.06, 1.11), followed by Dallas (OR = 1.04, 95% PI: 1.01, 1.07) and St. Louis (OR = 1.03, 95% PI: 0.99, 1.07). Patterns of association across levels of neighborhood SES in each city suggested stronger ORs in low compared to high SES areas, with some evidence of non-linear effect modification. Results suggest that ozone is associated with pediatric respiratory morbidity in multiple US cities; neighborhood SES may modify this association in a non-linear manner. In each city, children living in low SES environments appear to be especially vulnerable given positive ORs and high underlying rates of respiratory morbidity.
Methodology, status and plans for development and assessment of TUF and CATHENA codes
DOE Office of Scientific and Technical Information (OSTI.GOV)
Luxat, J.C.; Liu, W.S.; Leung, R.K.
1997-07-01
An overview is presented of the Canadian two-fluid computer codes TUF and CATHENA with specific focus on the constraints imposed during development of these codes and the areas of application for which they are intended. Additionally a process for systematic assessment of these codes is described which is part of a broader, industry based initiative for validation of computer codes used in all major disciplines of safety analysis. This is intended to provide both the licensee and the regulator in Canada with an objective basis for assessing the adequacy of codes for use in specific applications. Although focused specifically onmore » CANDU reactors, Canadian experience in developing advanced two-fluid codes to meet wide-ranging application needs while maintaining past investment in plant modelling provides a useful contribution to international efforts in this area.« less
The revised burn diagram and its effect on diagnosis-related group coding.
Turner, D G; Berger, N; Weiland, A P; Jordan, M H
1996-01-01
Diagnosis-related group (DRG) codes for burn injuries are defined by thresholds of the percentage of total body surface area and depth of burns, and by whether surgery, debridement, or grafting or both occurred. This prospective study was designed to determine whether periodic revisions of the burn diagram resulted in more accurate assignment of the International Classification of Diseases and DRG codes. The admission burn diagrams were revised after admission and after each surgical procedure. All areas grafted (deep second-and third-degree burns) were diagrammed as "third-degree," after the current convention that both are biologically the same and require grafting. The multiple diagrams from 82 charts were analyzed to determine the disparities in the percentage of total body surface area burn and the percentage of body surface area third-degree burn. The revised diagrams differed from the admission diagrams in 96.5% of the cases. In 77% of the cases, the revised diagram correctly depicted the percentage of body surface area third-degree burn as confirmed intraoperatively. In 7.3% of the cases, diagram revision changed the DRG code. Documenting wound evolution in this manner allows more accurate assignment of the International Classification of Diseases and DRG codes, assuring optimal reimbursement under the prospective payment system.
Wafula, Charles Ouma; Edwards, Nancy; Kaseje, Dan C O
2017-02-28
Many low and middle income countries have developed community health strategies involving lay health workers, to complement and strengthen public health services. This study explores variations in costing parameters pertinent to deployment of community health volunteers across different contexts outlining considerations for costing program scale-up. The study used quasi experimental study design and employed both quantitative and qualitative methods to explore community health unit implementation activities and costs and compare costs across purposively selected sites that differed socially, economically and ecologically. Data were collected from November 2010 to December 2013 through key informant interviews and focus group discussions. We interviewed 16 key informants (eight District community health strategy focal persons, eight frontline field officers), and eight focus group discussions (four with community health volunteers and four with community health committee) and 560 sets of monthly cost data. Cost data were tabulated using Microsoft Excel. Qualitative data were transcribed and coded using a content analysis framework. Four critical elements: attrition rates for community health volunteers, geography and population density, livelihood opportunity costs and benefits, and social opportunity benefits, drove cost variations across the three sites. Attrition rate was highest in peri-urban site where population is highly mobile and lowest in nomadic site. More households were covered by community health workers in the peri-urban area making per capita costs considerably less than in the nomadic settings where long distances had to be covered to reach sparsely distributed households. Livelihood opportunity costs for Community Health Volunteers were highest in nomadic setting, while peri-urban ones reported substantial employability benefits resulting from training. Social opportunity benefits were highest in rural site. Results show that costs of implementing community health strategy varied due to different area contextual factors in Kenya. This study identified four critical elements that drive cost variations: attrition rates for community health volunteers, geography and population density, livelihood opportunity costs and benefits, and social opportunity benefits. Health programme managers and policy-makers need to pay attention to details of contextual factors in costing for effective implementation of community health strategies.
NASA Technical Reports Server (NTRS)
Mcgoogan, J. T.; Leitao, C. D.; Wells, W. T.
1975-01-01
The SKYLAB S-193 altimeter altitude results are presented in a concise format for further use and analysis by the scientific community. The altimeter mission and instrumentation is described along with the altimeter processing techniques and values of parameters used for processing. The determination of reference orbits is discussed, and the tracking systems utilized are tabulated. Techniques for determining satellite pointing are presented and a tabulation of pointing for each data mission included. The geographical location, the ocean bottom topography, the altimeter-determined ocean surface topography, and the altimeter automatic gain control history is presented. Some typical applications of this data are suggested.
NASA Technical Reports Server (NTRS)
Johnson, F. S.; Mo, T.; Green, A. E. S.
1976-01-01
Tabulated values are presented for ultraviolet radiation at the earth's surface as a function of wavelength, latitude, and season, for clear sky and seasonally and latitudinally averaged ozone amounts. These tabulations can be combined with any biological sensitivity function in order to obtain the seasonal and latitudinal variation of the corresponding effective doses. The integrated dosages, based on the erythemal sensitivity curve and on the Robertson-Berger sunburn-meter sensitivity curve, have also been calculated, and these are found to vary with latitude and season in very nearly the same way as 307 and 314 nm radiation, respectively.
Study of activation of metal samples from LDEF-1 and Spacelab-2
NASA Technical Reports Server (NTRS)
Laird, C. E.
1991-01-01
The activation of metal samples and other material orbited onboard the Long Duration Exposure Facility (LDEF) and Spacelab-2 were studied. Measurements of the radioactivities of spacecraft materials were made, and corrections for self-absorption and efficiency were calculated. Activation cross sections for specific metal samples were updated while cross sections for other materials were tabulated from the scientific literature. Activation cross sections for 200 MeV neutrons were experimentally determined. Linear absorption coefficients, half lives, branching ratios and other pertinent technical data needed for LDEF sample analyses were tabulated. The status of the sample counting at low background facilities at national laboratories is reported.
Zernike Basis to Cartesian Transformations
NASA Astrophysics Data System (ADS)
Mathar, R. J.
2009-12-01
The radial polynomials of the 2D (circular) and 3D (spherical) Zernike functions are tabulated as powers of the radial distance. The reciprocal tabulation of powers of the radial distance in series of radial polynomials is also given, based on projections that take advantage of the orthogonality of the polynomials over the unit interval. They play a role in the expansion of products of the polynomials into sums, which is demonstrated by some examples. Multiplication of the polynomials by the angular bases (azimuth, polar angle) defines the Zernike functions, for which we derive transformations to and from the Cartesian coordinate system centered at the middle of the circle or sphere.
X-ray natural widths, level widths and Coster-Kronig transition probabilities
NASA Astrophysics Data System (ADS)
Papp, T.; Campbell, J. L.; Varga, D.
1997-01-01
A critical review is given for the K-N7 atomic level widths. The experimental level widths were collected from x-ray photoelectron spectroscopy (XPS), x-ray emission spectroscopy (XES), x-ray spectra fluoresced by synchrotron radiation, and photoelectrons from x-ray absorption (PAX). There are only limited atomic number ranges for a few atomic levels where data are available from more than one source. Generally the experimental level widths have large scatter compared to the reported error bars. The experimental data are compared with the recent tabulation of Perkins et al. and of Ohno et al. Ohno et al. performed a many body approach calculation for limited atomic number ranges and have obtained reasonable agreement with the experimental data. Perkins et al. presented a tabulation covering the K-Q1 shells of all atoms, based on extensions of the Scofield calculations for radiative rates and extensions of the Chen calculations for non-radiative rates. The experimental data are in disagreement with this tabulation, in excess of a factor of two in some cases. A short introduction to the experimental Coster-Kronig transition probabilities is presented. It is our opinion that the different experimental approaches result in systematically different experimental data.
NASA Astrophysics Data System (ADS)
Wan, Kaidi; Xia, Jun; Vervisch, Luc; Liu, Yingzu; Wang, Zhihua; Cen, Kefa
2018-03-01
The numerical modelling of alkali metal reacting dynamics in turbulent pulverised-coal combustion is discussed using tabulated sodium chemistry in large eddy simulation (LES). A lookup table is constructed from a detailed sodium chemistry mechanism including five sodium species, i.e. Na, NaO, NaO2, NaOH and Na2O2H2, and 24 elementary reactions. This sodium chemistry table contains four coordinates, i.e. the equivalence ratio, the mass fraction of the sodium element, the gas-phase temperature, and a progress variable. The table is first validated against the detailed sodium chemistry mechanism by zero-dimensional simulations. Then, LES of a turbulent pulverised-coal jet flame is performed and major coal-flame parameters compared against experiments. The chemical percolation devolatilisation (CPD) model and the partially stirred reactor (PaSR) model are employed to predict coal pyrolysis and gas-phase combustion, respectively. The response of the five sodium species in the pulverised-coal jet flame is subsequently examined. Finally, a systematic global sensitivity analysis of the sodium lookup table is performed and the accuracy of the proposed tabulated sodium chemistry approach has been calibrated.
Content analysis of CMS Healthcare Innovation Awards from a physiatric perspective.
Smith, Matthew J
2014-11-01
On June 15, 2012, the Centers for Medicare and Medicaid Services (CMS) announced the 107 recipients of the Healthcare Innovation Awards (HCIA). The 107 awardees received a total of $874,074,900 in funding, with a projected 3-year savings of $1,863,119,104. A word frequency calculator was used to perform a content analysis on the document that announced the projects receiving funding in the 2012 HCIA program. Results were tabulated and categorized to look for prevailing themes and trends. The words generated by the word frequency calculator were grouped into common roots and tabulated to better understand how CMS was rewarding value. Some of the most common words were "manage," "community," "coordinate," "team," "system" and "integrate." Additionally, the job positions that the projects propose to create were tabulated and grouped into categories. Physicians, including physiatrists, were not often mentioned, whereas nursing and nonclinical positions were frequently listed. This content analysis showed that the concepts emphasized in the HCIA projects parallel fundamental physiatric principles. The findings may help physiatrists understand how reform is unfolding, prepare for the evolving health care landscape, and recognize future opportunities. Copyright © 2014. Published by Elsevier Inc.
Predicting Regulatory Compliance in Beer Advertising on Facebook.
Noel, Jonathan K; Babor, Thomas F
2017-11-01
The prevalence of alcohol advertising has been growing on social media platforms. The purpose of this study was to evaluate alcohol advertising on Facebook for regulatory compliance and thematic content. A total of 50 Budweiser and Bud Light ads posted on Facebook within 1 month of the 2015 NFL Super Bowl were evaluated for compliance with a self-regulated alcohol advertising code and for thematic content. An exploratory sensitivity/specificity analysis was conducted to determine if thematic content could predict code violations. The code violation rate was 82%, with violations prevalent in guidelines prohibiting the association of alcohol with success (Guideline 5) and health benefits (Guideline 3). Overall, 21 thematic content areas were identified. Displaying the product (62%) and adventure/sensation seeking (52%) were the most prevalent. There was perfect specificity (100%) for 10 content areas for detecting any code violation (animals, negative emotions, positive emotions, games/contests/promotions, female characters, minorities, party, sexuality, night-time, sunrise) and high specificity (>80%) for 10 content areas for detecting violations of guidelines intended to protect minors (animals, negative emotions, famous people, friendship, games/contests/promotions, minorities, responsibility messages, sexuality, sunrise, video games). The high prevalence of code violations indicates a failure of self-regulation to prevent potentially harmful content from appearing in alcohol advertising, including explicit code violations (e.g. sexuality). Routine violations indicate an unwillingness to restrict advertising content for public health purposes, and statutory restrictions may be necessary to sufficiently deter alcohol producers from repeatedly violating marketing codes. Violations of a self-regulated alcohol advertising code are prevalent in a sample of beer ads published on Facebook near the US National Football League's Super Bowl. Overall, 16 thematic content areas demonstrated high specificity for code violations. Alcohol advertising codes should be updated to expressly prohibit the use of such content. © The Author 2017. Medical Council on Alcohol and Oxford University Press. All rights reserved.
NASA Technical Reports Server (NTRS)
Baumeister, K. J.; Eversman, W.; Astley, R. J.; White, J. W.
1981-01-01
Experimental data are presented for sound propagation in a simulated infinite hard wall duct with a large change in duct cross sectional area. The data are conveniently tabulated for further use. The 'steady' state finite element theory of Astley and Eversman (1981) and the transient finite difference theory of White (1981) are in good agreement with the data for both the axial and transverse pressure profiles and the axial phase angle. Therefore, numerical finite difference and finite element theories appear to be ideally suited for handling duct propagation problems which encounter large axial gradients in acoustic parameters. The measured energy reflection coefficient agrees with the values from the Astley-Eversman modal coupling model.
Use of ERTS-1 imagery in forest inventory
NASA Technical Reports Server (NTRS)
Rennie, J. C.; Birth, E. E.
1974-01-01
The utility of ERTS-1 imagery when combined with field observations and with aircraft imagery and field observations is evaluated. Satellite imagery consisted of 9-1/2 inch black and white negatives of four multispectral scanner bands taken over Polk County, Tennessee. Aircraft imagery was obtained by a C-130 flying at 23,000 ft over the same area and provided the basis for locating ground plots for field observations. Correspondence between aircraft and satellite imagery was somewhat inaccurate due to seasonal differences in observations and lack of good photogrammetry with the data processing system used. Better correspondence was found between satellite imagery and ground observations. Ways to obtain more accurate data are discussed, and comparisons between aircraft and satellite observations are tabulated.
Rocky Mountain Arsenal Ecological Chemical Data (1984-1985)
1986-03-01
Type Wet Areas Code Area LK Lake MT Marshy Type PD Pond C-1 APPENDIX D TISSUE CODES Code Tissue BRA Brain FIL Filet EDP Edible Portion LIV Liver MUS...ctus 472 Craton texcisis Crotoni 473 Cryptantha fcndleri Fcnder’s Cryptantha 474 Cucurbita foctidissinma Wild Gourd 475 Cytroptcrus montanus Pink Cym...adcrnocaulon Nlorthern wilIlow- herb 484 Eragrostis cilianensis Stinkgrass 485 Erfcgonum annuum Tall Erlogonum 486 Erigercn divergcns Spreading Fleabine 487
Welch, Alan H.; Bright, Daniel J.; Knochenmus, Lari A.
2008-01-01
INTRODUCTION This report summarizes results of a water-resources study for White Pine County, Nevada, and adjacent areas in east-central Nevada and western Utah. The Basin and Range carbonate-rock aquifer system (BARCAS) study was initiated in December 2004 through Federal legislation (Section 301(e) of the Lincoln County Conservation, Recreation, and Development Act of 2004; PL108-424) directing the Secretary of the Interior to complete a water-resources study through the U.S. Geological Survey, Desert Research Institute, and State of Utah. The study was designed as a regional water-resource assessment, with particular emphasis on summarizing the hydrogeologic framework and hydrologic processes that influence ground-water resources. The study area includes 13 hydrographic areas that cover most of White Pine County; in this report however, results for the northern and central parts of Little Smoky Valley were combined and presented as one hydrographic area. Hydrographic areas are the basic geographic units used by the State of Nevada and Utah and local agencies for water-resource planning and management, and are commonly defined on the basis of surface-water drainage areas. Hydrographic areas were further divided into subbasins that are separated by areas where bedrock is at or near the land surface. Subbasins are the subdivisions used in this study for estimating recharge, discharge, and water budget. Hydrographic areas are the subdivision used for reporting summed and tabulated subbasin estimates.
Welch, Alan H.; Bright, Daniel J.
2007-01-01
Summary of Major Findings This report summarizes results of a water-resources study for White Pine County, Nevada, and adjacent areas in east-central Nevada and western Utah. The Basin and Range carbonate-rock aquifer system (BARCAS) study was initiated in December 2004 through Federal legislation (Section 131 of the Lincoln County Conservation, Recreation, and Development Act of 2004) directing the Secretary of the Interior to complete a water-resources study through the U.S. Geological Survey, Desert Research Institute, and State of Utah. The study was designed as a regional water-resource assessment, with particular emphasis on summarizing the hydrogeologic framework and hydrologic processes that influence ground-water resources. The study area includes 13 hydrographic areas that cover most of White Pine County; in this report however, results for the northern and central parts of Little Smoky Valley were combined and presented as one hydrographic area. Hydrographic areas are the basic geographic units used by the State of Nevada and Utah and local agencies for water-resource planning and management, and are commonly defined on the basis of surface-water drainage areas. Hydrographic areas were further divided into subbasins that are separated by areas where bedrock is at or near the land surface. Subbasins represent subdivisions used in this study for estimating recharge, discharge, and water budget. Hydrographic areas represent the subdivision used for reporting summed and tabulated subbasin estimates.
From Phonemes to Articulatory Codes: An fMRI Study of the Role of Broca's Area in Speech Production
de Zwart, Jacco A.; Jansma, J. Martijn; Pickering, Martin J.; Bednar, James A.; Horwitz, Barry
2009-01-01
We used event-related functional magnetic resonance imaging to investigate the neuroanatomical substrates of phonetic encoding and the generation of articulatory codes from phonological representations. Our focus was on the role of the left inferior frontal gyrus (LIFG) and in particular whether the LIFG plays a role in sublexical phonological processing such as syllabification or whether it is directly involved in phonetic encoding and the generation of articulatory codes. To answer this question, we contrasted the brain activation patterns elicited by pseudowords with high– or low–sublexical frequency components, which we expected would reveal areas related to the generation of articulatory codes but not areas related to phonological encoding. We found significant activation of a premotor network consisting of the dorsal precentral gyrus, the inferior frontal gyrus bilaterally, and the supplementary motor area for low– versus high–sublexical frequency pseudowords. Based on our hypothesis, we concluded that these areas and in particular the LIFG are involved in phonetic and not phonological encoding. We further discuss our findings with respect to the mechanisms of phonetic encoding and provide evidence in support of a functional segregation of the posterior part of Broca's area, the pars opercularis. PMID:19181696
Schubotz, Ricarda I.; Wurm, Moritz F.; Wittmann, Marco K.; von Cramon, D. Yves
2014-01-01
Objects are reminiscent of actions often performed with them: knife and apple remind us on peeling the apple or cutting it. Mnemonic representations of object-related actions (action codes) evoked by the sight of an object may constrain and hence facilitate recognition of unrolling actions. The present fMRI study investigated if and how action codes influence brain activation during action observation. The average number of action codes (NAC) of 51 sets of objects was rated by a group of n = 24 participants. In an fMRI study, different volunteers were asked to recognize actions performed with the same objects presented in short videos. To disentangle areas reflecting the storage of action codes from those exploiting them, we showed object-compatible and object-incompatible (pantomime) actions. Areas storing action codes were considered to positively co-vary with NAC in both object-compatible and object-incompatible action; due to its role in tool-related tasks, we here hypothesized left anterior inferior parietal cortex (aIPL). In contrast, areas exploiting action codes were expected to show this correlation only in object-compatible but not incompatible action, as only object-compatible actions match one of the active action codes. For this interaction, we hypothesized ventrolateral premotor cortex (PMv) to join aIPL due to its role in biasing competition in IPL. We found left anterior intraparietal sulcus (IPS) and left posterior middle temporal gyrus (pMTG) to co-vary with NAC. In addition to these areas, action codes increased activity in object-compatible action in bilateral PMv, right IPS, and lateral occipital cortex (LO). Findings suggest that during action observation, the brain derives possible actions from perceived objects, and uses this information to shape action recognition. In particular, the number of expectable actions quantifies the activity level at PMv, IPL, and pMTG, but only PMv reflects their biased competition while observed action unfolds. PMID:25009519
DOE Office of Scientific and Technical Information (OSTI.GOV)
Walker, Andrew; Lawrence, Earl
The Response Surface Modeling (RSM) Tool Suite is a collection of three codes used to generate an empirical interpolation function for a collection of drag coefficient calculations computed with Test Particle Monte Carlo (TPMC) simulations. The first code, "Automated RSM", automates the generation of a drag coefficient RSM for a particular object to a single command. "Automated RSM" first creates a Latin Hypercube Sample (LHS) of 1,000 ensemble members to explore the global parameter space. For each ensemble member, a TPMC simulation is performed and the object drag coefficient is computed. In the next step of the "Automated RSM" code,more » a Gaussian process is used to fit the TPMC simulations. In the final step, Markov Chain Monte Carlo (MCMC) is used to evaluate the non-analytic probability distribution function from the Gaussian process. The second code, "RSM Area", creates a look-up table for the projected area of the object based on input limits on the minimum and maximum allowed pitch and yaw angles and pitch and yaw angle intervals. The projected area from the look-up table is used to compute the ballistic coefficient of the object based on its pitch and yaw angle. An accurate ballistic coefficient is crucial in accurately computing the drag on an object. The third code, "RSM Cd", uses the RSM generated by the "Automated RSM" code and the projected area look-up table generated by the "RSM Area" code to accurately compute the drag coefficient and ballistic coefficient of the object. The user can modify the object velocity, object surface temperature, the translational temperature of the gas, the species concentrations of the gas, and the pitch and yaw angles of the object. Together, these codes allow for the accurate derivation of an object's drag coefficient and ballistic coefficient under any conditions with only knowledge of the object's geometry and mass.« less
Michalaki, M; Oulis, C J; Pandis, N; Eliades, G
2016-12-01
This in vitro study was to classify questionable for caries occlusal surfaces (QCOS) of permanent teeth according to ICDAS codes 1, 2, and 3 and to compare them in terms of enamel mineral composition with the areas of sound tissue of the same tooth. Partially impacted human molars (60) extracted for therapeutic reasons with QCOS were used in the study, photographed via a polarised light microscope and classified according to the ICDAS II (into codes 1, 2, or 3). The crowns were embedded in clear self-cured acrylic resin and longitudinally sectioned at the levels of the characterised lesions and studied by SEM/EDX, to assess enamel mineral composition of the QCOS. Univariate and multivariate random effect regressions were used for Ca (wt%), P (wt%), and Ca/P (wt%). The EDX analysis indicated changes in the Ca and P contents that were more prominent in ICDAS-II code 3 lesions compared to codes 1 and 2 lesions. In these lesions, Ca (wt%) and P (wt%) concentrations were significantly decreased (p = 0.01) in comparison with sound areas. Ca and P (wt%) contents were significantly lower (p = 0.02 and p = 0.01 respectively) for code 3 areas in comparison with codes 1 and 2 areas. Significantly higher (p = 0.01) Ca (wt%) and P (wt%) contents were found on sound areas compared to the lesion areas. The enamel of occlusal surfaces of permanent teeth with ICDAS 1, 2, and 3 lesions was found to have different Ca/P compositions, necessitating further investigation on whether these altered surfaces might behave differently on etching preparation before fissure sealant placement, compared to sound surfaces.
One Speaker, Two Languages. Cross-Disciplinary Perspectives on Code-Switching.
ERIC Educational Resources Information Center
Milroy, Lesley, Ed.; Muysken, Pieter, Ed.
Fifteen articles review code-switching in the four major areas: policy implications in specific institutional and community settings; perspectives of social theory of code-switching as a form of speech behavior in particular social contexts; the grammatical analysis of code-switching, including factors that constrain switching even within a…
Application of LEED (trademark) and SPiRiT to a Proposed Building Design
2003-03-01
TOTAL AREA S CAT ARLOC INSTALLATION NO CODE C QTY UM OCPD F CODE 1) 24004 Aberdeen Provi E1890...ARLOC INSTALLATION NO CODE C QTY UM OCPD F CODE 1) 24004 Aberdeen Provi E3832 31010 P 13,000 SF 13,000 R
Recent advances in coding theory for near error-free communications
NASA Technical Reports Server (NTRS)
Cheung, K.-M.; Deutsch, L. J.; Dolinar, S. J.; Mceliece, R. J.; Pollara, F.; Shahshahani, M.; Swanson, L.
1991-01-01
Channel and source coding theories are discussed. The following subject areas are covered: large constraint length convolutional codes (the Galileo code); decoder design (the big Viterbi decoder); Voyager's and Galileo's data compression scheme; current research in data compression for images; neural networks for soft decoding; neural networks for source decoding; finite-state codes; and fractals for data compression.
Manimegalai, C T; Gauni, Sabitha; Kalimuthu, K
2017-12-04
Wireless body area network (WBAN) is a breakthrough technology in healthcare areas such as hospital and telemedicine. The human body has a complex mixture of different tissues. It is expected that the nature of propagation of electromagnetic signals is distinct in each of these tissues. This forms the base for the WBAN, which is different from other environments. In this paper, the knowledge of Ultra Wide Band (UWB) channel is explored in the WBAN (IEEE 802.15.6) system. The measurements of parameters in frequency range from 3.1-10.6 GHz are taken. The proposed system, transmits data up to 480 Mbps by using LDPC coded APSK Modulated Differential Space-Time-Frequency Coded MB-OFDM to increase the throughput and power efficiency.
NASA Technical Reports Server (NTRS)
Glassman, Arthur J.; Jones, Scott M.
1991-01-01
This analysis and this computer code apply to full, split, and dual expander cycles. Heat regeneration from the turbine exhaust to the pump exhaust is allowed. The combustion process is modeled as one of chemical equilibrium in an infinite-area or a finite-area combustor. Gas composition in the nozzle may be either equilibrium or frozen during expansion. This report, which serves as a users guide for the computer code, describes the system, the analysis methodology, and the program input and output. Sample calculations are included to show effects of key variables such as nozzle area ratio and oxidizer-to-fuel mass ratio.
Winterstein, T.A.
1982-01-01
This inventory of reports and data concerning the Mississippi and Minnesota Rivers in the Twin Cities metropolitan area was compiled from November 1981 through January 1982 for a planned river-quality assessment to be conducted cooperatively by the U.S. Geological Survey and the Metropolitan Waste Control Commission. There are 260 annotated citations: 176 citations of reports; 8 citations of computer models that have been used to model either or both rivers; and 76 citations of data in reports , in field notes, lab sheets, or handwritten tabulations, and in computer data bases. Citations of all the reports and data located that might conceivably be useful in understanding and interpreting the biological and chemical quality of the Mississippi and Minnesota Rivers in the past, present, or future were included. The accuracy of the citations was not verified and secondary sources, such as other annotated bibliographies, were used in the compilation of this inventory.
Static Footprint Local Forces, Areas, and Aspect Ratios for Three Type 7 Aircraft Tires
NASA Technical Reports Server (NTRS)
Howell, William E.; Perez, Sharon E.; Vogler, William A.
1991-01-01
The National Tire Modeling Program (NTMP) is a joint NASA/industry effort to improve the understanding of tire mechanics and develop accurate analytical design tools. This effort includes fundamental analytical and experimental research on the structural mechanics of tires. Footprint local forces, areas, and aspect ratios were measured. Local footprint forces in the vertical, lateral, and drag directions were measured with a special footprint force transducer. Measurements of the local forces in the footprint were obtained by positioning the transducer at specified locations within the footprint and externally loading the tires. Three tires were tested: (1) one representative of those used on the main landing gear of B-737 and DC-9 commercial transport airplanes, (2) a nose landing gear tire for the Space Shuttle Orbiter, and (3) a main landing gear tire for the Space Shuttle Orbiter. Data obtained for various inflation pressures and vertical loads are presented for two aircraft tires. The results are presented in graphical and tabulated forms.
The Pelican Nebula and its Vicinity: a New Look at Stellar Population in the Cloud and around It
NASA Astrophysics Data System (ADS)
Boyle, Richard P.; Janusz, R.; Vrba, F. J.; Straizys, V.; Laugalys, V.; Kazlauskas, A.; Stott, J.; Philip, A. G. D.
2011-01-01
A region of active star formation is located in the complex of dust and molecular clouds known as the Pelican Nebula and the dark cloud L935. In this paper we describe the results of our investigation in the area bounded by the coordinates (2000) RA 20h50m - 20h54m and DEC +44d20m - 44m55d. Our CCD photometry in the Vilnius seven-color system, obtained on the 1.8 m Vatican Advanced Technology Telescope, Mt. Graham, and the 1 m telescope of the USNO Flagstaff Station, is used to classify stars down to V = 17 mag in spectral and luminosity classes. The interstellar extinction values and distances to these stars are determined. Additionally, the data from the 2MASS, MegaCam, IPHAS and Spitzer surveys are analyzed. We present star population maps in the foreground and background of the complex and within it. The known and newly identified YSOs in the area are tabulated.
Infectious diseases affect marine fisheries and aquaculture economics
Lafferty, Kevin D.; Harvell, C. Drew; Conrad, Jonathan M.; Friedman, Carolyn S.; Kent, Michael L.; Kuris, Armand M.; Powell, Eric N.; Rondeau, Daniel; Saksida, Sonja M.
2015-01-01
Seafood is a growing part of the economy, but its economic value is diminished by marine diseases. Infectious diseases are common in the ocean, and here we tabulate 67 examples that can reduce commercial species' growth and survivorship or decrease seafood quality. These impacts seem most problematic in the stressful and crowded conditions of aquaculture, which increasingly dominates seafood production as wild fishery production plateaus. For instance, marine diseases of farmed oysters, shrimp, abalone, and various fishes, particularly Atlantic salmon, cost billions of dollars each year. In comparison, it is often difficult to accurately estimate disease impacts on wild populations, especially those of pelagic and subtidal species. Farmed species often receive infectious diseases from wild species and can, in turn, export infectious agents to wild species. However, the impact of disease export on wild fisheries is controversial because there are few quantitative data demonstrating that wild species near farms suffer more from infectious diseases than those in other areas. The movement of exotic infectious agents to new areas continues to be the greatest concern.
Hydrogeologic data for the northern Rocky Mountains intermontane basins, Montana
Dutton, DeAnn M.; Lawlor, Sean M.; Briar, D.W.; Tresch, R.E.
1995-01-01
The U.S. Geological Survey began a Regional Aquifer- System Analysis of the Northern Rocky Mountains Intermontane Basins of western Montana and central and central and northern Idaho in 1990 to establish a regional framework of information for aquifers in 54 intermontane basins in an area of about 77,500 square miles. Selected hydrogeologic data have been used as part of this analysis to define the hydro- logic systems. Records of 1,376 wells completed in 31 of the 34 intermontane basins in the Montana part of the study area are tabulated in this report. Data consist of location, alttiude of land surface, date well constructed, geologic unit, depth of well, diameter of casing, type of finish, top of open interval, primary use of water, water level, date water level measured, discharge, specific capacity, source of discharge data, type of log available, date water-quality parameters measured, specific conductance, pH, and temperature. Hydrographs for selected wells also are included. Locations of wells and basins are shown on the accompanying plate.
Infectious Diseases Affect Marine Fisheries and Aquaculture Economics
NASA Astrophysics Data System (ADS)
Lafferty, Kevin D.; Harvell, C. Drew; Conrad, Jon M.; Friedman, Carolyn S.; Kent, Michael L.; Kuris, Armand M.; Powell, Eric N.; Rondeau, Daniel; Saksida, Sonja M.
2015-01-01
Seafood is a growing part of the economy, but its economic value is diminished by marine diseases. Infectious diseases are common in the ocean, and here we tabulate 67 examples that can reduce commercial species' growth and survivorship or decrease seafood quality. These impacts seem most problematic in the stressful and crowded conditions of aquaculture, which increasingly dominates seafood production as wild fishery production plateaus. For instance, marine diseases of farmed oysters, shrimp, abalone, and various fishes, particularly Atlantic salmon, cost billions of dollars each year. In comparison, it is often difficult to accurately estimate disease impacts on wild populations, especially those of pelagic and subtidal species. Farmed species often receive infectious diseases from wild species and can, in turn, export infectious agents to wild species. However, the impact of disease export on wild fisheries is controversial because there are few quantitative data demonstrating that wild species near farms suffer more from infectious diseases than those in other areas. The movement of exotic infectious agents to new areas continues to be the greatest concern.
An analysis of the magnitude and frequency of floods on Oahu, Hawaii
Nakahara, R.H.
1980-01-01
An analysis of available peak-flow data for the island of Oahu, Hawaii, was made by using multiple regression techniques which related flood-frequency data to basin and climatic characteristics for 74 gaging stations on Oahu. In the analysis, several different groupings of stations were investigated, including divisions by geographic location and size of drainage area. The grouping consisting of two leeward divisions and one windward division produced the best results. Drainage basins ranged in area from 0.03 to 45.7 square miles. Equations relating flood magnitudes of selected frequencies to basin characteristics were developed for the three divisions of Oahu. These equations can be used to estimate the magnitude and frequency of floods for any site, gaged or ungaged, for any desired recurrence interval from 2 to 100 years. Data on basin characteristics, flood magnitudes for various recurrence intervals from individual station-frequency curves, and computed flood magnitudes by use of the regression equation are tabulated to provide the needed data. (USGS)
Infectious diseases affect marine fisheries and aquaculture economics.
Lafferty, Kevin D; Harvell, C Drew; Conrad, Jon M; Friedman, Carolyn S; Kent, Michael L; Kuris, Armand M; Powell, Eric N; Rondeau, Daniel; Saksida, Sonja M
2015-01-01
Seafood is a growing part of the economy, but its economic value is diminished by marine diseases. Infectious diseases are common in the ocean, and here we tabulate 67 examples that can reduce commercial species' growth and survivorship or decrease seafood quality. These impacts seem most problematic in the stressful and crowded conditions of aquaculture, which increasingly dominates seafood production as wild fishery production plateaus. For instance, marine diseases of farmed oysters, shrimp, abalone, and various fishes, particularly Atlantic salmon, cost billions of dollars each year. In comparison, it is often difficult to accurately estimate disease impacts on wild populations, especially those of pelagic and subtidal species. Farmed species often receive infectious diseases from wild species and can, in turn, export infectious agents to wild species. However, the impact of disease export on wild fisheries is controversial because there are few quantitative data demonstrating that wild species near farms suffer more from infectious diseases than those in other areas. The movement of exotic infectious agents to new areas continues to be the greatest concern.
Yancey, Antronette K; Cole, Brian L; Brown, Rochelle; Williams, Jerome D; Hillier, Amy; Kline, Randolph S; Ashe, Marice; Grier, Sonya A; Backman, Desiree; McCarthy, William J
2009-01-01
Context: Commercial marketing is a critical but understudied element of the sociocultural environment influencing Americans' food and beverage preferences and purchases. This marketing also likely influences the utilization of goods and services related to physical activity and sedentary behavior. A growing literature documents the targeting of racial/ethnic and income groups in commercial advertisements in magazines, on billboards, and on television that may contribute to sociodemographic disparities in obesity and chronic disease risk and protective behaviors. This article examines whether African Americans, Latinos, and people living in low-income neighborhoods are disproportionately exposed to advertisements for high-calorie, low nutrient–dense foods and beverages and for sedentary entertainment and transportation and are relatively underexposed to advertising for nutritious foods and beverages and goods and services promoting physical activities. Methods: Outdoor advertising density and content were compared in zip code areas selected to offer contrasts by area income and ethnicity in four cities: Los Angeles, Austin, New York City, and Philadelphia. Findings: Large variations were observed in the amount, type, and value of advertising in the selected zip code areas. Living in an upper-income neighborhood, regardless of its residents' predominant ethnicity, is generally protective against exposure to most types of obesity-promoting outdoor advertising (food, fast food, sugary beverages, sedentary entertainment, and transportation). The density of advertising varied by zip code area race/ethnicity, with African American zip code areas having the highest advertising densities, Latino zip code areas having slightly lower densities, and white zip code areas having the lowest densities. Conclusions: The potential health and economic implications of differential exposure to obesity-related advertising are substantial. Although substantive legal questions remain about the government's ability to regulate advertising, the success of limiting tobacco advertising offers lessons for reducing the marketing contribution to the obesigenicity of urban environments. PMID:19298419
Sopite Syndrome in Operational Flight Training.
1998-09-01
1 B. PURPOSE ...................................................................... 3 C ...7 C . SOPITE SYNDROME ...................................................... 10 D. TREATMENT ASSESSMENT...25 C . DATA TABULATION ..................................................... 26 vii IV. RESULTS
Host status of Meyer and Eureka lemons for Anastrepha ludens.
Mangan, Robert L; Moreno, Aleena Tarshis
2012-04-01
Host status for Mexican fruit fly (Anastrepha ludens (Loew)) was examined under laboratory conditions in cage infested Eureka and Meyer lemons. Our approach was to allow females to oviposit on the two cultivars in separate laboratory cages with aluminum foil covering to restrict the areas where females had access to fruit surface. Fruit of each cultivar were placed in covered trays for incubations and at approximately weekly intervals, fruit were removed, dissected, and live and dead eggs and larvae tabulated in each tissue of the fruit. Infestation and survival were tabulated and analyzed for the effects of harvest date, fruit color and brix indices, postoviposition period, and cultivar. Infestation rate, determined by counts of total eggs and larvae was significantly higher in Meyer lemons. In both cultivars, females deposited eggs into both albedo and pulp tissue but not into flavedo. Both cultivars showed high resistance (> 90% mortality) to egg and first instars survival in albedo and pulp. Second and third instars surviving in the pulp had high survival rates (> 60%) in both cultivars in fruit dissected at weeks 2-4 after infestation. Total adults produced were slightly higher, and total second and third stage larvae were also higher for Meyer lemons. Numbers of adults and total second and third stage larvae increased in Eureka lemons in more mature fruit, but the higher numbers in Meyer lemons were not associated with fruit maturity, at time of infestation. Numbers of second and third stage larvae were significantly correlated with some fruit color indices in Eureka but not in Meyer lemons. Application of these results to quarantine risk analysis is discussed.
NASA Astrophysics Data System (ADS)
York, B. J.; Sinha, N.; Dash, S. M.; Hosangadi, A.; Kenzakowski, D. C.; Lee, R. A.
1992-07-01
The analysis of steady and transient aerodynamic/propulsive/plume flowfield interactions utilizing several state-of-the-art computer codes (PARCH, CRAFT, and SCHAFT) is discussed. These codes have been extended to include advanced turbulence models, generalized thermochemistry, and multiphase nonequilibrium capabilities. Several specialized versions of these codes have been developed for specific applications. This paper presents a brief overview of these codes followed by selected cases demonstrating steady and transient analyses of conventional as well as advanced missile systems. Areas requiring upgrades include turbulence modeling in a highly compressible environment and the treatment of particulates in general. Recent progress in these areas are highlighted.
Error control techniques for satellite and space communications
NASA Technical Reports Server (NTRS)
Costello, Daniel J., Jr.
1994-01-01
Brief summaries of research in the following areas are presented: (1) construction of optimum geometrically uniform trellis codes; (2) a statistical approach to constructing convolutional code generators; and (3) calculating the exact performance of a convolutional code.
Anigstein, Robert; Olsher, Richard H; Loomis, Donald A; Ansari, Armin
2016-12-01
The detonation of a radiological dispersion device or other radiological incidents could result in widespread releases of radioactive materials and intakes of radionuclides by affected individuals. Transportable radiation monitoring instruments could be used to measure radiation from gamma-emitting radionuclides in the body for triaging individuals and assigning priorities to their bioassay samples for in vitro assessments. The present study derived sets of calibration factors for four instruments: the Ludlum Model 44-2 gamma scintillator, a survey meter containing a 2.54 × 2.54-cm NaI(Tl) crystal; the Captus 3000 thyroid uptake probe, which contains a 5.08 × 5.08-cm NaI(Tl) crystal; the Transportable Portal Monitor Model TPM-903B, which contains two 3.81 × 7.62 × 182.9-cm polyvinyltoluene plastic scintillators; and a generic instrument, such as an ionization chamber, that measures exposure rates. The calibration factors enable these instruments to be used for assessing inhaled or ingested intakes of any of four radionuclides: Co, I, Cs, and Ir. The derivations used biokinetic models embodied in the DCAL computer software system developed by the Oak Ridge National Laboratory and Monte Carlo simulations using the MCNPX radiation transport code. The three physical instruments were represented by MCNP models that were developed previously. The affected individuals comprised children of five ages who were represented by the revised Oak Ridge National Laboratory pediatric phantoms, and adult men and adult women represented by the Adult Reference Computational Phantoms described in Publication 110 of the International Commission on Radiological Protection. These calibration factors can be used to calculate intakes; the intakes can be converted to committed doses by the use of tabulated dose coefficients. These calibration factors also constitute input data to the ICAT computer program, an interactive Microsoft Windows-based software package that estimates intakes of radionuclides and cumulative and committed effective doses, based on measurements made with these instruments. This program constitutes a convenient tool for assessing intakes and doses without consulting tabulated calibration factors and dose coefficients.
Anigstein, Robert; Olsher, Richard H.; Loomis, Donald A.; Ansari, Armin
2017-01-01
The detonation of a radiological dispersion device or other radiological incidents could result in widespread releases of radioactive materials and intakes of radionuclides by affected individuals. Transportable radiation monitoring instruments could be used to measure radiation from gamma-emitting radionuclides in the body for triaging individuals and assigning priorities to their bioassay samples for in vitro assessments. The present study derived sets of calibration factors for four instruments: the Ludlum Model 44-2 gamma scintillator, a survey meter containing a 2.54 × 2.54-cm NaI(Tl) crystal; the Captus 3000 thyroid uptake probe, which contains a 5.08 × 5.08-cm NaI(Tl) crystal; the Transportable Portal Monitor Model TPM-903B, which contains two 3.81 × 7.62 × 182.9-cm polyvinyltoluene plastic scintillators; and a generic instrument, such as an ionization chamber, that measures exposure rates. The calibration factors enable these instruments to be used for assessing inhaled or ingested intakes of any of four radionuclides: 60Co, 131I, 137Cs, and 192Ir. The derivations used biokinetic models embodied in the DCAL computer software system developed by the Oak Ridge National Laboratory and Monte Carlo simulations using the MCNPX radiation transport code. The three physical instruments were represented by MCNP models that were developed previously. The affected individuals comprised children of five ages who were represented by the revised Oak Ridge National Laboratory pediatric phantoms, and adult men and adult women represented by the Adult Reference Computational Phantoms described in Publication 110 of the International Commission on Radiological Protection. These calibration factors can be used to calculate intakes; the intakes can be converted to committed doses by the use of tabulated dose coefficients. These calibration factors also constitute input data to the ICAT computer program, an interactive Microsoft Windows-based software package that estimates intakes of radionuclides and cumulative and committed effective doses, based on measurements made with these instruments. This program constitutes a convenient tool for assessing intakes and doses without consulting tabulated calibration factors and dose coefficients. PMID:27798478
1985-06-02
was declared a few days later under the auspices of the guarantors of the Rio Protocol of 1942 (Argentina, Brazil, Chile and the USA). Further...Charge d’affaires: Marin Kostov. Canada: Edif. Belmonte 6, Avda Corea 126 y Amazonas, Wuito; tel, 458-102; Ambassador: (Vacant) Chile : Avda...Availability Status In 1861 adopted Civil Code of Chile - based on Napoleonic Code, Roman Code, Louisiana Code, the Austrian and Prussian Codes and Seven
Two related numerical codes, 3DFEMWATER and 3DLEWASTE, are presented sed to delineate wellhead protection areas in agricultural regions using the assimilative capacity criterion. DFEMWATER (Three-dimensional Finite Element Model of Water Flow Through Saturated-Unsaturated Media) ...
77 FR 59629 - Statutorily Mandated Designation of Difficult Development Areas for 2013
Federal Register 2010, 2011, 2012, 2013, 2014
2012-09-28
... Economist, Economic Development and Public Finance Division, Office of Policy Development and Research... evaluative list of metropolitan zip codes that would be designated Small Area DDAs using this methodology and... Research. [FR Doc. 2012-23900 Filed 9-27-12; 8:45 am] BILLING CODE 4210-67-P ...
The Social Interactive Coding System (SICS): An On-Line, Clinically Relevant Descriptive Tool.
ERIC Educational Resources Information Center
Rice, Mabel L.; And Others
1990-01-01
The Social Interactive Coding System (SICS) assesses the continuous verbal interactions of preschool children as a function of play areas, addressees, script codes, and play levels. This paper describes the 26 subjects and the setting involved in SICS development, coding definitions and procedures, training procedures, reliability, sample…
Villani, Melanie; Earnest, Arul; Smith, Karen; de Courten, Barbora; Zoungas, Sophia
2018-03-23
Geographical variation of diabetic emergencies attended by prehospital emergency medical services (EMS) and the relationship between area-level social and demographic factors and risk of a diabetic emergency were examined. All cases of hypoglycaemia and hyperglycaemia attended by Ambulance Victoria between 1/01/2009 and 31/12/2015 were tabulated by Local Government Area (LGA). Conditional autoregressive models were used to create smoothed maps of age and gender standardised incidence ratio (SIR) of prehospital EMS attendance for a diabetic emergency. Spatial regression models were used to examine the relationship between risk of a diabetic emergency and area-level factors. The areas with the greatest risk of prehospital EMS attendance for a diabetic emergency were disperse. Area-level factors associated with risk of a prehospital EMS-attended diabetic emergency were socioeconomic status (SIR 0.70 95% CrI [0.51, 0.96]), proportion of overseas-born residents (SIR 2.02 95% CrI [1.37, 2.91]) and motor vehicle access (SIR 1.47 95% CrI [1.08, 1.99]). Recognition of areas of increased risk of prehospital EMS-attended diabetic emergencies may be used to assist prehospital EMS resource planning to meet increased need. In addition, identification of associated factors can be used to target preventative interventions tailored to individual regions to reduce demand.
Variation in Drug Prices at Pharmacies: Are Prices Higher in Poorer Areas?
Gellad, Walid F; Choudhry, Niteesh K; Friedberg, Mark W; Brookhart, M Alan; Haas, Jennifer S; Shrank, William H
2009-01-01
Objective To determine whether retail prices for prescription drugs are higher in poorer areas. Data Sources The MyFloridarx.com website, which provides retail prescription prices at Florida pharmacies, and median ZIP code income from the 2000 Census. Study Design We compared mean pharmacy prices for each of the four study drugs across ZIP code income groups. Pharmacies were classified as either chain pharmacies or independent pharmacies. Data Collection Prices were downloaded in November 2006. Principal Findings Across the four study drugs, mean prices were highest in the poorest ZIP codes: 9 percent above the statewide average. Independent pharmacies in the poorest ZIP codes charged the highest mean prices. Conclusions Retail prescription prices appear to be higher in poorer ZIP codes of Florida. PMID:19178584
Overview of NASA Multi-dimensional Stirling Convertor Code Development and Validation Effort
NASA Technical Reports Server (NTRS)
Tew, Roy C.; Cairelli, James E.; Ibrahim, Mounir B.; Simon, Terrence W.; Gedeon, David
2002-01-01
A NASA grant has been awarded to Cleveland State University (CSU) to develop a multi-dimensional (multi-D) Stirling computer code with the goals of improving loss predictions and identifying component areas for improvements. The University of Minnesota (UMN) and Gedeon Associates are teamed with CSU. Development of test rigs at UMN and CSU and validation of the code against test data are part of the effort. The one-dimensional (1-D) Stirling codes used for design and performance prediction do not rigorously model regions of the working space where abrupt changes in flow area occur (such as manifolds and other transitions between components). Certain hardware experiences have demonstrated large performance gains by varying manifolds and heat exchanger designs to improve flow distributions in the heat exchangers. 1-D codes were not able to predict these performance gains. An accurate multi-D code should improve understanding of the effects of area changes along the main flow axis, sensitivity of performance to slight changes in internal geometry, and, in general, the understanding of various internal thermodynamic losses. The commercial CFD-ACE code has been chosen for development of the multi-D code. This 2-D/3-D code has highly developed pre- and post-processors, and moving boundary capability. Preliminary attempts at validation of CFD-ACE models of MIT gas spring and "two space" test rigs were encouraging. Also, CSU's simulations of the UMN oscillating-flow fig compare well with flow visualization results from UMN. A complementary Department of Energy (DOE) Regenerator Research effort is aiding in development of regenerator matrix models that will be used in the multi-D Stirling code. This paper reports on the progress and challenges of this
A table of semiempirical gf values. Part 2. Wavelengths: 272. 3395 nm to 599. 3892 nm
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kurucz, R.L.; Peytremann, E.
1975-02-14
The gf values for 265,587 atomic lines selectedfrom the line data used to calculate line blanketed model atmospheres are tabulated. These data are especially useful for line identification and spectral synthesis in solar and stellar spectra. The gf values are calculated semiempirically by using scaled Thomas--Fermi--Dirac radial wave functions and eigenvectors found through least-squares fits to observed energy levels. Included in the calculation are the first five or six stages of ionization for sequences up through nickel. Published gf values are included for elements heavier than nickel. The tabulation is restricted to lines with wavelengths less than 10 micrometers. (auth)
NASA Technical Reports Server (NTRS)
Jenkins, R. V.
1983-01-01
The tabulated data from tests of a six inch chord NPL 9510 airfoil in the Langley 0.3-Meter Transonic Cryogenic Tunnel. The tests were performed over the following range of conditions: Mach numbers of 0.35 to 0.82, total temperature of 94 K to 300 K, total pressure of 1.20 to 5.81 atm, Reynolds number based on chord of 1.34 x 10 to the 6th to 48.23 x 10 to the 6th, and angle of attack of 0 deg to 6 deg. The NPL 9510 airfoil was observed to have decreasing drag coefficient up to the highest test Reynolds number.
NASA Technical Reports Server (NTRS)
Cannon, Michael D.
1956-01-01
This paper contains tail and hull loads data obtained in an investigation of a l/15-scale model of the Goodyear XZP5K airship. Data are presented in the form of tabulated pressure coefficients over a pitch and yaw range of +/-20 deg and 0 deg to 30 deg respectively, with various rudder and elevator deflections. Two tail configurations of different plan forms were tested on the model. The investigation was conducted in the Langley full-scale tunnel at a Reynolds number of approximately 16.5 x 10(exp 6) based on hull length, which corresponds to a Mach number of about 0.12.
Tabulation and summary of thermodynamic effects data for developed cavitation on ogive-nosed bodies
NASA Technical Reports Server (NTRS)
Holl, J. W.; Billet, M. L.; Weir, D. S.
1978-01-01
Thermodynamic effects data for developed cavitation on zero and quarter caliber ogives in Freon 113 and water are tabulated and summarized. These data include temperature depression (delta T), flow coefficient (C sub Q), and various geometrical characteristics of the cavity. For the delta T tests, the free-stream temperature varied from 35 C to 95 C in Freon 113 and from 60 C to 125 C in water for a velocity range of 19.5 m/sec to 36.6 m/sec. Two correlations of the delta T data by the entrainment method are presented. These correlations involve different combinations of the Nusselt, Reynolds, Froude, Weber, and Peclet numbers and dimensionless cavity length.
Atomic and ionic spectrum lines below 2000A: hydrogen through argon
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kelly, R.L.
1982-10-01
A critical tabulation of observed spectral lines below 2000 angstroms has been prepared from the published literature up to July 1978. It is intended principally as an aid to those physicists and astronomers who deal with the spectra of highly stripped atoms. This report includes the first 18 elements, from hydrogen (including deuterium) through argon. The tabulation is divided into two main sections: the spectrum lines by spectrum, and a finding list. The entries for each element give the ionization species, ground state term, and ionization potential, as well as the best values of vacuum wavelength, intensity, and classification. Amore » list of the pertinent references is appended at the end.« less
Anthropometric survey of the astronaut applicants and astronauts from 1985 to 1991
NASA Technical Reports Server (NTRS)
Rajulu, Sudhakar L.; Klute, Glenn K.
1993-01-01
The Anthropometry and Biomechanics Laboratory at the Johnson Space Center has been collecting anthropometric data from astronaut applicants since 1977. These anthropometric measurements had been taken from 473 applicants. Based on the position they applied for, these applicants were classified as either mission specialists, payload specialists, pilots, or observers. The main objective was to document the variations among these applicants and tabulate the percentile data for each anthropometric dimension. The percentile and the descriptive statistics data were tabulated and graphed for the whole astronaut candidate population; for the male and female groups; for each subject classification such as pilot, mission specialist, and payload specialist; and finally, for those who were selected as astronauts.
XSECT: A computer code for generating fuselage cross sections - user's manual
NASA Technical Reports Server (NTRS)
Ames, K. R.
1982-01-01
A computer code, XSECT, has been developed to generate fuselage cross sections from a given area distribution and wing definition. The cross sections are generated to match the wing definition while conforming to the area requirement. An iterative procedure is used to generate each cross section. Fuselage area balancing may be included in this procedure if desired. The code is intended as an aid for engineers who must first design a wing under certain aerodynamic constraints and then design a fuselage for the wing such that the contraints remain satisfied. This report contains the information necessary for accessing and executing the code, which is written in FORTRAN to execute on the Cyber 170 series computers (NOS operating system) and produces graphical output for a Tektronix 4014 CRT. The LRC graphics software is used in combination with the interface between this software and the PLOT 10 software.
Hierarchical differences in population coding within auditory cortex.
Downer, Joshua D; Niwa, Mamiko; Sutter, Mitchell L
2017-08-01
Most models of auditory cortical (AC) population coding have focused on primary auditory cortex (A1). Thus our understanding of how neural coding for sounds progresses along the cortical hierarchy remains obscure. To illuminate this, we recorded from two AC fields: A1 and middle lateral belt (ML) of rhesus macaques. We presented amplitude-modulated (AM) noise during both passive listening and while the animals performed an AM detection task ("active" condition). In both fields, neurons exhibit monotonic AM-depth tuning, with A1 neurons mostly exhibiting increasing rate-depth functions and ML neurons approximately evenly distributed between increasing and decreasing functions. We measured noise correlation ( r noise ) between simultaneously recorded neurons and found that whereas engagement decreased average r noise in A1, engagement increased average r noise in ML. This finding surprised us, because attentive states are commonly reported to decrease average r noise We analyzed the effect of r noise on AM coding in both A1 and ML and found that whereas engagement-related shifts in r noise in A1 enhance AM coding, r noise shifts in ML have little effect. These results imply that the effect of r noise differs between sensory areas, based on the distribution of tuning properties among the neurons within each population. A possible explanation of this is that higher areas need to encode nonsensory variables (e.g., attention, choice, and motor preparation), which impart common noise, thus increasing r noise Therefore, the hierarchical emergence of r noise -robust population coding (e.g., as we observed in ML) enhances the ability of sensory cortex to integrate cognitive and sensory information without a loss of sensory fidelity. NEW & NOTEWORTHY Prevailing models of population coding of sensory information are based on a limited subset of neural structures. An important and under-explored question in neuroscience is how distinct areas of sensory cortex differ in their population coding strategies. In this study, we compared population coding between primary and secondary auditory cortex. Our findings demonstrate striking differences between the two areas and highlight the importance of considering the diversity of neural structures as we develop models of population coding. Copyright © 2017 the American Physiological Society.
Video coding for 3D-HEVC based on saliency information
NASA Astrophysics Data System (ADS)
Yu, Fang; An, Ping; Yang, Chao; You, Zhixiang; Shen, Liquan
2016-11-01
As an extension of High Efficiency Video Coding ( HEVC), 3D-HEVC has been widely researched under the impetus of the new generation coding standard in recent years. Compared with H.264/AVC, its compression efficiency is doubled while keeping the same video quality. However, its higher encoding complexity and longer encoding time are not negligible. To reduce the computational complexity and guarantee the subjective quality of virtual views, this paper presents a novel video coding method for 3D-HEVC based on the saliency informat ion which is an important part of Human Visual System (HVS). First of all, the relationship between the current coding unit and its adjacent units is used to adjust the maximum depth of each largest coding unit (LCU) and determine the SKIP mode reasonably. Then, according to the saliency informat ion of each frame image, the texture and its corresponding depth map will be divided into three regions, that is, salient area, middle area and non-salient area. Afterwards, d ifferent quantization parameters will be assigned to different regions to conduct low complexity coding. Finally, the compressed video will generate new view point videos through the renderer tool. As shown in our experiments, the proposed method saves more bit rate than other approaches and achieves up to highest 38% encoding time reduction without subjective quality loss in compression or rendering.
NASA Technical Reports Server (NTRS)
Peterson, Victor L.; Menees, Gene P.
1961-01-01
Tabulated results of a wind-tunnel investigation of the aerodynamic loads on a canard airplane model with a single vertical tail are presented for Mach numbers from 0.70 to 2.22. The Reynolds number for the measurements was 2.9 x 10(exp 6) based on the wing mean aerodynamic chord. The results include local static pressure coefficients measured on the wing, body, and vertical tail for angles of attack from -4 deg to + 16 deg, angles of sideslip of 0 deg and 5.3 deg, vertical-tail settings of 0 deg and 5 deg, and nominal canard deflections of 0 deg and 10 deg. Also included are section force and moment coefficients obtained from integrations of the local pressures and model-component force and moment coefficients obtained from integrations of the section coefficients. Geometric details of the model and the locations of the pressure orifices are shown. An index to the data contained herein is presented and definitions of nomenclature are given.
VizieR Online Data Catalog: Historical and HST Astrometry of Sirius A,B (Bond+, 2017)
NASA Astrophysics Data System (ADS)
Bond, H. E.; Schaefer, G. H.; Gilliland, R. L.; Holberg, J. B.; Mason, B. D.; Lindenblad, I. W.; Seitz-McLeese, M.; Arnett, W. D.; Demarque, P.; Spada, F.; Young, P. A.; Barstow, M. A.; Burleigh, M. R.; Gudehus, D.
2017-05-01
We have assembled a compilation of published historical measurements of the position angle (PA) and the angular separation of Sirius B relative to Sirius A. Our tabulation is based on a critical review of measures contained in the Washington Double Star Catalog maintained at the USNO and from our additional literature searches. Notes included in the tabulation give extensive commentary on the historical observations. Many early publications provided measures averaged over multiple nights or even an entire observing season for the purpose of reducing computational labor in subsequent analyses. With modern computers, there is no need for such averaging, so we opted to present the individual measures whenever available. However, if an observer reported more than one measurement on a given night, we did compute the mean position for that night. If the original publication only reported a mean across several nights, we tabulated that mean as reported. The visual micrometer observations did not always include a contemporaneous measurement of both the PA and separation. These omissions are listed as -99.0 in the table. The measurement uncertainties were assigned through our orbital fitting method described in the paper. Measurements that were rejected from the orbital solution are identified in the Notes column and are listed with uncertainties of 0. (3 data files).
Schiff, Elad; Ben-Arye, Eran; Shilo, Margalit; Levy, Moti; Schachter, Leora; Weitchner, Na'ama; Golan, Ofra; Stone, Julie
2011-02-01
Recently, ethical guidelines regarding safe touch in CAM were developed in Israel. Publishing ethical codes does not imply that they will actually help practitioners to meet ethical care standards. The effectiveness of ethical rules depends on familiarity with the code and its content. In addition, critical self-examination of the code by individual members of the profession is required to reflect on the moral commitments encompassed in the code. For the purpose of dynamic self-appraisal, we devised a survey to assess how CAM practitioners view the suggested ethical guidelines for safe touch. We surveyed 781 CAM practitioners regarding their perspectives on the safe-touch code. There was a high level of agreement with general statements regarding ethics pertaining to safe touch with a mean rate of agreement of 4.61 out of a maximum of 5. Practitioners concurred substantially with practice guidelines for appropriate touch with a mean rate of agreement of 4.16 out of a maximum of 5. Attitudes toward the necessity to touch intimate areas for treatment purposes varied with 78.6% of respondents strongly disagreeing with any notion of need to touch intimate areas during treatment. 7.9% neither disagreed nor agreed, 7.9% slightly agreed, and 7.6% strongly agreed with the need for touching intimate areas during treatment. There was a direct correlation between disagreement with touching intimate areas for therapeutic purposes and agreement with general statements regarding ethics of safe touch (Spearman r=0.177, p<0.0001), and practice guidelines for appropriate touch (r=0.092, p=0.012). A substantial number of practitioners agreed with the code, although some findings regarding the need to touch intimate area during treatments were disturbing. Our findings can serve as a basis for ethical code development and implementation, as well as for educating CAM practitioners on the ethics of touch. Copyright © 2010 Elsevier Ltd. All rights reserved.
Standard interface files and procedures for reactor physics codes, version III
DOE Office of Scientific and Technical Information (OSTI.GOV)
Carmichael, B.M.
Standards and procedures for promoting the exchange of reactor physics codes are updated to Version-III status. Standards covering program structure, interface files, file handling subroutines, and card input format are included. The implementation status of the standards in codes and the extension of the standards to new code areas are summarized. (15 references) (auth)
Code of Federal Regulations, 2010 CFR
2010-07-01
..., maps, graphs, pamphlets, notes, charts, tabulations, analyses, statistical or informational... Office. Official business means the authorized business of the Office. General Counsel means the General...
Census data for studying elderly populations.
Martin, L G
1987-06-01
As fertility and population growth rates decline in the Asian and Pacific region, populations of the elderly (aged 65 and older) are growing rapidly. In 1980, the number of elderly was roughly equally divided between more developed and less developed countries; in the future, the majority of the elderly will live in less developed countries, especially in Asia. Most countries will double their elderly populations between 1980 and 2000. Censuses can provide information for planning ways to accommodate the growing numbers of elderly. Most censuses in the Asian and Pacific region provide population counts to fairly high ages. The elderly should not be stereotyped by assuming that they are all alike physically, emotionally, socially, and politically. Except for Indonesia and the Republic of Korea, all the censuses lower the last age for which provide detail in their published tabulations of marital status data compared to tabulations of population counts. Policy makers are more concerned with the extent to which the generations live together. Censuses potentially are a rich source of data on living arrangements, but, except for Japan, that potential is not being realized. Various types of information on the living arrangements of the elderly are contained in several censuses. An example is tabulations by 5-year groups to 85+ years of individuals by their relationship to the household head and by family type. It is important for the government to ascertain the extent to which some elderly may need some public assistance. Most of the censuses provide information on labor force participation to ages 60 or 65 and above. Censuses also provide information on income. More information will be needed on elderly in the future and much of it can be made available by increasing the amount of detail, by age, in published tabulations. Censuses can provide a regular, reliable flow of data and information that will form the foundation for designing programs for the expanding elderly population of Asia and the Pacific.
Thermodynamic Data for Fifty Reference Elements
NASA Technical Reports Server (NTRS)
McBride, Bonnie J.; Gordon, Sanford; Reno, Martin A.
2001-01-01
This report is a compilation of thermodynamic functions of 50 elements in their reference state. The functions are: C(sup 0, sub p), {H(T)-H(sup 0)(0)}, S(sup 0)(T), and - {G(sup 0)(T) - H(sup 0)(0)} for the elements Ag, Al, Ar, B, Ba, Be, Br2, C, Ca, Cd, Cl2, Co, Cr, Cs, Cu, F2, Fe, Ge, H2, He, Hg, I2, K, Kr, Li, Mg, Mn, Mo, N2, Na, Nb, Ne, Ni, O2, P, Pb, Rb, S, Si, Sn, Sr, Th, Th, Ti, U, V, W, Xe, Zn, and Zr. Deuterium D, and electron gas e(sup -) are also included. The data are tabulated as functions of temperature as well as given in the form of least-squares coefficients for two functional forms for C(sup 0, sub p) with integration constants for enthalpy and entropy. One functional form for C(sup 0, sub p) is a fourth-order polynomial and the other has two additional terms, one with T(sup -1) and the other with T(sup -2). The gases Ar, D2, e(sup -), H2, He, Kr, N2, Ne, O2, and Xe are tabulated for temperatures from 100 to 20 000 K. The remaining gases Cl2 and F2 are tabulated from 100 to 6000 K. The polynomial functional form for C(sup 0, sub p) for all these gases is split into two temperature intervals of 200 to 1000 K and 1000 to 6000 K. The second functional form for (sup 0, sub p) has an additional interval from 6000 to 20 000 K for the gases tabulated to 20 000 K. The fits are constrained so that the properties match at the common temperature endpoints. The temperature ranges for the condensed species vary with range of the data, phase changes, and shapes of the C(sup 0, sub p) curves.
NASA Technical Reports Server (NTRS)
Campbell, Bryan A.; Applin, Zachary T.; Kemmerly, Guy T.; Coe, Paul L., Jr.; Owens, D. Bruce; Gile, Brenda E.; Parikh, Pradip G.; Smith, Don
1999-01-01
A wind tunnel investigation of a leading edge boundary layer control system was conducted on a High Speed Civil Transport (HSCT) configuration in the Langley 14- by 22-Foot Subsonic Tunnel. Data were obtained over a Mach number range of 0.08 to 0.27, with corresponding chord Reynolds numbers of 1.79 x 10(exp 6) to 5.76 x 10(exp 6). Variations in the amount of suction, as well as the size and location of the suction area, were tested with outboard leading edge flaps deflected 0 and 30 deg and trailing-edge flaps deflected 0 and 20 deg. The longitudinal and lateral aerodynamic data are presented without analysis. A complete tabulated data listing is also presented herein.
SYNMAG PHOTOMETRY: A FAST TOOL FOR CATALOG-LEVEL MATCHED COLORS OF EXTENDED SOURCES
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bundy, Kevin; Yasuda, Naoki; Hogg, David W.
2012-12-01
Obtaining reliable, matched photometry for galaxies imaged by different observatories represents a key challenge in the era of wide-field surveys spanning more than several hundred square degrees. Methods such as flux fitting, profile fitting, and PSF homogenization followed by matched-aperture photometry are all computationally expensive. We present an alternative solution called 'synthetic aperture photometry' that exploits galaxy profile fits in one band to efficiently model the observed, point-spread-function-convolved light profile in other bands and predict the flux in arbitrarily sized apertures. Because aperture magnitudes are the most widely tabulated flux measurements in survey catalogs, producing synthetic aperture magnitudes (SYNMAGs) enablesmore » very fast matched photometry at the catalog level, without reprocessing imaging data. We make our code public and apply it to obtain matched photometry between Sloan Digital Sky Survey ugriz and UKIDSS YJHK imaging, recovering red-sequence colors and photometric redshifts with a scatter and accuracy as good as if not better than FWHM-homogenized photometry from the GAMA Survey. Finally, we list some specific measurements that upcoming surveys could make available to facilitate and ease the use of SYNMAGs.« less
Simultaneous storage of medical images in the spatial and frequency domain: a comparative study.
Nayak, Jagadish; Bhat, P Subbanna; Acharya U, Rajendra; Uc, Niranjan
2004-06-05
Digital watermarking is a technique of hiding specific identification data for copyright authentication. This technique is adapted here for interleaving patient information with medical images, to reduce storage and transmission overheads. The patient information is encrypted before interleaving with images to ensure greater security. The bio-signals are compressed and subsequently interleaved with the image. This interleaving is carried out in the spatial domain and Frequency domain. The performance of interleaving in the spatial, Discrete Fourier Transform (DFT), Discrete Cosine Transform (DCT) and Discrete Wavelet Transform (DWT) coefficients is studied. Differential pulse code modulation (DPCM) is employed for data compression as well as encryption and results are tabulated for a specific example. It can be seen from results, the process does not affect the picture quality. This is attributed to the fact that the change in LSB of a pixel changes its brightness by 1 part in 256. Spatial and DFT domain interleaving gave very less %NRMSE as compared to DCT and DWT domain. The Results show that spatial domain the interleaving, the %NRMSE was less than 0.25% for 8-bit encoded pixel intensity. Among the frequency domain interleaving methods, DFT was found to be very efficient.
Portrayals of food practices and exercise behavior in popular American films.
Bell, Robert A; Berger, Charles R; Cassady, Diana; Townsend, Marilyn S
2005-01-01
To describe depictions of food, alcohol, and exercise and sport in popular films. Content analysis of the 10 top-grossing films each year from 1991 to 2000 (N = 100 films). Coding reliabilities were assessed with Cohen's kappa. Research questions were addressed with basic descriptive statistics. Associations among variables were examined through a cross-tabulation procedure that corrects for the clustering of exercise and food depictions within movies. Food and drink appeared regularly in the films analyzed, typically as a background element or prop. On average, food appeared on the screen once every 4.2 minutes. The foods shown were most likely to be fats, oils, and sweets; fruits, vegetables, and dairy products were rarely seen. More than 20% of the food items shown were alcoholic beverages, which were nearly 2 times more likely to be ingested as nonalcohol food items. Planned exercise and sports appeared at a rate of about 2 incidents per hour. These activities rarely received explicit evaluation by characters. Films frequently portray foods and exercise. These depictions have the potential to cultivate beliefs about normative eating and exercise behavior, suggesting a need for public health professionals to encourage the media to offer healthier images.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Murray, R.C.
1991-09-01
Policy for addressing natural phenomenon comprises a hierarchy of interrelated documents. The top level of policy is contained in the code of Federal Regulations which establishes the framework and intent to ensure overall safety of DOE facilities when subjected to the effects of natural phenomena. The natural phenomena to be considered include earthquakes and tsunami, winds, hurricanes and tornadoes, floods, volcano effects and seiches. Natural phenomena criteria have been established for design of new facilities; evaluation of existing facilities; additions, modifications, and upgrades to existing facilities; and evaluation criteria for new or existing sites. Steps needed to implement these fourmore » general criteria are described. The intent of these criteria is to identify WHAT needs to be done to ensure adequate protection from natural phenomena. The commentary provides discussion of WHY this is needed for DOE facilities within the complex. Implementing procedures identifying HOW to carry out these criteria are next identified. Finally, short and long term tasks needed to identify the implementing procedure are tabulated. There is an overall need for consistency throughout the DOE complex related to natural phenomena including consistent terminology, policy, and implementation. 1 fig, 6 tabs.« less
NASA Astrophysics Data System (ADS)
Kaur, Jagjit; Gorczyca, T. W.; Badnell, N. R.
2018-02-01
Context. We aim to present a comprehensive theoretical investigation of dielectronic recombination (DR) of the silicon-like isoelectronic sequence and provide DR and radiative recombination (RR) data that can be used within a generalized collisional-radiative modelling framework. Aims: Total and final-state level-resolved DR and RR rate coefficients for the ground and metastable initial levels of 16 ions between P+ and Zn16+ are determined. Methods: We carried out multi-configurational Breit-Pauli DR calculations for silicon-like ions in the independent processes, isolated resonance, distorted wave approximation. Both Δnc = 0 and Δnc = 1 core excitations are included using LS and intermediate coupling schemes. Results: Results are presented for a selected number of ions and compared to all other existing theoretical and experimental data. The total dielectronic and radiative recombination rate coefficients for the ground state are presented in tabulated form for easy implementation into spectral modelling codes. These data can also be accessed from the Atomic Data and Analysis Structure (ADAS) OPEN-ADAS database. This work is a part of an assembly of a dielectronic recombination database for the modelling of dynamic finite-density plasmas.
Intenzo, Charles M; Parker, Laurence; Levin, David C; Kim, Sung M; Rao, Vijay M
2016-01-01
Both radiologists as well as nonimaging physicians perform dual-energy X-ray absorptiometry (DXA) imaging in the United States. This study aims to compare provider distribution between these physician groups on the Medicare population, which is the predominant age group of patients evaluated by this imaging procedure. Using the 2 relevant Current Procedural Terminology, Fourth Edition codes for DXA scans, source data were obtained from the CMS Physician Supplier Procedure Summary Master Files from 2003 through 2013. DXA scan procedure volumes for radiologists and nonradiologists on Medicare patients were tabulated. Utilization rates were calculated. From 2003 to 2013, the total number of DXA scans performed on Medicare patients decreased by 2%. However, over the same period, the number of scans performed by radiologists had increased by 25% over nonimaging specialists, whose utilization had declined by approximately the same amount. From 2003 to 2013, the rate of utilization of DXA scans in the Medicare fee-for-service population declined somewhat. However, radiologists continue to gain market share from other specialists and now predominate in this type of imaging by a substantial margin. Copyright © 2016 International Society for Clinical Densitometry. Published by Elsevier Inc. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rose, Amy N.; Nagle, Nicholas N.
Techniques such as Iterative Proportional Fitting have been previously suggested as a means to generate new data with the demographic granularity of individual surveys and the spatial granularity of small area tabulations of censuses and surveys. This article explores internal and external validation approaches for synthetic, small area, household- and individual-level microdata using a case study for Bangladesh. Using data from the Bangladesh Census 2011 and the Demographic and Health Survey, we produce estimates of infant mortality rate and other household attributes for small areas using a variation of an iterative proportional fitting method called P-MEDM. We conduct an internalmore » validation to determine: whether the model accurately recreates the spatial variation of the input data, how each of the variables performed overall, and how the estimates compare to the published population totals. We conduct an external validation by comparing the estimates with indicators from the 2009 Multiple Indicator Cluster Survey (MICS) for Bangladesh to benchmark how well the estimates compared to a known dataset which was not used in the original model. The results indicate that the estimation process is viable for regions that are better represented in the microdata sample, but also revealed the possibility of strong overfitting in sparsely sampled sub-populations.« less
Rose, Amy N.; Nagle, Nicholas N.
2016-08-01
Techniques such as Iterative Proportional Fitting have been previously suggested as a means to generate new data with the demographic granularity of individual surveys and the spatial granularity of small area tabulations of censuses and surveys. This article explores internal and external validation approaches for synthetic, small area, household- and individual-level microdata using a case study for Bangladesh. Using data from the Bangladesh Census 2011 and the Demographic and Health Survey, we produce estimates of infant mortality rate and other household attributes for small areas using a variation of an iterative proportional fitting method called P-MEDM. We conduct an internalmore » validation to determine: whether the model accurately recreates the spatial variation of the input data, how each of the variables performed overall, and how the estimates compare to the published population totals. We conduct an external validation by comparing the estimates with indicators from the 2009 Multiple Indicator Cluster Survey (MICS) for Bangladesh to benchmark how well the estimates compared to a known dataset which was not used in the original model. The results indicate that the estimation process is viable for regions that are better represented in the microdata sample, but also revealed the possibility of strong overfitting in sparsely sampled sub-populations.« less
Impact of changes to policy for Mexican risk travel on Canadian blood donor deferrals.
O'Brien, Sheila F; Uzicanin, Samra; Choquet, Karine; Yi, Qi-Long; Fan, Wenli; Goldman, Mindy
2013-10-01
Travel to malaria risk areas such as Mexico is a common source of donor deferral in Canada. On February 21st, 2011 the deferrable regions in Mexico were revised to permit donation if donors travelled to the state of Quintana Roo, Mexico, a popular ocean-side resort area. Canadian travel data and malaria deferral rates since 2007 were plotted to examine trends. Deferral records in one centre were accessed from January to April, 2011 to tabulate travel destinations of deferred donors immediately before and after the change. Travel to Mexico and the Caribbean accounts for 63% of general population travel, and travel to Mexico has been increasing (P <0.05). Deferral for short-term malaria risk travel has a strong seasonal trend with peaks in the winter and troughs in the summer. Approximately 36,000 fewer donations were lost following the change, a reduction of 37% from the previous year. Deferrals in one centre increased for Caribbean/Central America after the change (P <0.05) consistent with the seasonal trend, but decreased for Mexico (P <0.05). Deferrals for malaria risk travel are substantial. Careful revision and refinement of risk areas of travel can significantly reduce the burden of deferral.
Harada, K Y; Silva, J G; Schenkman, S; Hayama, E T; Santos, F R; Prado, M C; Pontes, R H
1999-01-07
The drawing up of adequate Public Health action planning to address the true needs of the population would increase the chances of effectiveness and decrease unnecessary expenses. To identify homogeneous regions in the UNIFESP/EPM healthcare center (HCC) coverage area based on sociodemographic indicators and to relate them to causes of deaths in 1995. Secondary data analysis. HCC coverage area; primary care. Sociodemographic indicators were obtained from special tabulations of the Demographic Census of 1991. Proportion of children and elderly in the population; family providers' education level (maximum: > 15 years, minimum: < 1 year) and income level (maximum: > 20 minimum wages, minimum: < 1 minimum wage); proportional mortality distribution The maximum income permitted the construction of four homogeneous regions, according to income ranking. Although the proportion of children and of elderly did not vary significantly among the regions, minimum income and education showed a statistically significant (p < 0.05) difference between the first region (least affluent) and the others. A clear trend of increasing maximum education was observed across the regions. Mortality also differed in the first region, with deaths generated by possibly preventable infections. The inequalities observed may contribute to primary health prevention.
Ramli, A T; Apriantoro, N H; Heryansyah, A; Basri, N A; Sanusi, M S M; Abu Hanifah, N Z H
2016-03-01
An extensive terrestrial gamma radiation dose (TGRD) rate survey has been conducted in Perak State, Peninsular Malaysia. The survey has been carried out taking into account geological and soil information, involving 2930 in situ surveys. Based on geological and soil information collected during TGRD rate measurements, TGRD rates have been predicted in Perak State using a statistical regression analysis which would be helpful to focus surveys in areas that are difficult to access. An equation was formulated according to a linear relationship between TGRD rates, geological contexts and soil types. The comparison of in situ measurements and predicted TGRD dose rates was tabulated and showed good agreement with the linear regression equation. The TGRD rates in the study area ranged from 38 nGy h(-1) to 1039 nGy h(-1) with a mean value of 224 ± 138 nGy h(-1). This value is higher than the world average as reported in UNSCEAR 2000. The TGRD rates contribute an average dose rate of 1.37 mSv per year. An isodose map for the study area was developed using a Kriging method based on predicted and in situ TGRD rate values.
Biro, Suzanne; Williamson, Tyler; Leggett, Jannet Ann; Barber, David; Morkem, Rachael; Moore, Kieran; Belanger, Paul; Mosley, Brian; Janssen, Ian
2016-03-11
Electronic medical records (EMRs) used in primary care contain a breadth of data that can be used in public health research. Patient data from EMRs could be linked with other data sources, such as a postal code linkage with Census data, to obtain additional information on environmental determinants of health. While promising, successful linkages between primary care EMRs with geographic measures is limited due to ethics review board concerns. This study tested the feasibility of extracting full postal code from primary care EMRs and linking this with area-level measures of the environment to demonstrate how such a linkage could be used to examine the determinants of disease. The association between obesity and area-level deprivation was used as an example to illustrate inequalities of obesity in adults. The analysis included EMRs of 7153 patients aged 20 years and older who visited a single, primary care site in 2011. Extracted patient information included demographics (date of birth, sex, postal code) and weight status (height, weight). Information extraction and management procedures were designed to mitigate the risk of individual re-identification when extracting full postal code from source EMRs. Based on patients' postal codes, area-based deprivation indexes were created using the smallest area unit used in Canadian censuses. Descriptive statistics and socioeconomic disparity summary measures of linked census and adult patients were calculated. The data extraction of full postal code met technological requirements for rendering health information extracted from local EMRs into anonymized data. The prevalence of obesity was 31.6 %. There was variation of obesity between deprivation quintiles; adults in the most deprived areas were 35 % more likely to be obese compared with adults in the least deprived areas (Chi-Square = 20.24(1), p < 0.0001). Maps depicting spatial representation of regional deprivation and obesity were created to highlight high risk areas. An area based socio-economic measure was linked with EMR-derived objective measures of height and weight to show a positive association between area-level deprivation and obesity. The linked dataset demonstrates a promising model for assessing health disparities and ecological factors associated with the development of chronic diseases with far reaching implications for informing public health and primary health care interventions and services.
ERIC Educational Resources Information Center
Rhodes, Gillian; Jeffery, Linda; Boeing, Alexandra; Calder, Andrew J.
2013-01-01
Despite the discovery of body-selective neural areas in occipitotemporal cortex, little is known about how bodies are visually coded. We used perceptual adaptation to determine how body identity is coded. Brief exposure to a body (e.g., anti-Rose) biased perception toward an identity with opposite properties (Rose). Moreover, the size of this…
24 CFR 200.926 - Minimum property standards for one and two family dwellings.
Code of Federal Regulations, 2012 CFR
2012-04-01
... property is to be located. (c) Standard for evaluating local or state building codes. The Secretary shall compare a local building code submitted under paragraph (d) of this section or a State code to the list of... each area and subarea on the list. (2) A State or local building code will be partially accepted if it...
24 CFR 200.926 - Minimum property standards for one and two family dwellings.
Code of Federal Regulations, 2014 CFR
2014-04-01
... property is to be located. (c) Standard for evaluating local or state building codes. The Secretary shall compare a local building code submitted under paragraph (d) of this section or a State code to the list of... each area and subarea on the list. (2) A State or local building code will be partially accepted if it...
24 CFR 200.926 - Minimum property standards for one and two family dwellings.
Code of Federal Regulations, 2013 CFR
2013-04-01
... property is to be located. (c) Standard for evaluating local or state building codes. The Secretary shall compare a local building code submitted under paragraph (d) of this section or a State code to the list of... each area and subarea on the list. (2) A State or local building code will be partially accepted if it...