Test Problem: Tilted Rayleigh-Taylor for 2-D Mixing Studies
Andrews, Malcolm J.; Livescu, Daniel; Youngs, David L.
2012-08-14
reasonable quality photographic data. The photographs in Figure 2 also reveal the appearance of a boundary layer at the left and right walls; this boundary layer has not been included in the test problem as preliminary calculations suggested it had a negligible effect on plume penetration and RT mixing. The significance of this test problem is that, unlike planar RT experiments such as the Rocket-Rig (Youngs, 1984), Linear Electric Motor - LEM (Dimonte, 1990), or the Water Tunnel (Andrews, 1992), the Tilted-Rig is a unique two-dimensional RT mixing experiment that has experimental data and now (in this TP) Direct Numerical Simulation data from Livescu and Wei. The availability of DNS data for the tilted-rig has made this TP viable as it provides detailed results for comparison purposes. The purpose of the test problem is to provide 3D simulation results, validated by comparison with experiment, which can be used for the development and validation of 2D RANS models. When such models are applied to 2D flows, various physics issues are raised such as double counting, combined buoyancy and shear, and 2-D strain, which have not yet been adequately addressed. The current objective of the test problem is to compare key results, which are needed for RANS model validation, obtained from high-Reynolds number DNS, high-resolution ILES or LES with explicit sub-grid-scale models. The experiment is incompressible and so is directly suitable for algorithms that are designed for incompressible flows (e.g. pressure correction algorithms with multi-grid); however, we have extended the TP so that compressible algorithms, run at low Mach number, may also be used if careful consideration is given to initial pressure fields. Thus, this TP serves as a useful tool for incompressible and compressible simulation codes, and mathematical models. In the remainder of this TP we provide a detailed specification; the next section provides the underlying assumptions for the TP, fluids, geometry details
A 2-D Test Problem for CFD Modeling Heat Transfer in Spent Fuel Transfer Cask Neutron Shields
Zigh, Ghani; Solis, Jorge; Fort, James A.
2011-01-14
well as the tradeoff between steady state and transient solutions. Solutions are compared for two commercial CFD codes, FLUENT and STAR-CCM+. The results can be used to provide input to the CFD Best Practices for this application. Following study results for the 2-D test problem, a comparison of simulation results is provided for a high Rayleigh number experiment with large annular gap. Because the geometry of this validation is significantly different from the neutron shield, and due to the critical nature of this application, the argument is made for new experiments at representative scales
On 2D bisection method for double eigenvalue problems
Ji, X.
1996-06-01
The two-dimensional bisection method presented in (SIAM J. Matrix Anal. Appl. 13(4), 1085 (1992)) is efficient for solving a class of double eigenvalue problems. This paper further extends the 2D bisection method of full matrix cases and analyses its stability. As in a single parameter case, the 2D bisection method is very stable for the tridiagonal matrix triples satisfying the symmetric-definite condition. Since the double eigenvalue problems arise from two-parameter boundary value problems, an estimate of the discretization error in eigenpairs is also given. Some numerical examples are included. 42 refs., 1 tab.
Validation and testing of the VAM2D computer code
Kool, J.B.; Wu, Y.S. )
1991-10-01
This document describes two modeling studies conducted by HydroGeoLogic, Inc. for the US NRC under contract no. NRC-04089-090, entitled, Validation and Testing of the VAM2D Computer Code.'' VAM2D is a two-dimensional, variably saturated flow and transport code, with applications for performance assessment of nuclear waste disposal. The computer code itself is documented in a separate NUREG document (NUREG/CR-5352, 1989). The studies presented in this report involve application of the VAM2D code to two diverse subsurface modeling problems. The first one involves modeling of infiltration and redistribution of water and solutes in an initially dry, heterogeneous field soil. This application involves detailed modeling over a relatively short, 9-month time period. The second problem pertains to the application of VAM2D to the modeling of a waste disposal facility in a fractured clay, over much larger space and time scales and with particular emphasis on the applicability and reliability of using equivalent porous medium approach for simulating flow and transport in fractured geologic media. Reflecting the separate and distinct nature of the two problems studied, this report is organized in two separate parts. 61 refs., 31 figs., 9 tabs.
2-D or not 2-D, that is the question: A Northern California test
Mayeda, K; Malagnini, L; Phillips, W S; Walter, W R; Dreger, D
2005-06-06
Reliable estimates of the seismic source spectrum are necessary for accurate magnitude, yield, and energy estimation. In particular, how seismic radiated energy scales with increasing earthquake size has been the focus of recent debate within the community and has direct implications on earthquake source physics studies as well as hazard mitigation. The 1-D coda methodology of Mayeda et al. has provided the lowest variance estimate of the source spectrum when compared against traditional approaches that use direct S-waves, thus making it ideal for networks that have sparse station distribution. The 1-D coda methodology has been mostly confined to regions of approximately uniform complexity. For larger, more geophysically complicated regions, 2-D path corrections may be required. The complicated tectonics of the northern California region coupled with high quality broadband seismic data provides for an ideal ''apples-to-apples'' test of 1-D and 2-D path assumptions on direct waves and their coda. Using the same station and event distribution, we compared 1-D and 2-D path corrections and observed the following results: (1) 1-D coda results reduced the amplitude variance relative to direct S-waves by roughly a factor of 8 (800%); (2) Applying a 2-D correction to the coda resulted in up to 40% variance reduction from the 1-D coda results; (3) 2-D direct S-wave results, though better than 1-D direct waves, were significantly worse than the 1-D coda. We found that coda-based moment-rate source spectra derived from the 2-D approach were essentially identical to those from the 1-D approach for frequencies less than {approx}0.7-Hz, however for the high frequencies (0.7{le} f {le} 8.0-Hz), the 2-D approach resulted in inter-station scatter that was generally 10-30% smaller. For complex regions where data are plentiful, a 2-D approach can significantly improve upon the simple 1-D assumption. In regions where only 1-D coda correction is available it is still preferable over 2
A linear analytical boundary element method (BEM) for 2D homogeneous potential problems
NASA Astrophysics Data System (ADS)
Friedrich, Jürgen
2002-06-01
The solution of potential problems is not only fundamental for geosciences, but also an essential part of related subjects like electro- and fluid-mechanics. In all fields, solution algorithms are needed that should be as accurate as possible, robust, simple to program, easy to use, fast and small in computer memory. An ideal technique to fulfill these criteria is the boundary element method (BEM) which applies Green's identities to transform volume integrals into boundary integrals. This work describes a linear analytical BEM for 2D homogeneous potential problems that is more robust and precise than numerical methods because it avoids numerical schemes and coordinate transformations. After deriving the solution algorithm, the introduced approach is tested against different benchmarks. Finally, the gained method was incorporated into an existing software program described before in this journal by the same author.
Parallel algorithms for 2-D cylindrical transport equations of Eigenvalue problem
Wei, J.; Yang, S.
2013-07-01
In this paper, aimed at the neutron transport equations of eigenvalue problem under 2-D cylindrical geometry on unstructured grid, the discrete scheme of Sn discrete ordinate and discontinuous finite is built, and the parallel computation for the scheme is realized on MPI systems. Numerical experiments indicate that the designed parallel algorithm can reach perfect speedup, it has good practicality and scalability. (authors)
Use of adaptive walls in 2D tests
NASA Technical Reports Server (NTRS)
Archambaud, J. P.; Chevallier, J. P.
1984-01-01
A new method for computing the wall effects gives precise answers to some questions arising in adaptive wall concept applications: length of adapted regions, fairings with up and downstream regions, residual misadjustments effects, reference conditions. The acceleration of the iterative process convergence and the development of an efficient technology used in CERT T2 wind tunnels give in a single run the required test conditions. Samples taken from CAST 7 tests demonstrate the efficiency of the whole process to obtain significant results with considerations of tridimensional case extension.
Analytical solution of boundary integral equations for 2-D steady linear wave problems
NASA Astrophysics Data System (ADS)
Chuang, J. M.
2005-10-01
Based on the Fourier transform, the analytical solution of boundary integral equations formulated for the complex velocity of a 2-D steady linear surface flow is derived. It has been found that before the radiation condition is imposed, free waves appear both far upstream and downstream. In order to cancel the free waves in far upstream regions, the eigensolution of a specific eigenvalue, which satisfies the homogeneous boundary integral equation, is found and superposed to the analytical solution. An example, a submerged vortex, is used to demonstrate the derived analytical solution. Furthermore, an analytical approach to imposing the radiation condition in the numerical solution of boundary integral equations for 2-D steady linear wave problems is proposed.
Structure-approximating inverse protein folding problem in the 2D HP model.
Gupta, Arvind; Manuch, Ján; Stacho, Ladislav
2005-12-01
The inverse protein folding problem is that of designing an amino acid sequence which has a particular native protein fold. This problem arises in drug design where a particular structure is necessary to ensure proper protein-protein interactions. In this paper, we show that in the 2D HP model of Dill it is possible to solve this problem for a broad class of structures. These structures can be used to closely approximate any given structure. One of the most important properties of a good protein (in drug design) is its stability--the aptitude not to fold simultaneously into other structures. We show that for a number of basic structures, our sequences have a unique fold. PMID:16379538
NASA Astrophysics Data System (ADS)
Tucciarelli, T.
2012-12-01
A new methodology for the solution of irrotational 2D flow problems in domains with strongly unstructured meshes is presented. A fractional time step procedure is applied to the original governing equations, solving consecutively a convective prediction system and a diffusive corrective system. The non linear components of the problem are concentrated in the prediction step, while the correction step leads to the solution of a linear system, of the order of the number of computational cells. A MArching in Space and Time (MAST) approach is applied for the solution of the convective prediction step. The major advantages of the model, as well as its ability to maintain the solution monotonicity even in strongly irregular meshes, are briefly described. The algorithm is applied to the solution of diffusive shallow water equations in a simple domain.
Coupling finite and boundary element methods for 2-D elasticity problems
NASA Technical Reports Server (NTRS)
Krishnamurthy, T.; Raju, I. S.; Sistla, R.
1993-01-01
A finite element-boundary element (FE-BE) coupling method for two-dimensional elasticity problems is developed based on a weighted residual variational method in which a portion of the domain of interest is modeled by FEs and the remainder of the region by BEs. The performance of the FE-BE coupling method is demonstrated via applications to a simple 'patch test' problem and three-crack problems. The method passed the patch tests for various modeling configurations and yielded accurate strain energy release rates for the crack problems studied.
Differential Sensitivity Theory applied to the MESA2D code for multi-material problems
Henninger, R.J.; Maudlin, P.J.; Harstad, E.N.
1996-05-01
The technique called Differential Sensitivity Theory (DST) is extended to the multi-component system of equations solved by the MESA2D hydrocode. DST uses adjoint techniques to determine exact sensitivity derivatives, i.e., if R is a calculation result of interest (response R) and {alpha}{sub i} is a calculation input (parameter {alpha}{sub i}), then {partial_derivative}R/{partial_derivative}{alpha}{sub i} is defined as the sensitivity. The advantage of using DST is that for an n-parameter problem {ital all} n sensitivities can be obtained by integrating the solutions from only {ital two} calculations, a MESA calculation and its corresponding adjoint calculation using an Adjoint Continuum Mechanics (ACM) code. Previous papers have described application of the technique to one-dimensional, single-material problems. This work presents the derivation and solution of the additional adjoint equations for the purpose of computing sensitivities for two-dimensional, multi-component problems. As an example, results for a multi-material flyer plate impact problem featuring an oblique impact are given. {copyright} {ital 1996 American Institute of Physics.}
Differential Sensitivity Theory applied to the MESA2D code for multi-material problems
NASA Astrophysics Data System (ADS)
Henninger, R. J.; Maudlin, P. J.; Harstad, E. N.
1996-05-01
The technique called Differential Sensitivity Theory (DST) is extended to the multi-component system of equations solved by the MESA2D hydrocode. DST uses adjoint techniques to determine exact sensitivity derivatives, i.e., if R is a calculation result of interest (response R) and αi is a calculation input (parameter αi), then ∂R/∂αi is defined as the sensitivity. The advantage of using DST is that for an n-parameter problem all n sensitivities can be obtained by integrating the solutions from only two calculations, a MESA calculation and its corresponding adjoint calculation using an Adjoint Continuum Mechanics (ACM) code. Previous papers have described application of the technique to one-dimensional, single-material problems. This work presents the derivation and solution of the additional adjoint equations for the purpose of computing sensitivities for two-dimensional, multi-component problems. As an example, results for a multi-material flyer plate impact problem featuring an oblique impact are given.
Differential sensitivity theory applied to the MESA2D code for multi-material problems
Henninger, R.J.; Maudlin, P.J.; Harstad, E.N.
1995-09-01
The technique called Differential Sensitivity Theory (DST) is extended to the multi-component system of equations solved by the MESA2D hydrocode. DST uses adjoint techniques to determine exact sensitivity derivatives, i.e., if R is a calculation result of interest (response R) and {alpha}{sub i} is a calculation input (parameter {alpha}{sub i}), then {partial_derivative}R/{partial_derivative}{alpha}{sub i} is defined as the sensitivity. The advantage of using DST is that for an n-parameter problem all n sensitivities can be obtained by integrating the solutions from only two calculations, a MESA calculation and its corresponding adjoint calculation using an Adjoint Continuum Mechanics (ACM) code. Previous papers have described application of the technique to one-dimensional, single-material problems. This work presents the derivation and solution of the additional adjoint equations for the purpose of computing sensitivities for two-dimensional, multi-component problems. As an example, results for a multi-material flyer plate impact problem featuring an oblique impact are given.
Fluctuating Pressure Data from 2-D Nozzle Cold Flow Tests (Dual Bell)
NASA Technical Reports Server (NTRS)
Nesman, Tomas E.
2001-01-01
Rocket engines nozzle performance changes as a vehicle climbs through the atmosphere. An altitude compensating nozzle, ACN, is intended to improve on a fixed geometry bell nozzle that performs at optimum at only one trajectory point. In addition to nozzle performance, nozzle transient loads are an important consideration. Any nozzle experiences large transient toads when shocks pass through the nozzle at start and shutdown. Additional transient toads will occur at transitional flow conditions. The objectives of cold flow nozzle testing at MSFC are CFD benchmark / calibration and Unsteady flow / sideloads. Initial testing performed with 2-D inserts to 14" transonic wind tunnel. Recent review of 2-D data in preparation for nozzle test facility 3-D testing. This presentation shows fluctuating pressure data and some observations from 2-D dual-bell nozzle cold flow tests.
NASA Astrophysics Data System (ADS)
Stone, James M.; Norman, Michael L.
1992-06-01
A detailed description of ZEUS-2D, a numerical code for the simulation of fluid dynamical flows including a self-consistent treatment of the effects of magnetic fields and radiation transfer is presented. Attention is given to the hydrodynamic (HD) algorithms which form the foundation for the more complex MHD and radiation HD algorithms. The effect of self-gravity on the flow dynamics is accounted for by an iterative solution of the sparse-banded matrix resulting from discretizing the Poisson equation in multidimensions. The results of an extensive series of HD test problems are presented. A detailed description of the MHD algorithms in ZEUS-2D is presented. A new method of computing the electromotive force is developed using the method of characteristics (MOC). It is demonstrated through the results of an extensive series of MHD test problems that the resulting hybrid MOC-constrained transport method provides for the accurate evolution of all modes of MHD wave families.
A 2D inverse problem of predicting boiling heat transfer in a long fin
NASA Astrophysics Data System (ADS)
Orzechowski, Tadeusz
2015-12-01
A method for the determination of local values of the heat transfer coefficient on non-isothermal surfaces was analyzed on the example of a long smooth-surfaced fin made of aluminium. On the basis of the experimental data, two cases were taken into consideration: one-dimensional model for Bi < 0.1 and two-dimensional model for thicker elements. In the case when the drop in temperature over the thickness could be omitted, the rejected local values of heat fluxes were calculated from the integral of the equation describing temperature distribution on the fin. The corresponding boiling curve was plotted on the basis of temperature gradient distribution as a function of superheat. For thicker specimens, where Bi > 0.1, the problem was modelled using a 2-D heat conduction equation, for which the boundary conditions were posed on the surface observed with a thermovision camera. The ill-conditioned inverse problem was solved using a method of heat polynomials, which required validation.
Fung, Jimmy; Masser, Thomas; Morgan, Nathaniel R.
2012-06-25
The Sedov test is classically defined as a point blast problem. The Sedov problem has led us to advances in algorithms and in their understanding. Vorticity generation can be physical or numerical. Both play a role in Sedov calculations. The RAGE code (Eulerian) resolves the shock well, but produces vorticity. The source definition matters. For the FLAG code (Lagrange), CCH is superior to SGH by avoiding spurious vorticity generation. FLAG SGH currently has a number of options that improve results over traditional settings. Vorticity production, not shock capture, has driven the Sedov work. We are pursuing treatments with respect to the hydro discretization as well as to artificial viscosity.
Design and true Reynolds number 2-D testing of an advanced technology airfoil
NASA Technical Reports Server (NTRS)
Reaser, J. S.; Hallissy, J. B.; Campbell, R. L.
1983-01-01
A NASA-industry program has been conducted to determine the accuracy of available 2-D airfoil analysis procedures over a wide range of Reynolds numbers. The program also served to develop and demonstrate effective wind tunnel model designs for use in a cryogenic environment. A Lockheed design, CRYO 12X, supercritical, shockfree airfoil was configured using a continuous curvature analytical definition of the ordinates. Test results show a very close ordinate tolerance was necessary to realize the intended pressure distribution. Correlation of test with Korn-Garabedian 2-D analysis pressure data were generally good. GRUMFOIL analysis with a sidewall correction gave a better correlation.
Moran, B
2007-08-08
We present analytic solutions to two test problems that can be used to check the hydrodynamic implementation in computer codes designed to calculate the propagation of shocks in spherically convergent geometry. Our analysis is restricted to fluid materials with constant bulk modulus. In the first problem we present the exact initial acceleration and pressure gradient at the outer surface of a sphere subjected to an exponentially decaying pressure of the form P(t) = P{sub 0}e{sup -at}. We show that finely-zoned hydro-code simulations are in good agreement with our analytic solution. In the second problem we discuss the implosions of incompressible spherical fluid shells and we present the radial pressure profile across the shell thickness. We also discuss a semi-analytic solution to the time-evolution of a nearly spherical shell with arbitrary but small initial 3-dimensional (3-D) perturbations on its inner and outer surfaces.
An analytical approach to estimate the number of small scatterers in 2D inverse scattering problems
NASA Astrophysics Data System (ADS)
Fazli, Roohallah; Nakhkash, Mansor
2012-07-01
This paper presents an analytical method to estimate the location and number of actual small targets in 2D inverse scattering problems. This method is motivated from the exact maximum likelihood estimation of signal parameters in white Gaussian noise for the linear data model. In the first stage, the method uses the MUSIC algorithm to acquire all possible target locations and in the next stage, it employs an analytical formula that works as a spatial filter to determine which target locations are associated to the actual ones. The ability of the method is examined for both the Born and multiple scattering cases and for the cases of well-resolved and non-resolved targets. Many numerical simulations using both the coincident and non-coincident arrays demonstrate that the proposed method can detect the number of actual targets even in the case of very noisy data and when the targets are closely located. Using the experimental microwave data sets, we further show that this method is successful in specifying the number of small inclusions.
2-D Path Corrections for Local and Regional Coda Waves: A Test of Transportability
Mayeda, K M; Malagnini, L; Phillips, W S; Walter, W R; Dreger, D S; Morasca, P
2005-07-13
Reliable estimates of the seismic source spectrum are necessary for accurate magnitude, yield, and energy estimation. In particular, how seismic radiated energy scales with increasing earthquake size has been the focus of recent debate within the community and has direct implications on earthquake source physics studies as well as hazard mitigation. The 1-D coda methodology of Mayeda et al. [2003] has provided the lowest variance estimate of the source spectrum when compared against traditional approaches that use direct S-waves, thus making it ideal for networks that have sparse station distribution. The 1-D coda methodology has been mostly confined to regions of approximately uniform complexity. For larger, more geophysically complicated regions, 2-D path corrections may be required. We will compare performance of 1-D versus 2-D path corrections in a variety of regions. First, the complicated tectonics of the northern California region coupled with high quality broadband seismic data provides for an ideal ''apples-to-apples'' test of 1-D and 2-D path assumptions on direct waves and their coda. Next, we will compare results for the Italian Alps using high frequency data from the University of Genoa. For Northern California, we used the same station and event distribution and compared 1-D and 2-D path corrections and observed the following results: (1) 1-D coda results reduced the amplitude variance relative to direct S-waves by roughly a factor of 8 (800%); (2) Applying a 2-D correction to the coda resulted in up to 40% variance reduction from the 1-D coda results; (3) 2-D direct S-wave results, though better than 1-D direct waves, were significantly worse than the 1-D coda. We found that coda-based moment-rate source spectra derived from the 2-D approach were essentially identical to those from the 1-D approach for frequencies less than {approx}0.7-Hz, however for the high frequencies (0.7 {le} f {le} 8.0-Hz), the 2-D approach resulted in inter-station scatter
An F2D analysis of the Flow Instability Test (FIT) experiment
Suo-Anttila, A.
1993-10-01
The F2D code was used to analyze the Flow-Instability-Test (FIT) experiments at Brookhaven National Laboratories. A one-dimensional analysis of the experiment indicated that at the higher temperature levels the element should be unstable. The experimental data corroborated this theory. The two-dimensional simulation behaved in a manner that was very similar to the experimentally measured behavior. In conclusion, the FIT experimental analysis yields partial code validation of F2D, and it also validates the methodology that is used in analyzing thermal flow stability.
Evaluation of a [13C]-Dextromethorphan Breath Test to Assess CYP2D6 Phenotype
Leeder, J. Steven; Pearce, Robin E.; Gaedigk, Andrea; Modak, Anil; Rosen, David I.
2016-01-01
A [13C]-dextromethorphan ([13C]-DM) breath test was evaluated to assess its feasibility as a rapid, phenotyping assay for CYP2D6 activity. [13C]-DM (0.5 mg/kg) was administered orally with water or potassium bicarbonate-sodium bicarbonate to 30 adult Caucasian volunteers (n = 1 each): CYP2D6 poor metabolizers (2 null alleles; PM-0) and extensive metabolizers with 1 (EM-1) or 2 functional alleles (EM-2). CYP2D6 phenotype was determined by 13CO2 enrichment measured by infrared spectrometry (delta-over-baseline [DOB] value) in expired breath samples collected before and up to 240 minutes after [13C]-DM ingestion and by 4-hour urinary metabolite ratio. The PM-0 group was readily distinguishable from either EM group by both the breath test and urinary metabolite ratio. Using a single point determination of phenotype at 40 minutes and defining PMs as subjects with a DOB ≤ 0.5, the sensitivity of the method was 100%; specificity was 95% with 95% accuracy and resulted in the misclassification of 1 EM-1 individual as a PM. Modification of the initial protocol (timing of potassium bicarbonate-sodium bicarbonate administration relative to dose) yielded comparable results, but there was a tendency toward increased DOB values. Although further development is required, these studies suggest that the [13C]-DM breath test offers promise as a rapid, minimally invasive phenotyping assay for CYP2D6 activity. PMID:18728242
NASA Astrophysics Data System (ADS)
Bouclier, R.; Elguedj, T.; Combescure, A.
2013-11-01
This work deals with the development of 2D solid shell non-uniform rational B-spline elements. We address a static problem, that can be solved with a 2D model, involving a thin slender structure under small perturbations. The plane stress, plane strain and axisymmetric assumption can be made. projection and reduced integration techniques are considered to deal with the locking phenomenon. The use of the approach leads to the implementation of two strategies insensitive to locking: the first strategy is based on a 1D projection of the mean strain across the thickness; the second strategy undertakes to project all the strains onto a suitably chosen 2D space. Conversely, the reduced integration approach based on Gauss points is less expensive, but only alleviates locking and is limited to quadratic approximations. The performance of the various 2D elements developed is assessed through several numerical examples. Simple extensions of these techniques to 3D are finally performed.
Criminality and the 2D:4D ratio: testing the prenatal androgen hypothesis.
Ellis, Lee; Hoskin, Anthony W
2015-03-01
A decade old theory hypothesizes that brain exposure to androgens promotes involvement in criminal behavior. General support for this hypothesis has been provided by studies of postpubertal circulating levels of testosterone, at least among males. However, the theory also predicts that for both genders, prenatal androgens will be positively correlated with persistent offending, an idea for which no evidence currently exists. The present study used an indirect measure of prenatal androgen exposure-the relative length of the second and fourth fingers of the right hand (r2D:4D)-to test the hypothesis that elevated prenatal androgens promote criminal tendencies later in life for males and females. Questionnaires were administered to 2,059 college students in Malaysia and 1,291 college students in the United States. Respondents reported their r2D:4D relative finger lengths along with involvement in 13 categories of delinquent and criminal acts. Statistically significant correlations between the commission of most types of offenses and r2D:4D ratios were found for males and females even after controlling for age. It is concluded that high exposure to androgens during prenatal development contributes to most forms of offending following the onset of puberty. PMID:24013770
Exact ground state for the four-electron problem in a 2D finite honeycomb lattice
NASA Astrophysics Data System (ADS)
Trencsényi, Réka; Glukhov, Konstantin; Gulácsi, Zsolt
2014-07-01
Working in a subspace with dimensionality much smaller than the dimension of the full Hilbert space, we deduce exact four-particle ground states in 2D samples containing hexagonal repeat units and described by Hubbard type of models. The procedure identifies first a small subspace ? in which the ground state ? is placed, than deduces ? by exact diagonalization in ?. The small subspace is obtained by the repeated application of the Hamiltonian ? on a carefully chosen starting wave vector describing the most interacting particle configuration, and the wave vectors resulting from the application of ?, till the obtained system of equations closes in itself. The procedure which can be applied in principle at fixed but arbitrary system size and number of particles is interesting on its own since it provides exact information for the numerical approximation techniques which use a similar strategy, but apply non-complete basis for ?. The diagonalization inside ? provides an incomplete image of the low lying part of the excitation spectrum, but provides the exact ?. Once the exact ground state is obtained, its properties can be easily analysed. The ? is found always as a singlet state whose energy, interestingly, saturates in the ? limit. The unapproximated results show that the emergence probabilities of different particle configurations in the ground state presents 'Zittern' (trembling) characteristics which are absent in 2D square Hubbard systems. Consequently, the manifestation of the local Coulomb repulsion in 2D square and honeycomb types of systems presents differences, which can be a real source in the differences in the many-body behaviour.
NASA Astrophysics Data System (ADS)
Tanaka, Satoyuki; Suzuki, Hirotaka; Sadamoto, Shota; Sannomaru, Shogo; Yu, Tiantang; Bui, Tinh Quoc
2016-08-01
Two-dimensional (2D) in-plane mixed-mode fracture mechanics problems are analyzed employing an efficient meshfree Galerkin method based on stabilized conforming nodal integration (SCNI). In this setting, the reproducing kernel function as meshfree interpolant is taken, while employing the SCNI for numerical integration of stiffness matrix in the Galerkin formulation. The strain components are smoothed and stabilized employing Gauss divergence theorem. The path-independent integral ( J-integral) is solved based on the nodal integration by summing the smoothed physical quantities and the segments of the contour integrals. In addition, mixed-mode stress intensity factors (SIFs) are extracted from the J-integral by decomposing the displacement and stress fields into symmetric and antisymmetric parts. The advantages and features of the present formulation and discretization in evaluation of the J-integral of in-plane 2D fracture problems are demonstrated through several representative numerical examples. The mixed-mode SIFs are evaluated and compared with reference solutions. The obtained results reveal high accuracy and good performance of the proposed meshfree method in the analysis of 2D fracture problems.
An ant colony optimisation algorithm for the 2D and 3D hydrophobic polar protein folding problem
Shmygelska, Alena; Hoos, Holger H
2005-01-01
Background The protein folding problem is a fundamental problems in computational molecular biology and biochemical physics. Various optimisation methods have been applied to formulations of the ab-initio folding problem that are based on reduced models of protein structure, including Monte Carlo methods, Evolutionary Algorithms, Tabu Search and hybrid approaches. In our work, we have introduced an ant colony optimisation (ACO) algorithm to address the non-deterministic polynomial-time hard (NP-hard) combinatorial problem of predicting a protein's conformation from its amino acid sequence under a widely studied, conceptually simple model – the 2-dimensional (2D) and 3-dimensional (3D) hydrophobic-polar (HP) model. Results We present an improvement of our previous ACO algorithm for the 2D HP model and its extension to the 3D HP model. We show that this new algorithm, dubbed ACO-HPPFP-3, performs better than previous state-of-the-art algorithms on sequences whose native conformations do not contain structural nuclei (parts of the native fold that predominantly consist of local interactions) at the ends, but rather in the middle of the sequence, and that it generally finds a more diverse set of native conformations. Conclusions The application of ACO to this bioinformatics problem compares favourably with specialised, state-of-the-art methods for the 2D and 3D HP protein folding problem; our empirical results indicate that our rather simple ACO algorithm scales worse with sequence length but usually finds a more diverse ensemble of native states. Therefore the development of ACO algorithms for more complex and realistic models of protein structure holds significant promise. PMID:15710037
An investigation of DTNS2D for use as an incompressible turbulence modelling test-bed
NASA Technical Reports Server (NTRS)
Steffen, Christopher J., Jr.
1992-01-01
This paper documents an investigation of a two dimensional, incompressible Navier-Stokes solver for use as a test-bed for turbulence modelling. DTNS2D is the code under consideration for use at the Center for Modelling of Turbulence and Transition (CMOTT). This code was created by Gorski at the David Taylor Research Center and incorporates the pseudo compressibility method. Two laminar benchmark flows are used to measure the performance and implementation of the method. The classical solution of the Blasius boundary layer is used for validating the flat plate flow, while experimental data is incorporated in the validation of backward facing step flow. Velocity profiles, convergence histories, and reattachment lengths are used to quantify these calculations. The organization and adaptability of the code are also examined in light of the role as a numerical test-bed.
Dong, Jianping
2014-03-15
The 2D space-fractional Schrödinger equation in the time-independent and time-dependent cases for the scattering problems in the fractional quantum mechanics is studied. We define the Green's functions for the two cases and give the mathematical expression of them in infinite series form and in terms of some special functions. The asymptotic formulas of the Green's functions are also given, and applied to get the approximate wave functions for the fractional quantum scattering problems. These results contain those in the standard (integer) quantum mechanics as special cases, and can be applied to study the complex quantum systems.
OECD/MCCI 2-D Core Concrete Interaction (CCI) tests : final report February 28, 2006.
Farmer, M. T.; Lomperski, S.; Kilsdonk, D. J.; Aeschlimann, R. W.; Basu, S.
2011-05-23
reactor material database for dry cavity conditions is solely one-dimensional. Although the MACE Scoping Test was carried out with a two-dimensional concrete cavity, the interaction was flooded soon after ablation was initiated to investigate debris coolability. Moreover, due to the scoping nature of this test, the apparatus was minimally instrumented and therefore the results are of limited value from the code validation viewpoint. Aside from the MACE program, the COTELS test series also investigated 2-D CCI under flooded cavity conditions. However, the input power density for these tests was quite high relative to the prototypic case. Finally, the BETA test series provided valuable data on 2-D core concrete interaction under dry cavity conditions, but these tests focused on investigating the interaction of the metallic (steel) phase with concrete. Due to these limitations, there is significant uncertainty in the partition of energy dissipated for the ablation of concrete in the lateral and axial directions under dry cavity conditions for the case of a core oxide melt. Accurate knowledge of this 'power split' is important in the evaluation of the consequences of an ex-vessel severe accident; e.g., lateral erosion can undermine containment structures, while axial erosion can penetrate the basemat, leading to ground contamination and/or possible containment bypass. As a result of this uncertainty, there are still substantial differences among computer codes in the prediction of 2-D cavity erosion behavior under both wet and dry cavity conditions. In light of the above issues, the OECD-sponsored Melt Coolability and Concrete Interaction (MCCI) program was initiated at Argonne National Laboratory. The project conducted reactor materials experiments and associated analysis to achieve the following technical objectives: (1) resolve the ex-vessel debris coolability issue through a program that focused on providing both confirmatory evidence and test data for the coolability
A quasi-spectral method for Cauchy problem of 2/D Laplace equation on an annulus
NASA Astrophysics Data System (ADS)
Saito, Katsuyoshi; Nakada, Manabu; Iijima, Kentaro; Onishi, Kazuei
2005-01-01
Real numbers are usually represented in the computer as a finite number of digits hexa-decimal floating point numbers. Accordingly the numerical analysis is often suffered from rounding errors. The rounding errors particularly deteriorate the precision of numerical solution in inverse and ill-posed problems. We attempt to use a multi-precision arithmetic for reducing the rounding error evil. The use of the multi-precision arithmetic system is by the courtesy of Dr Fujiwara of Kyoto University. In this paper we try to show effectiveness of the multi-precision arithmetic by taking two typical examples; the Cauchy problem of the Laplace equation in two dimensions and the shape identification problem by inverse scattering in three dimensions. It is concluded from a few numerical examples that the multi-precision arithmetic works well on the resolution of those numerical solutions, as it is combined with the high order finite difference method for the Cauchy problem and with the eigenfunction expansion method for the inverse scattering problem.
Sweetser, John David
2013-10-01
This report details Sculpt's implementation from a user's perspective. Sculpt is an automatic hexahedral mesh generation tool developed at Sandia National Labs by Steve Owen. 54 predetermined test cases are studied while varying the input parameters (Laplace iterations, optimization iterations, optimization threshold, number of processors) and measuring the quality of the resultant mesh. This information is used to determine the optimal input parameters to use for an unknown input geometry. The overall characteristics are covered in Chapter 1. The speci c details of every case are then given in Appendix A. Finally, example Sculpt inputs are given in B.1 and B.2.
Numerical solution of 2D-vector tomography problem using the method of approximate inverse
NASA Astrophysics Data System (ADS)
Svetov, Ivan; Maltseva, Svetlana; Polyakova, Anna
2016-08-01
We propose a numerical solution of reconstruction problem of a two-dimensional vector field in a unit disk from the known values of the longitudinal and transverse ray transforms. The algorithm is based on the method of approximate inverse. Numerical simulations confirm that the proposed method yields good results of reconstruction of vector fields.
National Prociency Testing Result of CYP2D6*10 Genotyping for Adjuvant Tamoxifen Therapy in China.
Lin, Guigao; Zhang, Kuo; Yi, Lang; Han, Yanxi; Xie, Jiehong; Li, Jinming
2016-01-01
Tamoxifen has been successfully used for treating breast cancer and preventing cancer recurrence. Cytochrome P450 2D6 (CYP2D6) plays a key role in the process of metabolizing tamoxifen to its active moiety, endoxifen. Patients with variants of the CYP2D6 gene may not receive the full benefit of tamoxifen treatment. The CYP2D6*10 variant (the most common variant in Asians) was analyzed to optimize the prescription of tamoxifen in China. To ensure referring clinicians have accurate information for genotype-guided tamoxifen treatment, the Chinese National Center for Clinical Laboratories (NCCL) organized a national proficiency testing (PT) to evaluate the performance of laboratories providing CYP2D6*10 genotyping. Ten genomic DNA samples with CYP2D6 wild-type or CYP2D6*10 variants were validated by PCR-sequencing and sent to 28 participant laboratories. The genotyping results and pharmacogenomic test reports were submitted and evaluated by NCCL experts. Additional information regarding the number of samples tested, the accreditation/certification status, and detecting technology was also requested. Thirty-one data sets were received, with a corresponding analytical sensitivity of 98.2% (548/558 challenges; 95% confidence interval: 96.7-99.1%) and an analytic specificity of 96.5% (675/682; 95% confidence interval: 97.9-99.5%). Overall, 25/28 participants correctly identified CYP2D6*10 status in 10 samples; however, two laboratories made serious genotyping errors. Most of the essential information was included in the 20 submitted CYP2D6*10 test reports. The majority of Chinese laboratories are reliable for detecting the CYP2D6*10 variant; however, several issues revealed in this study underline the importance of PT schemes in continued external assessment and provision of guidelines. PMID:27603206
A multiple-scale Pascal polynomial for 2D Stokes and inverse Cauchy-Stokes problems
NASA Astrophysics Data System (ADS)
Liu, Chein-Shan; Young, D. L.
2016-05-01
The polynomial expansion method is a useful tool for solving both the direct and inverse Stokes problems, which together with the pointwise collocation technique is easy to derive the algebraic equations for satisfying the Stokes differential equations and the specified boundary conditions. In this paper we propose two novel numerical algorithms, based on a third-first order system and a third-third order system, to solve the direct and the inverse Cauchy problems in Stokes flows by developing a multiple-scale Pascal polynomial method, of which the scales are determined a priori by the collocation points. To assess the performance through numerical experiments, we find that the multiple-scale Pascal polynomial expansion method (MSPEM) is accurate and stable against large noise.
A 2D forward and inverse code for streaming potential problems
NASA Astrophysics Data System (ADS)
Soueid Ahmed, A.; Jardani, A.; Revil, A.
2013-12-01
The self-potential method corresponds to the passive measurement of the electrical field in response to the occurrence of natural sources of current in the ground. One of these sources corresponds to the streaming current associated with the flow of the groundwater. We can therefore apply the self- potential method to recover non-intrusively some information regarding the groundwater flow. We first solve the forward problem starting with the solution of the groundwater flow problem, then computing the source current density, and finally solving a Poisson equation for the electrical potential. We use the finite-element method to solve the relevant partial differential equations. In order to reduce the number of (petrophysical) model parameters required to solve the forward problem, we introduced an effective charge density tensor of the pore water, which can be determined directly from the permeability tensor for neutral pore waters. The second aspect of our work concerns the inversion of the self-potential data using Tikhonov regularization with smoothness and weighting depth constraints. This approach accounts for the distribution of the electrical resistivity, which can be independently and approximately determined from electrical resistivity tomography. A numerical code, SP2DINV, has been implemented in Matlab to perform both the forward and inverse modeling. Three synthetic case studies are discussed.
Algebraic rings of integers and some 2D lattice problems in physics
NASA Astrophysics Data System (ADS)
Nanxian, Chen; Zhaodou, Chen; Shaojun, Liu; Yanan, Shen; Xijin, Ge
1996-09-01
This paper develops the Möbius inversion formula for the Gaussian integers and Eisenstein's integers, and gives two applications. The first application is to the two-dimensional arithmetic Fourier transform (AFT), which is suitable for parallel processing. The second application is to two-dimensional inverse lattice problems, and is illustrated with the recovery of interatomic potentials from the cohesive energy for monolayer graphite. The paper demonstrates the potential application in the physical science of integral domains other than the standard integers.
NASA Astrophysics Data System (ADS)
Mo, Yike; Greenhalgh, Stewart A.; Robertsson, Johan O. A.; Karaman, Hakki
2015-05-01
Lateral velocity variations and low velocity near-surface layers can produce strong scattered and guided waves which interfere with reflections and lead to severe imaging problems in seismic exploration. In order to investigate these specific problems by laboratory seismic modelling, a simple 2D ultrasonic model facility has been recently assembled within the Wave Propagation Lab at ETH Zurich. The simulated geological structures are constructed from 2 mm thick metal and plastic sheets, cut and bonded together. The experiments entail the use of a piezoelectric source driven by a pulse amplifier at ultrasonic frequencies to generate Lamb waves in the plate, which are detected by piezoelectric receivers and recorded digitally on a National Instruments recording system, under LabVIEW software control. The 2D models employed were constructed in-house in full recognition of the similitude relations. The first heterogeneous model features a flat uniform low velocity near-surface layer and deeper dipping and flat interfaces separating different materials. The second model is comparable but also incorporates two rectangular shaped inserts, one of low velocity, the other of high velocity. The third model is identical to the second other than it has an irregular low velocity surface layer of variable thickness. Reflection as well as transmission experiments (crosshole & vertical seismic profiling) were performed on each model. The two dominant Lamb waves recorded are the fundamental symmetric mode (non-dispersive) and the fundamental antisymmetric (flexural) dispersive mode, the latter normally being absent when the source transducer is located on a model edge but dominant when it is on the flat planar surface of the plate. Experimental group and phase velocity dispersion curves were determined and plotted for both modes in a uniform aluminium plate. For the reflection seismic data, various processing techniques were applied, as far as pre-stack Kirchhoff migration. The
NASA Astrophysics Data System (ADS)
Stone, James M.; Norman, Michael L.
1992-06-01
In this, the second of a series of three papers, we continue a detailed description of ZEUS-2D, a numerical code for the simulation of fluid dynamical flows in astrophysics including a self-consistent treatment of the effects of magnetic fields and radiation transfer. In this paper, we give a detailed description of the magnetohydrodynamical (MHD) algorithms in ZEUS-2D. The recently developed constrained transport (CT) algorithm is implemented for the numerical evolution of the components of the magnetic field for MHD simulations. This formalism guarantees the numerically evolved field components will satisfy the divergence-free constraint at all times. We find, however, that the method used to compute the electromotive forces must be chosen carefully to propagate accurately all modes of MHD wave families (in particular shear Alfvén waves). A new method of computing the electromotive force is developed using the method of characteristics (MOC). It is demonstrated through the results of an extensive series of MHD test problems that the resulting hybrid MOC-CT method provides for the accurate evolution of all modes of MHD wave families.
Veijola, Timo; Råback, Peter
2007-01-01
We present a straightforward method to solve gas damping problems for perforated structures in two dimensions (2D) utilising a Perforation Profile Reynolds (PPR) solver. The PPR equation is an extended Reynolds equation that includes additional terms modelling the leakage flow through the perforations, and variable diffusivity and compressibility profiles. The solution method consists of two phases: 1) determination of the specific admittance profile and relative diffusivity (and relative compressibility) profiles due to the perforation, and 2) solution of the PPR equation with a FEM solver in 2D. Rarefied gas corrections in the slip-flow region are also included. Analytic profiles for circular and square holes with slip conditions are presented in the paper. To verify the method, square perforated dampers with 16–64 holes were simulated with a three-dimensional (3D) Navier-Stokes solver, a homogenised extended Reynolds solver, and a 2D PPR solver. Cases for both translational (in normal to the surfaces) and torsional motion were simulated. The presented method extends the region of accurate simulation of perforated structures to cases where the homogenisation method is inaccurate and the full 3D Navier-Stokes simulation is too time-consuming.
NASA Astrophysics Data System (ADS)
Moustafa, Salli; Févotte, François; Lathuilière, Bruno; Plagne, Laurent
2014-06-01
The past few years have been marked by a noticeable increase in the interest in 3D whole-core heterogeneous deterministic neutron transport solvers for reference calculations. Due to the extremely large problem sizes tackled by such solvers, they need to use adapted numerical methods and need to be efficiently implemented to take advantage of the full computing power of modern systems. As for numerical methods, one possible approach consists in iterating over resolutions of 2D and 1D MOC problems by taking advantage of prismatic geometries. The MICADO solver, developed at EDF R&D, is a parallel implementation of such a method in distributed and shared memory systems. However it is currently unable to use SIMD vectorization to leverage the full computing power of modern CPUs. In this paper, we describe our first effort to support vectorization in MICADO, typically targeting Intel© SSE CPUs. Both the 2D and 1D algorithms are vectorized, allowing for high expected speedups for the whole spatial solver. We present benchmark computations, which show nearly optimal speedups for our vectorized implementation on the TAKEDA case.
A proposed experimental test to distinguish waves from 2-D turbulence
NASA Technical Reports Server (NTRS)
Dewan, E. M.
1986-01-01
A theory of buoyancy range turbulence that leads to a unique scale, K sub B, that allows one to differentiate between waves and turbulence for the special case of theta = 0 (i.e., horizontally propagating waves) is discussed. The theory does not seem to lead to a practical empirical distinction for the general situation. This is due to the fact that, as theta is increased, one has the ever-increasing presence of BRT for longer wavelengths. The fact that the numerical values of epsilon prime are not yet available compounds the difficulty. In addition, it does not appear possible to encompass true 2-D turbulence in the theory. We are thus driven to a test which circumvents all these difficulties. A proposed test is based on the idea that waves are coherent and propagate, while in turbulence we have the opposite situation. In particular, the test is suggested by the following quotation from MULLER (1984), on the nature of such turbulence: The turbulence in each horizontal plane is independent from the turbulence in the other planes. If this statement were to be taken literally, it would imply that the temporal coherence between horizontal speeds, separated only in altitude, would be zero. Any vertical separation would be forced to take into account the effects of viscosity: that is to say, a specific finite vertical separation would be needed to destroy coherence. In order to estimate this distance, L, one can use L = C(v/S) (1/2) were v is the kinematic viscosity, S is the shear scale, and C is a constant of order unity.
NASA Astrophysics Data System (ADS)
Velioǧlu, Deniz; Cevdet Yalçıner, Ahmet; Zaytsev, Andrey
2016-04-01
Tsunamis are huge waves with long wave periods and wave lengths that can cause great devastation and loss of life when they strike a coast. The interest in experimental and numerical modeling of tsunami propagation and inundation increased considerably after the 2011 Great East Japan earthquake. In this study, two numerical codes, FLOW 3D and NAMI DANCE, that analyze tsunami propagation and inundation patterns are considered. Flow 3D simulates linear and nonlinear propagating surface waves as well as long waves by solving three-dimensional Navier-Stokes (3D-NS) equations. NAMI DANCE uses finite difference computational method to solve 2D depth-averaged linear and nonlinear forms of shallow water equations (NSWE) in long wave problems, specifically tsunamis. In order to validate these two codes and analyze the differences between 3D-NS and 2D depth-averaged NSWE equations, two benchmark problems are applied. One benchmark problem investigates the runup of long waves over a complex 3D beach. The experimental setup is a 1:400 scale model of Monai Valley located on the west coast of Okushiri Island, Japan. Other benchmark problem is discussed in 2015 National Tsunami Hazard Mitigation Program (NTHMP) Annual meeting in Portland, USA. It is a field dataset, recording the Japan 2011 tsunami in Hilo Harbor, Hawaii. The computed water surface elevation and velocity data are compared with the measured data. The comparisons showed that both codes are in fairly good agreement with each other and benchmark data. The differences between 3D-NS and 2D depth-averaged NSWE equations are highlighted. All results are presented with discussions and comparisons. Acknowledgements: Partial support by Japan-Turkey Joint Research Project by JICA on earthquakes and tsunamis in Marmara Region (JICA SATREPS - MarDiM Project), 603839 ASTARTE Project of EU, UDAP-C-12-14 project of AFAD Turkey, 108Y227, 113M556 and 213M534 projects of TUBITAK Turkey, RAPSODI (CONCERT_Dis-021) of CONCERT
Safgren, Stephanie L.; Suman, Vera J.; Kosel, Matthew L.; Gilbert, Judith A; Buhrow, Sarah A.; Black, John L.; Northfelt, Donald W.; Modak, Anil S.; Rosen, David; Ingle, James N.; Ames, Matthew M.; Reid, Joel M.; Goetz, Matthew P.
2015-01-01
Background In tamoxifen-treated patients, breast cancer recurrence differs according to CYP2D6 genotype and endoxifen steady state concentrations (Endx Css). The 13Cdextromethorphan breath test (DM-BT), labeled with 13C at the O-CH3 moiety, measures CYP2D6 enzyme activity. We sought to examine the ability of the DM-BT to identify known CYP2D6 genotypic poor metabolizers and examine the correlation between DMBT and Endx Css. Methods DM-BT and tamoxifen pharmacokinetics were obtained at baseline (b), 3 month (3m) and 6 months (6m) following tamoxifen initiation. Potent CYP2D6 inhibitors were prohibited. The correlation between bDM-BT with CYP2D6 genotype and Endx Css was determined. The association between bDM-BT (where values ≤0.9 is an indicator of poor in vivo CYP2D6 metabolism) and Endx Css (using values ≤ 11.2 known to be associated with poorer recurrence free survival) was explored. Results 91 patients were enrolled and 77 were eligible. CYP2D6 genotype was positively correlated with b, 3m and 6m DMBT (r ranging from 0.457-0. 60 p < 0.001). Both CYP2D6 genotype (r = 0.47; 0.56, p <.0001), and bDM-BT (r=0.60; 0.54; p<.001) were associated with 3m and 6m Endx Css respectively. Seven of 9 patients (78%) with low (≤11.2 nM) 3m Endx Css also had low DM-BT (≤0.9) including 2/2 CYP2D6 PM/PM and 5/5 IM/PM. In contrast, 1 of 48 pts (2%) with a low DM-BT had Endx Css > 11.2 nM. Conclusions In patients not taking potent CYP2D6 inhibitors, DM-BT was associated with CYP2D6 genotype and 3m and 6 m Endx Css but did not provide better discrimination of Endx Css compared to CYP2D6 genotype alone. Further studies are needed to identify additional factors which alter Endx Css. PMID:25714002
NASA Astrophysics Data System (ADS)
Cockmartin, Lesley; Marshall, Nicholas W.; Van Ongeval, Chantal; Aerts, Gwen; Stalmans, Davina; Zanca, Federica; Shaheen, Eman; De Keyzer, Frederik; Dance, David R.; Young, Kenneth C.; Bosmans, Hilde
2015-05-01
This paper introduces a hybrid method for performing detection studies in projection image based modalities, based on image acquisitions of target objects and patients. The method was used to compare 2D mammography and digital breast tomosynthesis (DBT) in terms of the detection performance of spherical densities and microcalcifications. The method starts with the acquisition of spheres of different glandular equivalent densities and microcalcifications of different sizes immersed in a homogeneous breast tissue simulating medium. These target objects are then segmented and the subsequent templates are fused in projection images of patients and processed or reconstructed. This results in hybrid images with true mammographic anatomy and clinically relevant target objects, ready for use in observer studies. The detection study of spherical densities used 108 normal and 178 hybrid 2D and DBT images; 156 normal and 321 hybrid images were used for the microcalcifications. Seven observers scored the presence/absence of the spheres/microcalcifications in a square region via a 5-point confidence rating scale. Detection performance in 2D and DBT was compared via ROC analysis with sub-analyses for the density of the spheres, microcalcification size, breast thickness and z-position. The study was performed on a Siemens Inspiration tomosynthesis system using patient acquisitions with an average age of 58 years and an average breast thickness of 53 mm providing mean glandular doses of 1.06 mGy (2D) and 2.39 mGy (DBT). Study results showed that breast tomosynthesis (AUC = 0.973) outperformed 2D (AUC = 0.831) for the detection of spheres (p < 0.0001) and this applied for all spherical densities and breast thicknesses. By way of contrast, DBT was worse than 2D for microcalcification detection (AUC2D = 0.974, AUCDBT = 0.838, p < 0.0001), with significant differences found for all sizes (150-354 µm), for breast thicknesses above 40 mm and for heights
Simulation and Analysis of Converging Shock Wave Test Problems
Ramsey, Scott D.; Shashkov, Mikhail J.
2012-06-21
Results and analysis pertaining to the simulation of the Guderley converging shock wave test problem (and associated code verification hydrodynamics test problems involving converging shock waves) in the LANL ASC radiation-hydrodynamics code xRAGE are presented. One-dimensional (1D) spherical and two-dimensional (2D) axi-symmetric geometric setups are utilized and evaluated in this study, as is an instantiation of the xRAGE adaptive mesh refinement capability. For the 2D simulations, a 'Surrogate Guderley' test problem is developed and used to obviate subtleties inherent to the true Guderley solution's initialization on a square grid, while still maintaining a high degree of fidelity to the original problem, and minimally straining the general credibility of associated analysis and conclusions.
2D-Raman-THz spectroscopy: A sensitive test of polarizable water models
NASA Astrophysics Data System (ADS)
Hamm, Peter
2014-11-01
In a recent paper, the experimental 2D-Raman-THz response of liquid water at ambient conditions has been presented [J. Savolainen, S. Ahmed, and P. Hamm, Proc. Natl. Acad. Sci. U. S. A. 110, 20402 (2013)]. Here, all-atom molecular dynamics simulations are performed with the goal to reproduce the experimental results. To that end, the molecular response functions are calculated in a first step, and are then convoluted with the laser pulses in order to enable a direct comparison with the experimental results. The molecular dynamics simulation are performed with several different water models: TIP4P/2005, SWM4-NDP, and TL4P. As polarizability is essential to describe the 2D-Raman-THz response, the TIP4P/2005 water molecules are amended with either an isotropic or a anisotropic polarizability a posteriori after the molecular dynamics simulation. In contrast, SWM4-NDP and TL4P are intrinsically polarizable, and hence the 2D-Raman-THz response can be calculated in a self-consistent way, using the same force field as during the molecular dynamics simulation. It is found that the 2D-Raman-THz response depends extremely sensitively on details of the water model, and in particular on details of the description of polarizability. Despite the limited time resolution of the experiment, it could easily distinguish between various water models. Albeit not perfect, the overall best agreement with the experimental data is obtained for the TL4P water model.
2D-Raman-THz spectroscopy: A sensitive test of polarizable water models
Hamm, Peter
2014-11-14
In a recent paper, the experimental 2D-Raman-THz response of liquid water at ambient conditions has been presented [J. Savolainen, S. Ahmed, and P. Hamm, Proc. Natl. Acad. Sci. U. S. A. 110, 20402 (2013)]. Here, all-atom molecular dynamics simulations are performed with the goal to reproduce the experimental results. To that end, the molecular response functions are calculated in a first step, and are then convoluted with the laser pulses in order to enable a direct comparison with the experimental results. The molecular dynamics simulation are performed with several different water models: TIP4P/2005, SWM4-NDP, and TL4P. As polarizability is essential to describe the 2D-Raman-THz response, the TIP4P/2005 water molecules are amended with either an isotropic or a anisotropic polarizability a posteriori after the molecular dynamics simulation. In contrast, SWM4-NDP and TL4P are intrinsically polarizable, and hence the 2D-Raman-THz response can be calculated in a self-consistent way, using the same force field as during the molecular dynamics simulation. It is found that the 2D-Raman-THz response depends extremely sensitively on details of the water model, and in particular on details of the description of polarizability. Despite the limited time resolution of the experiment, it could easily distinguish between various water models. Albeit not perfect, the overall best agreement with the experimental data is obtained for the TL4P water model.
OECD 2-D Core Concrete Interaction (CCI) tests : CCI-2 test plan, Rev. 0 January 31, 2004.
Farmer, M. T.; Kilsdonk, D. J.; Lomperski, S.; Aeschlimann, R. W.; Basu, S.
2011-05-23
The Melt Attack and Coolability Experiments (MACE) program addressed the issue of the ability of water to cool and thermally stabilize a molten core-concrete interaction when the reactants are flooded from above. These tests provided data regarding the nature of corium interactions with concrete, the heat transfer rates from the melt to the overlying water pool, and the role of noncondensable gases in the mixing processes that contribute to melt quenching. As a follow-on program to MACE, The Melt Coolability and Concrete Interaction Experiments (MCCI) project is conducting reactor material experiments and associated analysis to achieve the following objectives: (1) resolve the ex-vessel debris coolability issue through a program that focuses on providing both confirmatory evidence and test data for the coolability mechanisms identified in MACE integral effects tests, and (2) address remaining uncertainties related to long-term two-dimensional molten core-concrete interactions under both wet and dry cavity conditions. Achievement of these two program objectives will demonstrate the efficacy of severe accident management guidelines for existing plants, and provide the technical basis for better containment designs for future plants. In terms of satisfying these objectives, the Management Board (MB) approved the conduct of two long-term 2-D Core-Concrete Interaction (CCI) experiments designed to provide information in several areas, including: (i) lateral vs. axial power split during dry core-concrete interaction, (ii) integral debris coolability data following late phase flooding, and (iii) data regarding the nature and extent of the cooling transient following breach of the crust formed at the melt-water interface. The first of these two tests, CCI-1, was conducted on December 19, 2003. This test investigated the interaction of a fully oxidized 400 kg PWR core melt, initially containing 8 wt % calcined siliceous concrete, with a specially designed two
HT2DINV: A 2D forward and inverse code for steady-state and transient hydraulic tomography problems
NASA Astrophysics Data System (ADS)
Soueid Ahmed, A.; Jardani, A.; Revil, A.; Dupont, J. P.
2015-12-01
Hydraulic tomography is a technique used to characterize the spatial heterogeneities of storativity and transmissivity fields. The responses of an aquifer to a source of hydraulic stimulations are used to recover the features of the estimated fields using inverse techniques. We developed a 2D free source Matlab package for performing hydraulic tomography analysis in steady state and transient regimes. The package uses the finite elements method to solve the ground water flow equation for simple or complex geometries accounting for the anisotropy of the material properties. The inverse problem is based on implementing the geostatistical quasi-linear approach of Kitanidis combined with the adjoint-state method to compute the required sensitivity matrices. For undetermined inverse problems, the adjoint-state method provides a faster and more accurate approach for the evaluation of sensitivity matrices compared with the finite differences method. Our methodology is organized in a way that permits the end-user to activate parallel computing in order to reduce the computational burden. Three case studies are investigated demonstrating the robustness and efficiency of our approach for inverting hydraulic parameters.
Comparison between 2D and 3D Numerical Modelling of a hot forging simulative test
Croin, M.; Ghiotti, A.; Bruschi, S.
2007-04-07
The paper presents the comparative analysis between 2D and 3D modelling of a simulative experiment, performed in laboratory environment, in which operating conditions approximate hot forging of a turbine aerofoil section. The plane strain deformation was chosen as an ideal case to analyze the process because of the thickness variations in the final section and the consequent distributions of contact pressure and sliding velocity at the interface that are closed to the conditions of the real industrial process. In order to compare the performances of 2D and 3D approaches, two different analyses were performed and compared with the experiments in terms of loads and temperatures peaks at the interface between the dies and the workpiece.
Fabrication and Testing of Low Cost 2D Carbon-Carbon Nozzle Extensions at NASA/MSFC
NASA Technical Reports Server (NTRS)
Greene, Sandra Elam; Shigley, John K.; George, Russ; Roberts, Robert
2015-01-01
Subscale liquid engine tests were conducted at NASA/MSFC using a 1.2 Klbf engine with liquid oxygen (LOX) and gaseous hydrogen. Testing was performed for main-stage durations ranging from 10 to 160 seconds at a chamber pressure of 550 psia and a mixture ratio of 5.7. Operating the engine in this manner demonstrated a new and affordable test capability for evaluating subscale nozzles by exposing them to long duration tests. A series of 2D C-C nozzle extensions were manufactured, oxidation protection applied and then tested on a liquid engine test facility at NASA/MSFC. The C-C nozzle extensions had oxidation protection applied using three very distinct methods with a wide range of costs and process times: SiC via Polymer Impregnation & Pyrolysis (PIP), Air Plasma Spray (APS) and Melt Infiltration. The tested extensions were about 6" long with an exit plane ID of about 6.6". The test results, material properties and performance of the 2D C-C extensions and attachment features will be discussed.
Liu, T.; Deptuch, G.; Hoff, J.; Jindariani, S.; Joshi, S.; Olsen, J.; Tran, N.; Trimpl, M.
2015-02-01
An associative memory-based track finding approach has been proposed for a Level 1 tracking trigger to cope with increasing luminosities at the LHC. The associative memory uses a massively parallel architecture to tackle the intrinsically complex combinatorics of track finding algorithms, thus avoiding the typical power law dependence of execution time on occupancy and solving the pattern recognition in times roughly proportional to the number of hits. This is of crucial importance given the large occupancies typical of hadronic collisions. The design of an associative memory system capable of dealing with the complexity of HL-LHC collisions and with the short latency required by Level 1 triggering poses significant, as yet unsolved, technical challenges. For this reason, an aggressive R&D program has been launched at Fermilab to advance state of-the-art associative memory technology, the so called VIPRAM (Vertically Integrated Pattern Recognition Associative Memory) project. The VIPRAM leverages emerging 3D vertical integration technology to build faster and denser Associative Memory devices. The first step is to implement in conventional VLSI the associative memory building blocks that can be used in 3D stacking, in other words, the building blocks are laid out as if it is a 3D design. In this paper, we report on the first successful implementation of a 2D VIPRAM demonstrator chip (protoVIPRAM00). The results show that these building blocks are ready for 3D stacking.
NASA Astrophysics Data System (ADS)
Liu, T.; Deptuch, G.; Hoff, J.; Jindariani, S.; Joshi, S.; Olsen, J.; Tran, N.; Trimpl, M.
2015-02-01
An associative memory-based track finding approach has been proposed for a Level 1 tracking trigger to cope with increasing luminosities at the LHC. The associative memory uses a massively parallel architecture to tackle the intrinsically complex combinatorics of track finding algorithms, thus avoiding the typical power law dependence of execution time on occupancy and solving the pattern recognition in times roughly proportional to the number of hits. This is of crucial importance given the large occupancies typical of hadronic collisions. The design of an associative memory system capable of dealing with the complexity of HL-LHC collisions and with the short latency required by Level 1 triggering poses significant, as yet unsolved, technical challenges. For this reason, an aggressive R&D program has been launched at Fermilab to advance state of-the-art associative memory technology, the so called VIPRAM (Vertically Integrated Pattern Recognition Associative Memory) project. The VIPRAM leverages emerging 3D vertical integration technology to build faster and denser Associative Memory devices. The first step is to implement in conventional VLSI the associative memory building blocks that can be used in 3D stacking; in other words, the building blocks are laid out as if it is a 3D design. In this paper, we report on the first successful implementation of a 2D VIPRAM demonstrator chip (protoVIPRAM00). The results show that these building blocks are ready for 3D stacking.
Ignition problems in scramjet testing
Mitani, Tohru
1995-05-01
Ignition of H{sub 2} in heated air containing H{sub 2}O, radicals, and dust was investigated for scramjet testing. Using a reduced kinetic model for H{sub 2}{minus}O{sub 2} systems, the effects of H{sub 2}O and radicals in nozzles are discussed in relation to engine testing with vitiation heaters. Analysis using linearized rate-equations suggested that the addition of O atoms was 1.5 times more effective than the addition of H atoms for ignition. This result can be applied to the problem of premature ignition caused by residual radicals and to plasma-jet igniters. Thermal and chemical effects of dust, inevitable in storage air heaters, were studied next. The effects of heat capacity and size of dust were expressed in terms of an exponential integral function. It was found that the radical termination on the surface of dust produces an effect equivalent to heat loss. The inhibition of ignition by dust may result, if the mass fraction of dust becomes 10{sup {minus}3}.
Implementation of a system to life test 2-D laser arrays
NASA Astrophysics Data System (ADS)
Faltus, Thomas H.; Bicket, Daniel J.
1992-02-01
Multi-emitter laser devices, stacked to form 2-dimensional arrays, have been shown to effectively pump Nd:YAG slabs in solid state laser systems. Using these arrays as substitutes for flashlamps provides the potential for increased reliability of laser systems. However, to quantify this reliability improvement, laser arrays must be life tested. To ensure that the life test data accurately describes the array lifetimes, the life test system must possess the following characteristics: adequate control of operating stresses, to ensure that the test results apply to true use-conditions; continuous monitoring and recording of array health, to capture unpredictable variations in array performance; in-situ parameter measurement, to measure array performance without inducing handling damage; and extensive safety interlocks, to protect personnel from laser hazards. This paper describes an array life test system possessing these characteristics. It describes the system hardware, operating and test software, and the methodology behind the system's use. We demonstrate the system's performance by life testing 2-dimensional laser arrays having previously documented front facet anomalies. Disadvantages as well as advantages of design decisions are discussed.
Altitude testing of the 2D V/STOL ADEN demonstrator on an F404 engine
NASA Technical Reports Server (NTRS)
Blozy, J. T.
1985-01-01
The Augmented Deflector Exhaust Nozzle (ADEN) exhaust system was tested in the PSL-3 altitude chamber at the NASA Lewis Research Center in order to evaluate aerodynamic performance, cooling-system effectiveness, and mechanical operation at flight-type conditions. The ADEN, a flight-weight, two-dimensional, thrust-vectoring nozzle, was successfully tested on the F404 engine using a remote engine control system for automatic or manual setting of the throat-area control and available fan air for the nozzle internal cooling system. Throughout the test, the ADEN performed with no adverse effects on the engine or augmentor operation.
Altitude testing of a flight weight, self-cooled, 2D thrust vectoring exhaust nozzle
NASA Technical Reports Server (NTRS)
Wooten, W. H.; Blozy, J. T.; Speir, D. W.; Lottig, R. A.
1984-01-01
The Augmented Deflector Exhaust Nozzle (ADEN) was tested in PSL-3 at NASA-Lewis Research Center using an F404 engine. The ADEN is a flight weight Single Expansion Ramp Nozzle with thrust vectoring, an internal cooling system utilizing the available engine fan flow, and a variable area throat controlled by the engine control system. Test conditions included dry and max A/B operation at nozzle pressure ratios from 2.0 to 15.0. High nozzle pressure loading was simulated to verify structural integrity at near maximum design pressure. Nozzle settings covered the full range in throat area and + or - 15 deg deflection angle. Test results demonstrated expected aerodynamic performance, cooling system effectiveness, control system stability, and mechanical integrity.
Critical Heat Flux Experiments on the Reactor Vessel Wall Using 2-D Slice Test Section
Jeong, Yong Hoon; Chang, Soon Heung; Baek, Won-Pil
2005-11-15
The critical heat flux (CHF) on the reactor vessel outer wall was measured using the two-dimensional slice test section. The radius and the channel area of the test section were 2.5 m and 10 cm x 15 cm, respectively. The flow channel area and the heater width were smaller than those of the ULPU experiments, but the radius was greater than that of the ULPU. The CHF data under the inlet subcooling of 2 to 25 deg. C and the mass flux 0 to 300 kg/m{sup 2}.s had been acquired. The measured CHF value was generally slightly lower than that of the ULPU. The difference possibly comes from the difference of the test section material and the thickness. However, the general trend of CHF according to the mass flux was similar with that of the ULPU. The experimental CHF data were compared with the predicted values by SULTAN correlation. The SULTAN correlation predicted well this study's data only for the mass flux higher than 200 kg/m{sup 2}.s, and for the exit quality lower than 0.05. The local condition-based correlation was developed, and it showed good prediction capability for broad quality (-0.01 to 0.5) and mass flux (<300 kg/m{sup 2}.s) conditions with a root-mean-square error of 2.4%. There were increases in the CHF with trisodium phosphate-added water.
Surrogate Guderley Test Problem Definition
Ramsey, Scott D.; Shashkov, Mikhail J.
2012-07-06
The surrogate Guderley problem (SGP) is a 'spherical shock tube' (or 'spherical driven implosion') designed to ease the notoriously subtle initialization of the true Guderley problem, while still maintaining a high degree of fidelity. In this problem (similar to the Guderley problem), an infinitely strong shock wave forms and converges in one-dimensional (1D) cylindrical or spherical symmetry through a polytropic gas with arbitrary adiabatic index {gamma}, uniform density {rho}{sub 0}, zero velocity, and negligible pre-shock pressure and specific internal energy (SIE). This shock proceeds to focus on the point or axis of symmetry at r = 0 (resulting in ostensibly infinite pressure, velocity, etc.) and reflect back out into the incoming perturbed gas.
Medical Tests for Prostate Problems
... to be related to urine blockage, the health care provider may recommend tests that measure bladder pressure and urine flow rate. ... pain, chills, or fever—should call their health care provider immediately. [ Top ] How soon will prostate test results be available? Results for simple medical tests ...
Medical Tests for Prostate Problems
... appears to be related to urine blockage, the health care provider may recommend tests that measure bladder pressure and urine flow rate. ... including pain, chills, or fever—should call their health care provider ... soon will prostate test results be available? Results for simple medical tests ...
Najjar, F M; Solberg, J; White, D
2008-04-17
A verification test suite has been assessed with primary focus on low reynolds number flow of liquid metals. This is representative of the interface between the armature and rail in gun applications. The computational multiphysics framework, ALE3D, is used. The main objective of the current study is to provide guidance and gain confidence in the results obtained with ALE3D. A verification test suite based on 2-D cases is proposed and includes the lid-driven cavity and the Couette flow are investigated. The hydro and thermal fields are assumed to be steady and laminar in nature. Results are compared with analytical solutions and previously published data. Mesh resolution studies are performed along with various models for the equation of state.
NASA Astrophysics Data System (ADS)
Pérez-Corona, M.; García, J. A.; Taller, G.; Polgár, D.; Bustos, E.; Plank, Z.
2016-02-01
The purpose of geophysical electrical surveys is to determine the subsurface resistivity distribution by making measurements on the ground surface. From these measurements, the true resistivity of the subsurface can be estimated. The ground resistivity is related to various geological parameters, such as the mineral and fluid content, porosity and degree of water saturation in the rock. Electrical resistivity surveys have been used for many decades in hydrogeological, mining and geotechnical investigations. More recently, they have been used for environmental surveys. To obtain a more accurate subsurface model than is possible with a simple 1-D model, a more complex model must be used. In a 2-D model, the resistivity values are allowed to vary in one horizontal direction (usually referred to as the x direction) but are assumed to be constant in the other horizontal (the y) direction. A more realistic model would be a fully 3-D model where the resistivity values are allowed to change in all three directions. In this research, a simulation of the cone penetration test and 2D imaging resistivity are used as tools to simulate the distribution of hydrocarbons in soil.
Problem-Solving Test: Pyrosequencing
ERIC Educational Resources Information Center
Szeberenyi, Jozsef
2013-01-01
Terms to be familiar with before you start to solve the test: Maxam-Gilbert sequencing, Sanger sequencing, gel electrophoresis, DNA synthesis reaction, polymerase chain reaction, template, primer, DNA polymerase, deoxyribonucleoside triphosphates, orthophosphate, pyrophosphate, nucleoside monophosphates, luminescence, acid anhydride bond,…
Inverse Problem in Nondestructive Testing Using Arrayed Eddy Current Sensors
Zaoui, Abdelhalim; Menana, Hocine; Feliachi, Mouloud; Berthiau, Gérard
2010-01-01
A fast crack profile reconstitution model in nondestructive testing is developed using an arrayed eddy current sensor. The inverse problem is based on an iterative solving of the direct problem using genetic algorithms. In the direct problem, assuming a current excitation, the incident field produced by all the coils of the arrayed sensor is obtained by the translation and superposition of the 2D axisymmetric finite element results obtained for one coil; the impedance variation of each coil, due to the crack, is obtained by the reciprocity principle involving the dyadic Green’s function. For the inverse problem, the surface of the crack is subdivided into rectangular cells, and the objective function is expressed only in terms of the depth of each cell. The evaluation of the dyadic Green’s function matrix is made independently of the iterative procedure, making the inversion very fast. PMID:22163680
Inverse problem in nondestructive testing using arrayed eddy current sensors.
Zaoui, Abdelhalim; Menana, Hocine; Feliachi, Mouloud; Berthiau, Gérard
2010-01-01
A fast crack profile reconstitution model in nondestructive testing is developed using an arrayed eddy current sensor. The inverse problem is based on an iterative solving of the direct problem using genetic algorithms. In the direct problem, assuming a current excitation, the incident field produced by all the coils of the arrayed sensor is obtained by the translation and superposition of the 2D axisymmetric finite element results obtained for one coil; the impedance variation of each coil, due to the crack, is obtained by the reciprocity principle involving the dyadic Green's function. For the inverse problem, the surface of the crack is subdivided into rectangular cells, and the objective function is expressed only in terms of the depth of each cell. The evaluation of the dyadic Green's function matrix is made independently of the iterative procedure, making the inversion very fast. PMID:22163680
Photoluminescence and the gallium problem for highest-mobility GaAs/AlGaAs-based 2d electron gases
NASA Astrophysics Data System (ADS)
Schläpfer, F.; Dietsche, W.; Reichl, C.; Faelt, S.; Wegscheider, W.
2016-05-01
The quest for extremely high mobilities of 2d electron gases in MBE-grown heterostructures is hampered by the available purity of the starting materials, particularly of the gallium. Here we compare the role of different Ga lots having nominally the highest possible quality on the mobility and the photoluminescence (PL) of modulation doped single interface structures and find significant differences. A weak exciton PL reveals that the purity of the Ga is insufficient. No high mobility can be reached with such a lot with a reasonable effort. On the other hand, a strong exciton PL indicates a high initial Ga purity, allowing to reach mobilities of 15 million (single interface) or 28 million cm2/V s (doped quantum wells) in our MBE systems. We discuss possible origins of the inconsistent Ga quality. Furthermore, we compare samples grown in different MBE systems over a period of several years and find that mobility and PL are correlated if similar structures and growth procedures are used.
NASA Astrophysics Data System (ADS)
Kh., Lotfy
2012-06-01
In the present paper, we introduce the coupled theory (CD), Lord-Schulman (LS) theory, and Green-Lindsay (GL) theory to study the influences of a magnetic field and rotation on a two-dimensional problem of fibre-reinforced thermoelasticity. The material is a homogeneous isotropic elastic half-space. The method applied here is to use normal mode analysis to solve a thermal shock problem. Some particular cases are also discussed in the context of the problem. Deformation of a body depends on the nature of the force applied as well as the type of boundary conditions. Numerical results for the temperature, displacement, and thermal stress components are given and illustrated graphically in the absence and the presence of the magnetic field and rotation.
ERIC Educational Resources Information Center
Leighty, Katherine A.; Menzel, Charles R.; Fragaszy, Dorothy M.
2008-01-01
Object recognition research is typically conducted using 2D stimuli in lieu of 3D objects. This study investigated the amount and complexity of knowledge gained from 2D stimuli in adult chimpanzees ("Pan troglodytes") and young children (aged 3 and 4 years) using a titrated series of cross-dimensional search tasks. Results indicate that 3-year-old…
Fevotte, F.; Lathuiliere, B.
2013-07-01
The large increase in computing power over the past few years now makes it possible to consider developing 3D full-core heterogeneous deterministic neutron transport solvers for reference calculations. Among all approaches presented in the literature, the method first introduced in [1] seems very promising. It consists in iterating over resolutions of 2D and ID MOC problems by taking advantage of prismatic geometries without introducing approximations of a low order operator such as diffusion. However, before developing a solver with all industrial options at EDF, several points needed to be clarified. In this work, we first prove the convergence of this iterative process, under some assumptions. We then present our high-performance, parallel implementation of this algorithm in the MICADO solver. Benchmarking the solver against the Takeda case shows that the 2D-1D coupling algorithm does not seem to affect the spatial convergence order of the MOC solver. As for performance issues, our study shows that even though the data distribution is suited to the 2D solver part, the efficiency of the ID part is sufficient to ensure a good parallel efficiency of the global algorithm. After this study, the main remaining difficulty implementation-wise is about the memory requirement of a vector used for initialization. An efficient acceleration operator will also need to be developed. (authors)
Farmer, M. T.; Kilsdonk, D. J.; Lomperski, S.; Aeschliman, R. W.; Basu, S.
2011-05-23
experiments to address remaining uncertainties related to long-term two-dimensional molten core-concrete interaction. In particular, for both wet and dry cavity conditions, there is uncertainty insofar as evaluating the lateral vs. axial power split during a core-concrete interaction due to a lack of experiment data. As a result, there are differences in the 2-D cavity erosion predicted by codes such as MELCOR, WECHSL, and COSACO. The first step towards generating this data is to produce a test plan for review by the Project Review Group (PRG). The purpose of this document is to provide this plan.
NASA Astrophysics Data System (ADS)
Leblond, Jean-Baptiste; Frelat, Joël
2014-03-01
It is experimentally well-known that a crack loaded in mode I+III propagates through formation of discrete fracture facets inclined at a certain tilt angle on the original crack plane, depending on the ratio of the mode III to mode I initial stress intensity factors. Pollard et al. (1982) have proposed to calculate this angle by considering the tractions on all possible future infinitesimal facets and assuming shear tractions to be zero on that which will actually develop. In this paper we consider the opposite case of well-developed facets; the stress field near the lateral fronts of such facets becomes independent of the initial crack and essentially 2D in a plane perpendicular to the main direction of crack propagation. To determine this stress field, we solve the model 2D problem of an infinite plate containing an infinite periodic array of cracks inclined at some angle on a straight line, and loaded through uniform stresses at infinity. This is done first analytically, for small values of this angle, by combining Muskhelishvili's (1953) formalism and a first-order perturbation procedure. The formulae found for the 2D stress intensity factors are then extended in an approximate way to larger angles by using another reference solution, and finally assessed through comparison with some finite element results. To finally illustrate the possible future application of these formulae to the prediction of the stationary tilt angle, we introduce the tentative assumption that the 2D mode II stress intensity factor is zero on the lateral fronts of the facets. An approximate formula providing the tilt angle as a function of the ratio of the mode III to mode I stress intensity factors of the initial crack is deduced from there. This formula, which slightly depends on the type of loading imposed, predicts somewhat smaller angles than that of Pollard et al. (1982).
NASA Technical Reports Server (NTRS)
Costiner, Sorin; Taasan, Shlomo
1994-01-01
This paper presents multigrid (MG) techniques for nonlinear eigenvalue problems (EP) and emphasizes an MG algorithm for a nonlinear Schrodinger EP. The algorithm overcomes the mentioned difficulties combining the following techniques: an MG projection coupled with backrotations for separation of solutions and treatment of difficulties related to clusters of close and equal eigenvalues; MG subspace continuation techniques for treatment of the nonlinearity; an MG simultaneous treatment of the eigenvectors at the same time with the nonlinearity and with the global constraints. The simultaneous MG techniques reduce the large number of self consistent iterations to only a few or one MG simultaneous iteration and keep the solutions in a right neighborhood where the algorithm converges fast.
Techniques utilized in the simulated altitude testing of a 2D-CD vectoring and reversing nozzle
NASA Technical Reports Server (NTRS)
Block, H. Bruce; Bryant, Lively; Dicus, John H.; Moore, Allan S.; Burns, Maureen E.; Solomon, Robert F.; Sheer, Irving
1988-01-01
Simulated altitude testing of a two-dimensional, convergent-divergent, thrust vectoring and reversing exhaust nozzle was accomplished. An important objective of this test was to develop test hardware and techniques to properly operate a vectoring and reversing nozzle within the confines of an altitude test facility. This report presents detailed information on the major test support systems utilized, the operational performance of the systems and the problems encountered, and test equipment improvements recommended for future tests. The most challenging support systems included the multi-axis thrust measurement system, vectored and reverse exhaust gas collection systems, and infrared temperature measurement systems used to evaluate and monitor the nozzle. The feasibility of testing a vectoring and reversing nozzle of this type in an altitude chamber was successfully demonstrated. Supporting systems performed as required. During reverser operation, engine exhaust gases were successfully captured and turned downstream. However, a small amount of exhaust gas spilled out the collector ducts' inlet openings when the reverser was opened more than 60 percent. The spillage did not affect engine or nozzle performance. The three infrared systems which viewed the nozzle through the exhaust collection system worked remarkably well considering the harsh environment.
Leighty, Katherine A; Menzel, Charles R; Fragaszy, Dorothy M
2008-09-01
Object recognition research is typically conducted using 2D stimuli in lieu of 3D objects. This study investigated the amount and complexity of knowledge gained from 2D stimuli in adult chimpanzees (Pan troglodytes) and young children (aged 3 and 4 years) using a titrated series of cross-dimensional search tasks. Results indicate that 3-year-old children utilize a response rule guided by local features to solve cross-dimensional tasks. Four-year-old toddlers and adult chimpanzees use information about object form and compositional structure from a 2D image to guide their search in three dimensions. Findings have specific implications to research conducted in object recognition/perception and broad relevance to all areas of research and daily living that incorporate 2D displays. PMID:18801134
The MINPACK-2 test problem collection
Averick, B.M.; Carter, R.G.; Xue, Guo-Liang; More, J.J.
1992-06-01
Optimization software has often been developed without any specific application in mind. This generic approach has worked well in many cases, but as we seek the solution of larger and more complex optimization problems on high-performance computers, the development of optimization software should take into account specific optimization problems that arise in a wide range of applications. This observation was the motivation for the development of the MINPACK-2 test problem collection. Each of the problems in this collection comes from a real application and is representative of other commonly encountered problems. There are problems from such diverse fields as fluid dynamics, medicine, elasticity, combustion, molecular conformation, nondestructive testing, chemical kinetics, lubrication, and superconductivity.
NASA Astrophysics Data System (ADS)
Lotfy, Kh.; Othman, Mohamed I. A.
2014-01-01
In the present paper, the coupled theory, Lord-Şhulman theory, and Green-Lindsay theory are introduced to study the influence of a magnetic field on the 2-D problem of a fiber-reinforced thermoelastic. These theories are also applied to study the influence of reinforcement on the total deformation of an infinite space weakened by a finite linear opening Mode-I crack. The material is homogeneous and an isotropic elastic half-space. The crack is subjected to a prescribed temperature and stress distribution. Normal mode analysis is used to solve the problem of a Mode-I crack. Numerical results for the temperature, the displacement, and thermal stress components are given and illustrated graphically in the absence and the presence of the magnetic field. A comparison between the three theories is also made for different depths.
Rua, Francesco; Sadeghi, Sheila J; Castrignanò, Silvia; Valetti, Francesca; Gilardi, Gianfranco
2015-10-01
This work reports for the first time the direct electron transfer of the Canis familiaris cytochrome P450 2D15 on glassy carbon electrodes to provide an analytical tool as an alternative to P450 animal testing in the drug discovery process. Cytochrome P450 2D15, that corresponds to the human homologue P450 2D6, was recombinantly expressed in Escherichia coli and entrapped on glassy carbon electrodes (GC) either with the cationic polymer polydiallyldimethylammonium chloride (PDDA) or in the presence of gold nanoparticles (AuNPs). Reversible electrochemical signals of P450 2D15 were observed with calculated midpoint potentials (E1/2) of −191 ± 5 and −233 ± 4 mV vs. Ag/AgCl for GC/PDDA/2D15 and GC/AuNPs/2D15, respectively. These experiments were then followed by the electro-catalytic activity of the immobilized enzyme in the presence of metoprolol. The latter drug is a beta-blocker used for the treatment of hypertension and is a specific marker of the human P450 2D6 activity. Electrocatalysis data showed that only in the presence of AuNps the expected α-hydroxy-metoprolol product was present as shown by HPLC. The successful immobilization of the electroactive C. familiaris cytochrome P450 2D15 on electrode surfaces addresses the ever increasing demand of developing alternative in vitromethods for amore detailed study of animal P450 enzymes' metabolism, reducing the number of animals sacrificed in preclinical tests. PMID:26092534
NASA Technical Reports Server (NTRS)
Miller, Franklin; Bagdanove, paul; Blake, Peter; Canavan, Ed; Cofie, Emmanuel; Crane, J. Allen; Dominquez, Kareny; Hagopian, John; Johnston, John; Madison, Tim; Miller, Dave; Oaks, Darrell; Williams, Pat; Young, Dan; Zukowski, Barbara; Zukowski, Tim
2007-01-01
The James Webb Space Telescope Instrument Support Integration Module (ISIM) is being designed and developed at the Goddard Space Flight Center. The ISM Thermal Distortion Testing (ITDT) program was started with the primary objective to validate the ISM mechanical design process. The ITDT effort seeks to establish confidence and demonstrate the ability to predict thermal distortion in composite structures at cryogenic temperatures using solid element models. This-program's goal is to better ensure that ISIM meets all the mechanical and structural requirements by using test results to verify or improve structural modeling techniques. The first step to accomplish the ITDT objectives was to design, and then construct solid element models of a series 2-D test assemblies that represent critical building blocks of the ISIM structure. Second, the actual test assemblies consisting of composite tubes and invar end fittings were fabricated and tested for thermal distortion. This paper presents the development of the GSFC Cryo Distortion Measurement Facility (CDMF) to meet the requirements of the ISIM 2-D test. assemblies, and other future ISIM testing needs. The CDMF provides efficient cooling with both a single, and two-stage cryo-cooler. Temperature uniformity of the test assemblies during thermal transients and at steady state is accomplished by using sapphire windows for all of the optical ports on the radiation shields and by using .thermal straps to cool the test assemblies. Numerical thermal models of the test assemblies were used to predict the temperature uniformity of the parts during cooldown and at steady state. Results of these models are compared to actual temperature data from the tests. Temperature sensors with a 0.25K precision were used to insure that test assembly gradients did not exceed 2K lateral, and 4K axially. The thermal distortions of two assemblies were measured during six thermal cycles from 320K to 35K using laser interferometers. The standard
Demidenko, Eugene
2011-01-01
An analytic solution of the potential distribution on a 2D homogeneous disk for electrical impedance tomography under the complete electrode model is expressed via an infinite system of linear equations. For the shunt electrode model with two electrodes, our solution coincides with the previously derived solution expressed via elliptic integral (Pidcock et al 1995). The Dirichlet-to-Neumann map is derived for statistical estimation via nonlinear least squares. The solution is validated in phantom experiments and applied for breast contact impedance estimation in vivo. Statistical hypothesis testing is used to test whether the contact impedances are the same across electrodes or all equal zero. Our solution can be especially useful for a rapid real-time test for bad surface contact in clinical setting. PMID:21799240
Testing Times: Problems Arising from Misdiagnosis.
ERIC Educational Resources Information Center
Vialle, Wilma; Konza, Deslea
1997-01-01
Three case studies illustrate problems in the identification of gifted students when tests are not used appropriately. The paper concludes that testing must occur within the context of intensive observations of and discussions with the child and family. The importance of all teachers receiving training in gifted education is stressed. (DB)
Transport Test Problems for Hybrid Methods Development
Shaver, Mark W.; Miller, Erin A.; Wittman, Richard S.; McDonald, Benjamin S.
2011-12-28
This report presents 9 test problems to guide testing and development of hybrid calculations for the ADVANTG code at ORNL. These test cases can be used for comparing different types of radiation transport calculations, as well as for guiding the development of variance reduction methods. Cases are drawn primarily from existing or previous calculations with a preference for cases which include experimental data, or otherwise have results with a high level of confidence, are non-sensitive, and represent problem sets of interest to NA-22.
Rogojerov, Marin; Keresztury, Gábor; Kamenova-Nacheva, Mariana; Sundius, Tom
2012-12-01
A new analytical approach for improving the precision in determination of vibrational transition moment directions of low symmetry molecules (lacking orthogonal axes) is discussed in this paper. The target molecules are partially uniaxially oriented in nematic liquid crystalline solvent and are studied by IR absorption spectroscopy using polarized light. The fundamental problem addressed is that IR linear dichroism measurements of low symmetry molecules alone cannot provide sufficient information on molecular orientation and transition moment directions. It is shown that computational prediction of these quantities can supply relevant complementary data, helping to reveal the hidden information content and achieve a more meaningful and more precise interpretation of the measured dichroic ratios. The combined experimental and theoretical/computational method proposed by us recently for determination of the average orientation of molecules with C(s) symmetry has now been replaced by a more precise analytical approach. The new method introduced and discussed in full detail here uses a mathematically evaluated angle between two vibrational transition moment vectors as a reference. The discussion also deals with error analysis and estimation of uncertainties of the orientational parameters. The proposed procedure has been tested in an analysis of the infrared linear dichroism (IR-LD) spectra of 1-D- and 2-D-naphthalene complemented with DFT calculations using the scaled quantum mechanical force field (SQM FF) method. PMID:22981590
Problem-Solving Test: Tryptophan Operon Mutants
ERIC Educational Resources Information Center
Szeberenyi, Jozsef
2010-01-01
This paper presents a problem-solving test that deals with the regulation of the "trp" operon of "Escherichia coli." Two mutants of this operon are described: in mutant A, the operator region of the operon carries a point mutation so that it is unable to carry out its function; mutant B expresses a "trp" repressor protein unable to bind…
Knowledge dimensions in hypothesis test problems
NASA Astrophysics Data System (ADS)
Krishnan, Saras; Idris, Noraini
2012-05-01
The reformation in statistics education over the past two decades has predominantly shifted the focus of statistical teaching and learning from procedural understanding to conceptual understanding. The emphasis of procedural understanding is on the formulas and calculation procedures. Meanwhile, conceptual understanding emphasizes students knowing why they are using a particular formula or executing a specific procedure. In addition, the Revised Bloom's Taxonomy offers a twodimensional framework to describe learning objectives comprising of the six revised cognition levels of original Bloom's taxonomy and four knowledge dimensions. Depending on the level of complexities, the four knowledge dimensions essentially distinguish basic understanding from the more connected understanding. This study identifiesthe factual, procedural and conceptual knowledgedimensions in hypothesis test problems. Hypothesis test being an important tool in making inferences about a population from sample informationis taught in many introductory statistics courses. However, researchers find that students in these courses still have difficulty in understanding the underlying concepts of hypothesis test. Past studies also show that even though students can perform the hypothesis testing procedure, they may not understand the rationale of executing these steps or know how to apply them in novel contexts. Besides knowing the procedural steps in conducting a hypothesis test, students must have fundamental statistical knowledge and deep understanding of the underlying inferential concepts such as sampling distribution and central limit theorem. By identifying the knowledge dimensions of hypothesis test problems in this study, suitable instructional and assessment strategies can be developed in future to enhance students' learning of hypothesis test as a valuable inferential tool.
Motor operated valves problems tests and simulations
Pinier, D.; Haas, J.L.
1996-12-01
An analysis of the two refusals of operation of the EAS recirculation shutoff valves enabled two distinct problems to be identified on the motorized valves: the calculation methods for the operating torques of valves in use in the power plants are not conservative enough, which results in the misadjustement of the torque limiters installed on their motorizations, the second problem concerns the pressure locking phenomenon: a number of valves may entrap a pressure exceeding the in-line pressure between the disks, which may cause a jamming of the valve. EDF has made the following approach to settle the first problem: determination of the friction coefficients and the efficiency of the valve and its actuator through general and specific tests and models, definition of a new calculation method. In order to solve the second problem, EDF has made the following operations: identification of the valves whose technology enables the pressure to be entrapped: the tests and numerical simulations carried out in the Research and Development Division confirm the possibility of a {open_quotes}boiler{close_quotes} effect: determination of the necessary modifications: development and testing of anti-boiler effect systems.
NASA Astrophysics Data System (ADS)
Biondi, Marco; Guarnieri, Daniela; Yu, Hui; Belli, Valentina; Netti, Paolo Antonio
2013-02-01
A big challenge in tumor targeting by nanoparticles (NPs), taking advantage of the enhanced permeability and retention effect, is the fabrication of small size devices for enhanced tumor penetration, which is considered fundamental to improve chemotherapy efficacy. The purposes of this study are (i) to engineer the formulation of doxorubicin-loaded poly(d,l-lactic-co-glycolic acid) (PLGA)-block-poly(ethylene glycol) (PEG) NPs to obtain <100 nm devices and (ii) to translate standard 2D cytotoxicity studies to 3D collagen systems in which an initial step gradient of the NPs is present. Doxorubicin release can be prolonged for days to weeks depending on the NP formulation and the pH of the release medium. Sub-100 nm NPs are effectively internalized by HeLa cells in 2D and are less cytotoxic than free doxorubicin. In 3D, <100 nm NPs are significantly more toxic than larger ones towards HeLa cells, and the cell death rate is affected by the contributions of drug release and device transport through collagen. Thus, the reduction of NP size is a fundamental feature from both a technological and a biological point of view and must be properly engineered to optimize the tumor response to the NPs.
Farmer, M. T.; Lomperski, S.; Aeschlimann, R. W.; Basu, S.
2011-05-23
The Melt Attack and Coolability Experiments (MACE) program addressed the issue of the ability of water to cool and thermally stabilize a molten core-concrete interaction when the reactants are flooded from above. These tests provided data regarding the nature of corium interactions with concrete, the heat transfer rates from the melt to the overlying water pool, and the role of noncondensable gases in the mixing processes that contribute to melt quenching. As a follow-on program to MACE, The Melt Coolability and Concrete Interaction Experiments (MCCI) project is conducting reactor material experiments and associated analysis to achieve the following objectives: (1) resolve the ex-vessel debris coolability issue through a program that focuses on providing both confirmatory evidence and test data for the coolability mechanisms identified in MACE integral effects tests, and (2) address remaining uncertainties related to long-term two-dimensional molten coreconcrete interactions under both wet and dry cavity conditions. Achievement of these two program objectives will demonstrate the efficacy of severe accident management guidelines for existing plants, and provide the technical basis for better containment designs for future plants. In terms of satisfying these objectives, the Management Board (MB) approved the conduct of two long-term 2-D Core-Concrete Interaction (CCI) experiments designed to provide information in several areas, including: (i) lateral vs. axial power split during dry core-concrete interaction, (ii) integral debris coolability data following late phase flooding, and (iii) data regarding the nature and extent of the cooling transient following breach of the crust formed at the melt-water interface. This data report provides thermal hydraulic test results from the CCI-1 experiment, which was conducted on December 19, 2003. Test specifications for CCI-1 are provided in Table 1-1. This experiment investigated the interaction of a fully oxidized 400 kg
Farmer, M. T.; Lomperski, S.; Kilsdonk, D. J.; Aeschlimann, R. W.; Basu, S.
2011-05-23
The Melt Attack and Coolability Experiments (MACE) program addressed the issue of the ability of water to cool and thermally stabilize a molten core-concrete interaction when the reactants are flooded from above. These tests provided data regarding the nature of corium interactions with concrete, the heat transfer rates from the melt to the overlying water pool, and the role of noncondensable gases in the mixing processes that contribute to melt quenching. As a follow-on program to MACE, The Melt Coolability and Concrete Interaction Experiments (MCCI) project is conducting reactor material experiments and associated analysis to achieve the following objectives: (1) resolve the ex-vessel debris coolability issue through a program that focuses on providing both confirmatory evidence and test data for the coolability mechanisms identified in MACE integral effects tests, and (2) address remaining uncertainties related to long-term two-dimensional molten core-concrete interactions under both wet and dry cavity conditions. Achievement of these two program objectives will demonstrate the efficacy of severe accident management guidelines for existing plants, and provide the technical basis for better containment designs for future plants. In terms of satisfying these objectives, the Management Board (MB) approved the conduct of a third long-term 2-D Core-Concrete Interaction (CCI) experiment designed to provide information in several areas, including: (i) lateral vs. axial power split during dry core-concrete interaction, (ii) integral debris coolability data following late phase flooding, and (iii) data regarding the nature and extent of the cooling transient following breach of the crust formed at the melt-water interface. This data report provides thermal hydraulic test results from the CCI-3 experiment, which was conducted on September 22, 2005. Test specifications for CCI-3 are provided in Table 1-1. This experiment investigated the interaction of a fully oxidized 375
Farmer, M. T.; Lomperski, S.; Kilsdonk, D. J.; Aeschlimann, R. W.; Basu, S.
2011-05-23
The Melt Attack and Coolability Experiments (MACE) program addressed the issue of the ability of water to cool and thermally stabilize a molten core-concrete interaction when the reactants are flooded from above. These tests provided data regarding the nature of corium interactions with concrete, the heat transfer rates from the melt to the overlying water pool, and the role of noncondensable gases in the mixing processes that contribute to melt quenching. As a follow-on program to MACE, The Melt Coolability and Concrete Interaction Experiments (MCCI) project is conducting reactor material experiments and associated analysis to achieve the following objectives: (1) resolve the ex-vessel debris coolability issue through a program that focuses on providing both confirmatory evidence and test data for the coolability mechanisms identified in MACE integral effects tests, and (2) address remaining uncertainties related to long-term two-dimensional molten core-concrete interactions under both wet and dry cavity conditions. Achievement of these two program objectives will demonstrate the efficacy of severe accident management guidelines for existing plants, and provide the technical basis for better containment designs for future plants. In terms of satisfying these objectives, the Management Board (MB) approved the conduct of two long-term 2-D Core-Concrete Interaction (CCI) experiments designed to provide information in several areas, including: (i) lateral vs. axial power split during dry core-concrete interaction, (ii) integral debris coolability data following late phase flooding, and (iii) data regarding the nature and extent of the cooling transient following breach of the crust formed at the melt-water interface. This data report provides thermal hydraulic test results from the CCI-2 experiment, which was conducted on August 24, 2004. Test specifications for CCI-2 are provided in Table 1-1. This experiment investigated the interaction of a fully oxidized 400 kg
Llop, Jordi; Gil, Emilio; Llorens, Jordi; Miranda-Fuentes, Antonio; Gallart, Montserrat
2016-01-01
Canopy characterization is essential for pesticide dosage adjustment according to vegetation volume and density. It is especially important for fresh exportable vegetables like greenhouse tomatoes. These plants are thin and tall and are planted in pairs, which makes their characterization with electronic methods difficult. Therefore, the accuracy of the terrestrial 2D LiDAR sensor is evaluated for determining canopy parameters related to volume and density and established useful correlations between manual and electronic parameters for leaf area estimation. Experiments were performed in three commercial tomato greenhouses with a paired plantation system. In the electronic characterization, a LiDAR sensor scanned the plant pairs from both sides. The canopy height, canopy width, canopy volume, and leaf area were obtained. From these, other important parameters were calculated, like the tree row volume, leaf wall area, leaf area index, and leaf area density. Manual measurements were found to overestimate the parameters compared with the LiDAR sensor. The canopy volume estimated with the scanner was found to be reliable for estimating the canopy height, volume, and density. Moreover, the LiDAR scanner could assess the high variability in canopy density along rows and hence is an important tool for generating canopy maps. PMID:27608025
The linear separability problem: some testing methods.
Elizondo, D
2006-03-01
The notion of linear separability is used widely in machine learning research. Learning algorithms that use this concept to learn include neural networks (single layer perceptron and recursive deterministic perceptron), and kernel machines (support vector machines). This paper presents an overview of several of the methods for testing linear separability between two classes. The methods are divided into four groups: Those based on linear programming, those based on computational geometry, one based on neural networks, and one based on quadratic programming. The Fisher linear discriminant method is also presented. A section on the quantification of the complexity of classification problems is included. PMID:16566462
Word Problems: Where Test Bias Creeps In.
ERIC Educational Resources Information Center
Chipman, Susan F.
The problem of sex bias in mathematics word problems is discussed, with references to the appropriate literature. Word problems are assessed via cognitive science analysis of word problem solving. It has been suggested that five basic semantic relations are adequate to classify nearly all story problems, namely, change, combine, compare, vary, and…
Radix, P.; Leonard, M.; Papantoniou, C.; Roman, G.; Saouter, E.; Gallotti-Schmitt, S.; Thiebaud, H.; Vasseur, P.
1999-10-01
The Daphnia magna 21-d test may be required by European authorities as a criterion for the assessment of aquatic chronic toxicity for the notification of new substances. However, this test has several drawbacks. It is labor-intensive, relatively expensive, and requires the breeding of test organisms. The Brachionous calyciflorus 2-d test and Microtox chronic 22-h test do not suffer from these disadvantages and could be used as substitutes for the Daphnia 21-d test for screening assays. During this study, the toxicity of 25 chemicals was measured using both the microtox chronic toxicity and B. calyciflorus 2-d tests, and the no-observed-effect concentrations (NOECs) were compared to the D. magna 21-d test. The Brachionus test was slightly less sensitive than the Daphnia test, but the correlation between the two tests was relatively good (r{sup 2} = 0.54). The B. calyciflorus 2-d test, and to a lesser extent the Microtox chronic 22-h test, were able to predict the chronic toxicity values of the Daphnia 21-d test. They constitute promising cost-effective tools for chronic toxicity screening.
Li, Yan; Zhu, Zhuo R; Ou, Bao C; Wang, Ya Q; Tan, Zhou B; Deng, Chang M; Gao, Yi Y; Tang, Ming; So, Ji H; Mu, Yang L; Zhang, Lan Q
2015-02-15
Major depressive disorder is one of the most prevalent and life-threatening forms of mental illnesses. The traditional antidepressants often take several weeks, even months, to obtain clinical effects. However, recent clinical studies have shown that ketamine, an N-methyl-D-aspartate (NMDA) receptor antagonist, exerts rapid antidepressant effects within 2h and are long-lasting. The aim of the present study was to investigate whether dopaminergic system was involved in the rapid antidepressant effects of ketamine. The acute administration of ketamine (20 mg/kg) significantly reduced the immobility time in the forced swim test. MK-801 (0.1 mg/kg), the more selective NMDA antagonist, also exerted rapid antidepressant-like effects. In contrast, fluoxetine (10 mg/kg) did not significantly reduced the immobility time in the forced swim test after 30 min administration. Notably, pretreatment with haloperidol (0.15 mg/kg, a nonselective dopamine D2/D3 antagonist), but not SCH23390 (0.04 and 0.1 mg/kg, a selective dopamine D1 receptor antagonist), significantly prevented the effects of ketamine or MK-801. Moreover, the administration of sub-effective dose of ketamine (10 mg/kg) in combination with pramipexole (0.3 mg/kg, a dopamine D2/D3 receptor agonist) exerted antidepressant-like effects compared with each drug alone. In conclusion, our results indicated that the dopamine D2/D3 receptors, but not D1 receptors, are involved in the rapid antidepressant-like effects of ketamine. PMID:25449845
Zelt, Colin A.; Haines, Seth; Powers, Michael H.; Sheehan, Jacob; Rohdewald, Siegfried; Link, Curtis; Hayashi, Koichi; Zhao, Don; Zhou, Hua-wei; Burton, Bethany L.; Petersen, Uni K.; Bonal, Nedra D.; Doll, William E.
2013-01-01
Seismic refraction methods are used in environmental and engineering studies to image the shallow subsurface. We present a blind test of inversion and tomographic refraction analysis methods using a synthetic first-arrival-time dataset that was made available to the community in 2010. The data are realistic in terms of the near-surface velocity model, shot-receiver geometry and the data's frequency and added noise. Fourteen estimated models were determined by ten participants using eight different inversion algorithms, with the true model unknown to the participants until it was revealed at a session at the 2011 SAGEEP meeting. The estimated models are generally consistent in terms of their large-scale features, demonstrating the robustness of refraction data inversion in general, and the eight inversion algorithms in particular. When compared to the true model, all of the estimated models contain a smooth expression of its two main features: a large offset in the bedrock and the top of a steeply dipping low-velocity fault zone. The estimated models do not contain a subtle low-velocity zone and other fine-scale features, in accord with conventional wisdom. Together, the results support confidence in the reliability and robustness of modern refraction inversion and tomographic methods.
A class of ejecta transport test problems
Hammerberg, James E; Buttler, William T; Oro, David M; Rousculp, Christopher L; Morris, Christopher; Mariam, Fesseha G
2011-01-31
Hydro code implementations of ejecta dynamics at shocked interfaces presume a source distribution function ofparticulate masses and velocities, f{sub 0}(m, v;t). Some of the properties of this source distribution function have been determined from extensive Taylor and supported wave experiments on shock loaded Sn interfaces of varying surface and subsurface morphology. Such experiments measure the mass moment of f{sub o} under vacuum conditions assuming weak particle-particle interaction and, usually, fully inelastic capture by piezo-electric diagnostic probes. Recently, planar Sn experiments in He, Ar, and Kr gas atmospheres have been carried out to provide transport data both for machined surfaces and for coated surfaces. A hydro code model of ejecta transport usually specifies a criterion for the instantaneous temporal appearance of ejecta with source distribution f{sub 0}(m, v;t{sub 0}). Under the further assumption of separability, f{sub 0}(m,v;t{sub 0}) = f{sub 1}(m)f{sub 2}(v), the motion of particles under the influence of gas dynamic forces is calculated. For the situation of non-interacting particulates, interacting with a gas via drag forces, with the assumption of separability and simplified approximations to the Reynolds number dependence of the drag coefficient, the dynamical equation for the time evolution of the distribution function, f(r,v,m;t), can be resolved as a one-dimensional integral which can be compared to a direct hydro simulation as a test problem. Such solutions can also be used for preliminary analysis of experimental data. We report solutions for several shape dependent drag coefficients and analyze the results of recent planar dsh experiments in Ar and Xe.
Testing Developmental Pathways to Antisocial Personality Problems
ERIC Educational Resources Information Center
Diamantopoulou, Sofia; Verhulst, Frank C.; van der Ende, Jan
2010-01-01
This study examined the development of antisocial personality problems (APP) in young adulthood from disruptive behaviors and internalizing problems in childhood and adolescence. Parent ratings of 507 children's (aged 6-8 years) symptoms of attention deficit hyperactivity disorder, oppositional defiant disorder, and anxiety, were linked to…
Opdam, F L; Modak, A S; Gelderblom, H; Guchelaar, H J
2015-06-01
In a previous study, we found that the CYP2D6 phenotype determined by (13)C-dextromethorphan breath test (DM-BT) might be used to predict tamoxifen treatment outcome in breast cancer patients in the adjuvant setting. However, large variation in the delta-over-baseline (DOB) values was observed in the extensive metabolizer predicted phenotype group based on single point measures. In the present work we aimed to analyze the variability of phenotype results and determine reproducibility to further characterize the clinical utility of DM-BT by introducing multiple breath sampling instead of single breath sampling and by administration of a fixed dose of (13)C-DM. PMID:25891764
Hoffman, E.L.; Ammerman, D.J.
1993-08-01
A series of tests investigating dynamic pulse buckling of a cylindrical shell under axial impact is compared to several finite element simulations of the event. The purpose of the study is to compare the performance of the various analysis codes and element types with respect to a problem which is applicable to radioactive material transport packages, and ultimately to develop a benchmark problem to qualify finite element analysis codes for the transport package design industry.
Problems and Issues in Translating International Educational Achievement Tests
ERIC Educational Resources Information Center
Arffman, Inga
2013-01-01
The article reviews research and findings on problems and issues faced when translating international academic achievement tests. The purpose is to draw attention to the problems, to help to develop the procedures followed when translating the tests, and to provide suggestions for further research. The problems concentrate on the following: the…
ERIC Educational Resources Information Center
Hill, Kennedy T.; Horton, Margaret W.
Educational solutions to the problem of test anxiety were explored. Test anxiety has a debilitating effect on performance which increases over the school years. The solution is, first, to measure test anxiety so that the extent of it, as well as the effectiveness of programs designed to alleviate it, can be measured. The seven-item Comfort Index,…
NASA Astrophysics Data System (ADS)
Gnanvo, Kondo; Bai, Xinzhan; Gu, Chao; Liyanage, Nilanga; Nelyubin, Vladimir; Zhao, Yuxiang
2016-02-01
A large-area and light-weight gas electron multiplier (GEM) detector was built at the University of Virginia as a prototype for the detector R&D program of the future Electron Ion Collider. The prototype has a trapezoidal geometry designed as a generic sector module in a disk layer configuration of a forward tracker in collider detectors. It is based on light-weight material and narrow support frames in order to minimize multiple scattering and dead-to-sensitive area ratio. The chamber has a novel type of two dimensional (2D) stereo-angle readout board with U-V strips that provides (r,φ) position information in the cylindrical coordinate system of a collider environment. The prototype was tested at the Fermilab Test Beam Facility in October 2013 and the analysis of the test beam data demonstrates an excellent response uniformity of the large area chamber with an efficiency higher than 95%. An angular resolution of 60 μrad in the azimuthal direction and a position resolution better than 550 μm in the radial direction were achieved with the U-V strip readout board. The results are discussed in this paper.
Errors in Standardized Tests: A Systemic Problem.
ERIC Educational Resources Information Center
Rhoades, Kathleen; Madaus, George
The nature and extent of human error in educational testing over the past 25 years were studied. In contrast to the random measurement error expected in all tests, the presence of human error is unexpected and brings unknown, often harmful, consequences for students and their schools. Using data from a variety of sources, researchers found 103…
Test problem construction for single-objective bilevel optimization.
Sinha, Ankur; Malo, Pekka; Deb, Kalyanmoy
2014-01-01
In this paper, we propose a procedure for designing controlled test problems for single-objective bilevel optimization. The construction procedure is flexible and allows its user to control the different complexities that are to be included in the test problems independently of each other. In addition to properties that control the difficulty in convergence, the procedure also allows the user to introduce difficulties caused by interaction of the two levels. As a companion to the test problem construction framework, the paper presents a standard test suite of 12 problems, which includes eight unconstrained and four constrained problems. Most of the problems are scalable in terms of variables and constraints. To provide baseline results, we have solved the proposed test problems using a nested bilevel evolutionary algorithm. The results can be used for comparison, while evaluating the performance of any other bilevel optimization algorithm. The code related to the paper may be accessed from the website http://bilevel.org . PMID:24364674
Problem-solving test: Tryptophan operon mutants.
Szeberényi, József
2010-09-01
Terms to be familiar with before you start to solve the test: tryptophan, operon, operator, repressor, inducer, corepressor, promoter, RNA polymerase, chromosome-polysome complex, regulatory gene, cis-acting element, trans-acting element, plasmid, transformation. PMID:21567855
ESTL: Innovative Solutions to Tribology Test Problems
NASA Astrophysics Data System (ADS)
Roberts, E.; Eiden, M.
2004-08-01
For over 30 years, ESTL, through the financial and technical support of ESA and ESTEC, has provided a unique service to the European space industry by ensuring the reliability of the moving parts of spacecraft mechanisms through the application of sound tribology. ESTL's activities range from fundamental measurements of adhesion, friction and wear of material couples to the full qualification and life-testing of primary spacecraft mechanisms. In all cases, test work is carried out under conditions that simulate the thermal and vacuum conditions that prevail in the space environment. Occasionally there have arisen specific measurement requirements which have proved challenging under the constraints imposed by thermal- vacuum conditions. How ESTL has met some of its more demanding test requirements, and the significance of the test results obtained, is the subject of this paper.
NASA Astrophysics Data System (ADS)
Ekberg, Peter; Stiblert, Lars; Mattsson, Lars
2014-05-01
The manufacturing of flat panel displays requires a number of photomasks for the placement of pixel patterns and supporting transistor arrays. For large area photomasks, dedicated ultra-precision writers have been developed for the production of these chromium patterns on glass or quartz plates. The dimensional tolerances in X and Y for absolute pattern placement on these plates, with areas measured in square meters, are in the range of 200-300 nm (3σ). To verify these photomasks, 2D ultra-precision coordinate measurement machines are used having even tighter tolerance requirements. This paper will present how the world standard metrology tool used for verifying large masks, the Micronic Mydata MMS15000, is calibrated without any other references than the wavelength of the interferometers in an extremely well-controlled temperature environment. This process is called self-calibration and is the only way to calibrate the metrology tool, as no square-meter-sized large area 2D traceable artifact is available. The only parameter that cannot be found using self-calibration is the absolute length scale. To make the MMS15000 traceable, a 1D reference rod, calibrated at a national metrology lab, is used. The reference plates used in the calibration of the MMS15000 may have sizes up to 1 m2 and a weight of 50 kg. Therefore, standard methods for self-calibration on a small scale with exact placements cannot be used in the large area case. A new, more general method had to be developed for the purpose of calibrating the MMS15000. Using this method, it is possible to calibrate the measurement tool down to an uncertainty level of <90 nm (3σ) over an area of (0.8 × 0.8) m2. The method used, which is based on the concept of iteration, does not introduce any more noise than the random noise introduced by the measurements, resulting in the lowest possible noise level that can be achieved by any self-calibration method.
ERIC Educational Resources Information Center
Veldkamp, Bernard P.; Verschoor, Angela J.; Eggen, Theo J. H. M.
2010-01-01
Overexposure and underexposure of items in the bank are serious problems in operational computerized adaptive testing (CAT) systems. These exposure problems might result in item compromise, or point at a waste of investments. The exposure control problem can be viewed as a test assembly problem with multiple objectives. Information in the test has…
NASA Astrophysics Data System (ADS)
Morgan, J. P.; de Monserrat, A.; Hall, R.; Taramon, J. M.; Perez-Gussinye, M.
2015-12-01
This work focuses on improving current 2D numerical approaches to modeling the boundary conditions associated with computing accurate deformation and melting associated with continental rifting. Recent models primarily use far-field boundary conditions that have been used for decades with little assessment of their effects on asthenospheric flow beneath the rifting region. All are clearly extremely oversimplified — Huismans and Buiter assume there is no vertical flow into the rifting region, with the asthenosphere flowing uniformly into the rifting region from the sides beneath lithosphere moving in the opposing direction, Armitage et al. and van Wijk use divergent velocities on the upper boundary to impose break-up within a Cartesian box, while other studies generally assume there is uniform horizontal flow away from the center of rifting, with uniform vertical flow replenishing the material pulled out of the sides of the computational region. All are likely to significantly shape the pattern of asthenospheric flow beneath the stretching lithosphere that is associated with pressure-release melting and rift volcanism. Thus while ALL may lead to similar predictions of the effects of crustal stretching and thinning, NONE may lead to accurate determination of the the asthenospheric flow and melting associated with lithospheric stretching and breakup. Here we discuss a suite of numerical experiments that compare these choices to likely more realistic boundary condition choices like the analytical solution for flow associated with two diverging plates stretching over a finite-width region, and a high-resolution 2-D region embedded within a cylindrical annulus 'whole mantle cross-section' at 5% extra numerical problem size. Our initial results imply that the choice of far-field boundary conditions does indeed significantly influence predicted melting distributions and melt volumes associated with continental breakup. For calculations including asthenospheric melting
Testing general relativity: Progress, problems, and prospects
NASA Technical Reports Server (NTRS)
Shapiro, I. I.
1971-01-01
The results from ground-based experimental testing are presented. Prospects for improving these experiments are discussed. Radar echo time delays, perihelion advance and solar oblateness, time variation of the gravitational constant, and radio wave deflection are considered. Ground-based and spacecraft techniques are compared on an accuracy vs. cost basis.
Problem-Solving Test: Southwestern Blotting
ERIC Educational Resources Information Center
Szeberényi, József
2014-01-01
Terms to be familiar with before you start to solve the test: Southern blotting, Western blotting, restriction endonucleases, agarose gel electrophoresis, nitrocellulose filter, molecular hybridization, polyacrylamide gel electrophoresis, proto-oncogene, c-abl, Src-homology domains, tyrosine protein kinase, nuclear localization signal, cDNA,…
Problem-Solving Test: Restriction Endonuclease Mapping
ERIC Educational Resources Information Center
Szeberenyi, Jozsef
2011-01-01
The term "restriction endonuclease mapping" covers a number of related techniques used to identify specific restriction enzyme recognition sites on small DNA molecules. A method for restriction endonuclease mapping of a 1,000-basepair (bp)-long DNA molecule is described in the fictitious experiment of this test. The most important fact needed to…
American History's Problem with Standardized Testing
ERIC Educational Resources Information Center
McCoog, Ian J.
2005-01-01
This article looks at current research concerning how students best learn the discipline of history, commentaries both in favor of and against standardized testing, and basic philosophical beliefs about the discipline. It explains methods of how to incorporate differentiated lessons and performance based assessments to NCLB standards and…
Crash test for the Copenhagen problem.
Nagler, Jan
2004-06-01
The Copenhagen problem is a simple model in celestial mechanics. It serves to investigate the behavior of a small body under the gravitational influence of two equally heavy primary bodies. We present a partition of orbits into classes of various kinds of regular motion, chaotic motion, escape and crash. Collisions of the small body onto one of the primaries turn out to be unexpectedly frequent, and their probability displays a scale-free dependence on the size of the primaries. The analysis reveals a high degree of complexity so that long term prediction may become a formidable task. Moreover, we link the results to chaotic scattering theory and the theory of leaking Hamiltonian systems. PMID:15244719
Group Testing: Four Student Solutions to a Classic Optimization Problem
ERIC Educational Resources Information Center
Teague, Daniel
2006-01-01
This article describes several creative solutions developed by calculus and modeling students to the classic optimization problem of testing in groups to find a small number of individuals who test positive in a large population.
2005-07-01
Aniso2d is a two-dimensional seismic forward modeling code. The earth is parameterized by an X-Z plane in which the seismic properties Can have monoclinic with x-z plane symmetry. The program uses a user define time-domain wavelet to produce synthetic seismograms anrwhere within the two-dimensional media.
Execution of Multidisciplinary Design Optimization Approaches on Common Test Problems
NASA Technical Reports Server (NTRS)
Balling, R. J.; Wilkinson, C. A.
1997-01-01
A class of synthetic problems for testing multidisciplinary design optimization (MDO) approaches is presented. These test problems are easy to reproduce because all functions are given as closed-form mathematical expressions. They are constructed in such a way that the optimal value of all variables and the objective is unity. The test problems involve three disciplines and allow the user to specify the number of design variables, state variables, coupling functions, design constraints, controlling design constraints, and the strength of coupling. Several MDO approaches were executed on two sample synthetic test problems. These approaches included single-level optimization approaches, collaborative optimization approaches, and concurrent subspace optimization approaches. Execution results are presented, and the robustness and efficiency of these approaches an evaluated for these sample problems.
Greg Flach, Frank Smith
2011-12-31
Mesh2d is a Fortran90 program designed to generate two-dimensional structured grids of the form [x(i),y(i,j)] where [x,y] are grid coordinates identified by indices (i,j). The x(i) coordinates alone can be used to specify a one-dimensional grid. Because the x-coordinates vary only with the i index, a two-dimensional grid is composed in part of straight vertical lines. However, the nominally horizontal y(i,j0) coordinates along index i are permitted to undulate or otherwise vary. Mesh2d also assigns an integer material type to each grid cell, mtyp(i,j), in a user-specified manner. The complete grid is specified through three separate input files defining the x(i), y(i,j), and mtyp(i,j) variations.
2011-12-31
Mesh2d is a Fortran90 program designed to generate two-dimensional structured grids of the form [x(i),y(i,j)] where [x,y] are grid coordinates identified by indices (i,j). The x(i) coordinates alone can be used to specify a one-dimensional grid. Because the x-coordinates vary only with the i index, a two-dimensional grid is composed in part of straight vertical lines. However, the nominally horizontal y(i,j0) coordinates along index i are permitted to undulate or otherwise vary. Mesh2d also assignsmore » an integer material type to each grid cell, mtyp(i,j), in a user-specified manner. The complete grid is specified through three separate input files defining the x(i), y(i,j), and mtyp(i,j) variations.« less
NASA Astrophysics Data System (ADS)
Lotsch, Bettina V.
2015-07-01
Graphene's legacy has become an integral part of today's condensed matter science and has equipped a whole generation of scientists with an armory of concepts and techniques that open up new perspectives for the postgraphene area. In particular, the judicious combination of 2D building blocks into vertical heterostructures has recently been identified as a promising route to rationally engineer complex multilayer systems and artificial solids with intriguing properties. The present review highlights recent developments in the rapidly emerging field of 2D nanoarchitectonics from a materials chemistry perspective, with a focus on the types of heterostructures available, their assembly strategies, and their emerging properties. This overview is intended to bridge the gap between two major—yet largely disjunct—developments in 2D heterostructures, which are firmly rooted in solid-state chemistry or physics. Although the underlying types of heterostructures differ with respect to their dimensions, layer alignment, and interfacial quality, there is common ground, and future synergies between the various assembly strategies are to be expected.
Group Work Tests for Context-Rich Problems
ERIC Educational Resources Information Center
Meyer, Chris
2016-01-01
The group work test is an assessment strategy that promotes higher-order thinking skills for solving context-rich problems. With this format, teachers are able to pose challenging, nuanced questions on a test, while providing the support weaker students need to get started and show their understanding. The test begins with a group discussion…
Development of a Test of Experimental Problem-Solving Skills.
ERIC Educational Resources Information Center
Ross, John A.; Maynes, Florence J.
1983-01-01
Multiple-choice tests were constructed for seven problem-solving skills using learning hierarchies based on expert-novice differences and refined in three phases of field testing. Includes test reliabilities (sufficient for making judgments of group performance but insufficient in single-administration for individual assessment), validity, and…
Problems and Alternatives in Testing Mexican American Students.
ERIC Educational Resources Information Center
Cervantes, Robert A.
The problems of standardized tests with regard to Mexican American students, particularly "ethnic validity", are reviewed. Inadequate norm group representation, cultural bias, and language bias are purported by the author to be the most common faults of standardized tests. Suggested is the elimination of standardized testing as a principal means…
Problems in Testing the Intonation of Advanced Foreign Learners.
ERIC Educational Resources Information Center
Mendelsohn, David
1978-01-01
It is argued that knowledge about the testing of intonation in English as a foreign language is inadequate; the major problems are outlined and tentative suggestions are given. The basic problem is that the traditional foreign language teacher's conception of intonation is limited. A three-part definition of intonation is favored, with suggestions…
Some Current Problems in Simulator Design, Testing and Use.
ERIC Educational Resources Information Center
Caro, Paul W.
Concerned with the general problem of the effectiveness of simulator training, this report reflects information developed during the conduct of aircraft simulator training research projects sponsored by the Air Force, Army, Navy, and Coast Guard. Problems are identified related to simulator design, testing, and use, all of which impact upon…
Invitational Conference on Testing Problems (New York, October 29, 1966).
ERIC Educational Resources Information Center
Educational Testing Service, Princeton, NJ.
The 1966 Invitational Conference on Testing Problems dealt with the innovations of the new age of flexibility and the problems of evaluating and preparing for them. Papers presented in Session I, Innovation and Evaluation, were: (1) "Innovation and Evaluation: In Whose Hands?" by Nils Y. Wessell; (2) "The Discovery and Development of Educational…
A description of the test problems in the TRAC-P standard test matrix
Steinke, R.G.
1996-03-01
This report describes 15 different test problems in the TRAC-P (Transient Reactor Analysis Code) standard test matrix of 42 test-problem calculations. Their TRACIN input-data files are listed in Appendix A. The description of each test problem includes the nature of what the test problem models and evaluates, the principal models of TRAC-P that the test problem serves to verify or validate, and the TRAC-P features and options that are being involved in its calculation. The test-problem calculations will determine the effect that changes made to a TRAC-P version have on the results. This will help the developers assess the acceptance of those changes to TRAC-P.
Brittle damage models in DYNA2D
Faux, D.R.
1997-09-01
DYNA2D is an explicit Lagrangian finite element code used to model dynamic events where stress wave interactions influence the overall response of the system. DYNA2D is often used to model penetration problems involving ductile-to-ductile impacts; however, with the advent of the use of ceramics in the armor-anti-armor community and the need to model damage to laser optics components, good brittle damage models are now needed in DYNA2D. This report will detail the implementation of four brittle damage models in DYNA2D, three scalar damage models and one tensor damage model. These new brittle damage models are then used to predict experimental results from three distinctly different glass damage problems.
Cattaneo, Cristina; Cantatore, Angela; Ciaffi, Romina; Gibelli, Daniele; Cigada, Alfredo; De Angelis, Danilo; Sala, Remo
2012-01-01
Identification from video surveillance systems is frequently requested in forensic practice. The "3D-2D" comparison has proven to be reliable in assessing identification but still requires standardization; this study concerns the validation of the 3D-2D profile comparison. The 3D models of the faces of five individuals were compared with photographs from the same subjects as well as from another 45 individuals. The difference in area and distance between maxima (glabella, tip of nose, fore point of upper and lower lips, pogonion) and minima points (selion, subnasale, stomion, suprapogonion) were measured. The highest difference in area between the 3D model and the 2D image was between 43 and 133 mm(2) in the five matches, always greater than 157 mm(2) in mismatches; the mean distance between the points was greater than 1.96 mm in mismatches, <1.9 mm in five matches (p < 0.05). These results indicate that this difference in areas may point toward a manner of distinguishing "correct" from "incorrect" matches. PMID:22074112
ERIC Educational Resources Information Center
Charalambous, Charalambos; Kyriakides, Leonidas; Philippou, George
2003-01-01
The study reported in this paper is an attempt to develop a comprehensive model of measuring problem solving and posing (PSP) skills based on Marshall's schema theory (ST). A battery of tests on PSP skills was administered to 5th and 6th grade Cypriot students (n=2519). The Rasch model was used and a scale was created for the battery of tests and…
Flat flame olympics: test problem a. Final report
Coffee, T.P.
1982-10-01
This report discusses a test problem for a computer program for numerically solving the equations governing a laminar, premixed, one-dimensional flame. The problem was proposed by GAMM (Committee for Numerical Methods in Fluid Mechanics), and has been solved for presentation at a workshop at the Technical University, Aachen, Germany, 12-14 Oct. 1981. The test problem is an unsteady propagating flame with one-step chemistry and Lewis number different from unity. A code developed for steady state problems with elementary chemistry was modified to use the simplified transport and chemistry of the test problem and to follow the details of the transient solution. The problem is solved for six cases. The cases differ in the Lewis number chosen and the activation energy of the single reaction. The initial conditions used are the steady state solutions predicted by the simplified analytic method of asymptotic analysis. In most cases, the numerical solutions rapidly converge, and the steady state solutions are similar to the asymptotic solutions. However, in one case, with activation energy and Lewis number equal to two, the solution does not converge. Instead, large oscillations in the flame speed and the profiles occur.
Group Work Tests for Context-Rich Problems
NASA Astrophysics Data System (ADS)
Meyer, Chris
2016-05-01
The group work test is an assessment strategy that promotes higher-order thinking skills for solving context-rich problems. With this format, teachers are able to pose challenging, nuanced questions on a test, while providing the support weaker students need to get started and show their understanding. The test begins with a group discussion phase, when students are given a "number-free" version of the problem. This phase allows students to digest the story-like problem, explore solution ideas, and alleviate some test anxiety. After 10-15 minutes of discussion, students inform the instructor of their readiness for the individual part of the test. What follows next is a pedagogical phase change from lively group discussion to quiet individual work. The group work test is a natural continuation of the group work in our daily physics classes and helps reinforce the importance of collaboration. This method has met with success at York Mills Collegiate Institute, in Toronto, Ontario, where it has been used consistently for unit tests and the final exam of the grade 12 university preparation physics course.
A one-loop test for construction of 4D N = 4 SYM from 2D SYM via fuzzy-sphere geometry
NASA Astrophysics Data System (ADS)
Matsuura, So; Sugino, Fumihiko
2016-04-01
As a perturbative check of the construction of 4D N=4 supersymmetric Yang-Mills theory (SYM) from mass-deformed N=(8,8) SYM on the 2D lattice, the one-loop effective action for scalar kinetic terms is computed in N=4 U(k) SYM on R^2 × (fuzzy S^2), which is obtained by expanding 2D N=(8,8) U(N) SYM with mass deformation around its fuzzy-sphere classical solution. The radius of the fuzzy sphere is proportional to the inverse of the mass. We consider two successive limits: (1) decompactify the fuzzy sphere to a noncommutative (Moyal) plane and (2) turn off the noncommutativity of the Moyal plane. It is straightforward at the classical level to obtain the ordinary N=4 SYM on R^4 in the limits, while it is nontrivial at the quantum level. The one-loop effective action for the SU(k) sector of the gauge group U(k) coincides with that of the ordinary 4D N=4 SYM in the above limits. Although a "noncommutative anomaly" appears in the overall U(1) sector of the U(k) gauge group, this can be expected to be a gauge artifact not affecting gauge-invariant observables.
Nyström, Monica E; Terris, Darcey D; Sparring, Vibeke; Tolf, Sara; Brown, Claire R
2012-01-01
Our objective was to test whether the Structured Problem and Success Inventory (SPI) instrument could capture mental representations of organizational and work-related problems as described by individuals working in health care organizations and to test whether these representations varied according to organizational position. A convenience sample (n = 56) of middle managers (n = 20), lower-level managers (n = 20), and staff (n = 16) from health care organizations in Stockholm (Sweden) attending organizational development courses during 2003-2004 was recruited. Participants used the SPI to describe the 3 most pressing organizational and work-related problems. Data were systematically reviewed to identify problem categories and themes. One hundred sixty-four problems were described, clustered into 13 problem categories. Generally, middle managers focused on organizational factors and managerial responsibilities, whereas lower-level managers and staff focused on operational issues and what others did or ought to do. Furthermore, we observed similarities and variation in perceptions and their association with respondents' position within an organization. Our results support the need for further evaluation of the SPI as a promising tool for health care organizations. Collecting structured inventories of organizational and work-related problems from multiple perspectives may assist in the development of shared understandings of organizational challenges and lead to more effective and efficient processes of solution planning and implementation. PMID:22453820
The measurand problem in infrared breath alcohol testing
NASA Astrophysics Data System (ADS)
Vosk, Ted
2012-02-01
Measurements are made to determine the value of a quantity known as a measurand. The measurand is not always the quantity subject to measurement, however. Often, a distinct quantity will be measured and related to the measurand through a measurement function. When the identities of the measurand and the quantity actually measured are not well defined or distinguished, it can lead to the misinterpretation of results. This is referred to as the measurand problem. The measurand problem can present significant difficulties when the law and not science determines the measurand. This arises when the law requires that a particular quantity be measured. Legal definitions are seldom as rigorous or complete as those utilized in science. Thus, legally defined measurands often fall prey to the measurand problem. An example is the measurement of breath alcohol concentration by infrared spectroscopy. All 50 states authorize such tests but the measurand differs by jurisdiction. This leads to misinterpretation of results in both the forensic and legal communities due to the measurand problem with the consequence that the innocent are convicted and guilty set free. Correct interpretation of breath test results requires that the measurand be properly understood and accounted for. I set forth the varying measurands defined by law, the impact these differing measurands have on the interpretation of breath test results and how the measurand problem can be avoided in the measurement of breath alcohol concentration.
2-d Finite Element Code Postprocessor
1996-07-15
ORION is an interactive program that serves as a postprocessor for the analysis programs NIKE2D, DYNA2D, TOPAZ2D, and CHEMICAL TOPAZ2D. ORION reads binary plot files generated by the two-dimensional finite element codes currently used by the Methods Development Group at LLNL. Contour and color fringe plots of a large number of quantities may be displayed on meshes consisting of triangular and quadrilateral elements. ORION can compute strain measures, interface pressures along slide lines, reaction forcesmore » along constrained boundaries, and momentum. ORION has been applied to study the response of two-dimensional solids and structures undergoing finite deformations under a wide variety of large deformation transient dynamic and static problems and heat transfer analyses.« less
MAZE96. Generates 2D Input for DYNA NIKE & TOPAZ
Sanford, L.; Hallquist, J.O.
1992-02-24
MAZE is an interactive program that serves as an input and two-dimensional mesh generator for DYNA2D, NIKE2D, TOPAZ2D, and CHEMICAL TOPAZ2D. MAZE also generates a basic template for ISLAND input. MAZE has been applied to the generation of input data to study the response of two-dimensional solids and structures undergoing finite deformations under a wide variety of large deformation transient dynamic and static problems and heat transfer analyses.
Generates 2D Input for DYNA NIKE & TOPAZ
1996-07-15
MAZE is an interactive program that serves as an input and two-dimensional mesh generator for DYNA2D, NIKE2D, TOPAZ2D, and CHEMICAL TOPAZ2D. MAZE also generates a basic template for ISLAND input. MAZE has been applied to the generation of input data to study the response of two-dimensional solids and structures undergoing finite deformations under a wide variety of large deformation transient dynamic and static problems and heat transfer analyses.
Internal Photoemission Spectroscopy of 2-D Materials
NASA Astrophysics Data System (ADS)
Nguyen, Nhan; Li, Mingda; Vishwanath, Suresh; Yan, Rusen; Xiao, Shudong; Xing, Huili; Cheng, Guangjun; Hight Walker, Angela; Zhang, Qin
Recent research has shown the great benefits of using 2-D materials in the tunnel field-effect transistor (TFET), which is considered a promising candidate for the beyond-CMOS technology. The on-state current of TFET can be enhanced by engineering the band alignment of different 2D-2D or 2D-3D heterostructures. Here we present the internal photoemission spectroscopy (IPE) approach to determine the band alignments of various 2-D materials, in particular SnSe2 and WSe2, which have been proposed for new TFET designs. The metal-oxide-2-D semiconductor test structures are fabricated and characterized by IPE, where the band offsets from the 2-D semiconductor to the oxide conduction band minimum are determined by the threshold of the cube root of IPE yields as a function of photon energy. In particular, we find that SnSe2 has a larger electron affinity than most semiconductors and can be combined with other semiconductors to form near broken-gap heterojunctions with low barrier heights which can produce a higher on-state current. The details of data analysis of IPE and the results from Raman spectroscopy and spectroscopic ellipsometry measurements will also be presented and discussed.
Discuss the testing problems of ultraviolet irradiance meters
NASA Astrophysics Data System (ADS)
Ye, Jun'an; Lin, Fangsheng
2014-09-01
Ultraviolet irradiance meters are widely used in many areas such as medical treatment, epidemic prevention, energy conservation and environment protection, computers, manufacture, electronics, ageing of material and photo-electric effect, for testing ultraviolet irradiance intensity. So the accuracy of value directly affects the sterile control in hospital, treatment, the prevention level of CDC and the control accuracy of curing and aging in manufacturing industry etc. Because the display of ultraviolet irradiance meters is easy to change, in order to ensure the accuracy, it needs to be recalibrated after being used period of time. By the comparison with the standard ultraviolet irradiance meters, which are traceable to national benchmarks, we can acquire the correction factor to ensure that the instruments working under accurate status and giving the accurate measured data. This leads to an important question: what kind of testing device is more accurate and reliable? This article introduces the testing method and problems of the current testing device for ultraviolet irradiance meters. In order to solve these problems, we have developed a new three-dimensional automatic testing device. We introduce structure and working principle of this system and compare the advantages and disadvantages of two devices. In addition, we analyses the errors in the testing of ultraviolet irradiance meters.
Qualification tests and electrical measurements: Practice and problems
NASA Technical Reports Server (NTRS)
Smokler, M. I.
1983-01-01
As part of the Flat-Plate Solar Array Project, 138 different module designs were subjected to qualification tests. Electrical measurements were subjected on well over a thousand modules representing more than 150 designs. From this experience, conclusions are drawn regarding results and problems, with discussion of the need for change or improvement. The qualification test sequence incuded application of environmental and electrical stresses to the module. With few exceptions, the tests have revealed defects necessitation of environmental and electrical stresses to the module. With few exceptions, the tests have revealed defects necessitating module design or process changes. However, the continued need for these tests may be questioned on the basis of technical and logistical factors. Technically, the current test sequence does not cover all design characteristics, does not include all field conditions and is not known to represent the desired 30-year module life. Logistically, the tests are time-consuming and costly, and there is a lack of, fully qualified independent test organizations. Alternatives to the current test program include simplification based on design specification and site environment, and/or the use of warranties or other commercial practices.
ERIC Educational Resources Information Center
Marchis, Iuliana
2009-01-01
The results of the Romanian pupils on international tests PISA and TIMSS in Mathematics are below the average. These poor results have many explications. In this article we compare the Mathematics problems given on these international tests with those given on national tests in Romania.
Testing problem-solving capacities: differences between individual testing and social group setting.
Krasheninnikova, Anastasia; Schneider, Jutta M
2014-09-01
Testing animals individually in problem-solving tasks limits distractions of the subjects during the test, so that they can fully concentrate on the problem. However, such individual performance may not indicate the problem-solving capacity that is commonly employed in the wild when individuals are faced with a novel problem in their social groups, where the presence of a conspecific influences an individual's behaviour. To assess the validity of data gathered from parrots when tested individually, we compared the performance on patterned-string tasks among parrots tested singly and parrots tested in social context. We tested two captive groups of orange-winged amazons (Amazona amazonica) with several patterned-string tasks. Despite the differences in the testing environment (singly vs. social context), parrots from both groups performed similarly. However, we found that the willingness to participate in the tasks was significantly higher for the individuals tested in social context. The study provides further evidence for the crucial influence of social context on individual's response to a challenging situation such as a problem-solving test. PMID:24668582
Reproducibility problems with the AMPLICOR PCR Chlamydia trachomatis test.
Peterson, E M; Darrow, V; Blanding, J; Aarnaes, S; de la Maza, L M
1997-01-01
In an attempt to use an expanded "gold standard" in an evaluation of an antigen detection test for Chlamydia trachomatis, the AMPLICOR (Roche Diagnostics Systems, Inc., Branchburg, N.J.) PCR Chlamydia trachomatis test and culture were used with 591 sets of cervical specimens. Of the 591 specimens assayed, 35 were retested due to either an equivocal result by the PCR (19 samples) or a discrepancy between the results of culture, PCR, and the antigen detection method. During the repeat testing of the samples with equivocal and discrepant results, all but one interpretation change was due to the PCR result. In addition, upon repeat testing the PCR assay value measured in optical density units varied widely for 13 of these specimens. These 13 specimens were then tested in triplicate by the manufacturer with primers to the chlamydia plasmid and in duplicate with primers to the major outer membrane protein. Only 3 of the 13 specimens gave the same interpretation with these five replicates. In summary, reproducibility problems with the AMPLICOR test should be considered before it is incorporated as part of routine testing or used as an expanded gold standard for chlamydia testing. PMID:9157161
NASA Astrophysics Data System (ADS)
Wang, Jin; Ma, Jianyong; Zhou, Changhe
2014-11-01
A 3×3 high divergent 2D-grating with period of 3.842μm at wavelength of 850nm under normal incidence is designed and fabricated in this paper. This high divergent 2D-grating is designed by the vector theory. The Rigorous Coupled Wave Analysis (RCWA) in association with the simulated annealing (SA) is adopted to calculate and optimize this 2D-grating.The properties of this grating are also investigated by the RCWA. The diffraction angles are more than 10 degrees in the whole wavelength band, which are bigger than the traditional 2D-grating. In addition, the small period of grating increases the difficulties of fabrication. So we fabricate the 2D-gratings by direct laser writing (DLW) instead of traditional manufacturing method. Then the method of ICP etching is used to obtain the high divergent 2D-grating.
Mason, W.E.
1983-03-01
A set of finite element codes for the solution of nonlinear, two-dimensional (TACO2D) and three-dimensional (TACO3D) heat transfer problems. Performs linear and nonlinear analyses of both transient and steady state heat transfer problems. Has the capability to handle time or temperature dependent material properties. Materials may be either isotropic or orthotropic. A variety of time and temperature dependent boundary conditions and loadings are available including temperature, flux, convection, radiation, and internal heat generation.
NASA Astrophysics Data System (ADS)
Slanger, T. G.; Cosby, P. C.; Huestis, D. L.
2003-04-01
N(^2D) is an important species in the nighttime ionosphere, as its reaction with O_2 is a principal source of NO. Its modeled concentration peaks near 200 km, at approximately 4 × 10^5 cm-3. Nightglow emission in the optically forbidden lines at 519.8 and 520.0 nm is quite weak, a consequence of the combination of an extremely long radiative lifetime, about 10^5 sec, and quenching by O-atoms, O_2, and N_2. The radiative lifetime is known only from theory, and various calculations lead to a range of possible values for the intensity ratio R = I(519.8)/I(520.0) of 1.5-2.5. On the observational side, Hernandez and Turtle [1969] determined a range of R = 1.3-1.9 in the nightglow, and Sivjee et al. [1981] reported a variable ratio in aurorae, between 1.2 and 1.6. From sky spectra obtained at the Keck II telescope on Mauna Kea, we have accumulated eighty-five 30-60 minute data sets, from March and October, 2000, and April, 2001, over 13 nights of astronomical observations. We find R to have a quite precise value of 1.760± 0.012 (2-σ). There is no difference between the three data sets in terms of the extracted ratio, which therefore seems to be independent of external conditions. At the same time, determination of the O(^1D - ^3P) doublet intensity ratio, I(630.0)/I(636.4), gives a value of 3.03 ± 0.01, the statistical expectation. G. Hernandez and J. P. Turtle, Planet. Space Sci. 17, 675, 1969. G. G. Sivjee, C. S. Deehr, and K. Henricksen, J. Geophys. Res. 86, 1581, 1981.
Test-state approach to the quantum search problem
Sehrawat, Arun; Nguyen, Le Huy; Englert, Berthold-Georg
2011-05-15
The search for 'a quantum needle in a quantum haystack' is a metaphor for the problem of finding out which one of a permissible set of unitary mappings - the oracles - is implemented by a given black box. Grover's algorithm solves this problem with quadratic speedup as compared with the analogous search for 'a classical needle in a classical haystack'. Since the outcome of Grover's algorithm is probabilistic - it gives the correct answer with high probability, not with certainty - the answer requires verification. For this purpose we introduce specific test states, one for each oracle. These test states can also be used to realize 'a classical search for the quantum needle' which is deterministic - it always gives a definite answer after a finite number of steps - and 3.41 times as fast as the purely classical search. Since the test-state search and Grover's algorithm look for the same quantum needle, the average number of oracle queries of the test-state search is the classical benchmark for Grover's algorithm.
Extended 2D generalized dilaton gravity theories
NASA Astrophysics Data System (ADS)
de Mello, R. O.
2008-09-01
We show that an anomaly-free description of matter in (1+1) dimensions requires a deformation of the 2D relativity principle, which introduces a non-trivial centre in the 2D Poincaré algebra. Then we work out the reduced phase space of the anomaly-free 2D relativistic particle, in order to show that it lives in a noncommutative 2D Minkowski space. Moreover, we build a Gaussian wave packet to show that a Planck length is well defined in two dimensions. In order to provide a gravitational interpretation for this noncommutativity, we propose to extend the usual 2D generalized dilaton gravity models by a specific Maxwell component, which guages the extra symmetry associated with the centre of the 2D Poincaré algebra. In addition, we show that this extension is a high energy correction to the unextended dilaton theories that can affect the topology of spacetime. Further, we couple a test particle to the general extended dilaton models with the purpose of showing that they predict a noncommutativity in curved spacetime, which is locally described by a Moyal star product in the low energy limit. We also conjecture a probable generalization of this result, which provides strong evidence that the noncommutativity is described by a certain star product which is not of the Moyal type at high energies. Finally, we prove that the extended dilaton theories can be formulated as Poisson Sigma models based on a nonlinear deformation of the extended Poincaré algebra.
Application of successive test feature classifier to dynamic recognition problems
NASA Astrophysics Data System (ADS)
Sakata, Yukinobu; Kaneko, Shun'ichi; Tanaka, Takayuki
2005-12-01
A novel successive learning algorithm is proposed for efficiently handling sequentially provided training data based on Test Feature Classifier (TFC), which is non-parametric and effective even for small data. We have proposed a novel classifier TFC utilizing prime test features (PTF) which is combination feature subsets for getting excellent performance. TFC has characteristics as follows: non-parametric learning, no mis-classification of training data. And then, in some real-world problems, the effectiveness of TFC is confirmed through way applications. However, TFC has a problem that it must be reconstructed even when any sub-set of data is changed. In the successive learning, after recognition of a set of unknown objects, they are fed into the classifier in order to obtain a modified classifier. We propose an efficient algorithm for reconstruction of PTFs, which is formalized in cases of addition and deletion of training data. In the verification experiment, using the successive learning algorithm, we can save about 70% on the total computational cost in comparison with a batch learning. We applied the proposed successive TFC to dynamic recognition problems from which the characteristic of training data changes with progress of time, and examine the characteristic by the fundamental experiments. Support Vector Machine (SVM) which is well established in algorithm and on practical application, was compared with the proposed successive TFC. And successive TFC indicated high performance compared with SVM.
NASA Astrophysics Data System (ADS)
Antinori, Samuele; Falchieri, Davide; Gabrielli, Alessandro; Gandolfi, Enzo
2004-09-01
CARLOSv3 is a third version of a chip that plays a significant role in the data acquisition chain of the A Large Ion Collider Experiment Inner Tracking System experiment. It has been designed and realized with a 0.25 μm CMOS 3-metal rad-hard digital library. The chip elaborates and compresses, by means of a bi-dimensional compressor, data belonging to a so-called event. The compressor looks for cross-shaped clusters within the whole data set coming from the silicon detector. To test the chip a specific PCB has been designed; it contains the connectors for probing the ASIC with a pattern generator and a logic state analyzer. The chip is inserted on the PCB using a ZIF socket. This allows to test the 35 packaged samples out of the total amount of bare chips we have from the foundry. The test phase has shown that 32 out of 35 chips under test work well. It is planned to redesign a new version of the chip by adding extra features and to submit the final version of CARLOS upon the final DAQ chain will be totally tested both in Bologna and at CERN.
ERIC Educational Resources Information Center
Congress of the U.S., Washington, DC. House Committee on Education and the Workforce.
H.R. 2846, a bill to prohibit spending Federal education funds on national testing without explicit and specific legislation was referred to the Committee on Education and the Workforce of the U.S. House of Representatives. The Committee, having reviewed the bill, reports favorably on it in this document, proposes some amendments, and recommends…
Review of measurement and testing problems. [of aircraft emissions
NASA Technical Reports Server (NTRS)
1976-01-01
Good instrumentation was required to obtain reliable and repeatable baseline data. Problems that were encountered in developing such a total system were: (1) accurate airflow measurement, (2) precise fuel flow measurement, and (3) the instrumentation used for pollutant measurement was susceptible to frequent malfunctions. Span gas quality had a significant effect on emissions test results. The Spindt method was used in the piston aircraft emissions program. The Spindt method provided a comparative computational procedure for fuel/air ratio based on measured emissions concentrations.
A class of self-similar hydrodynamics test problems
Ramsey, Scott D; Brown, Lowell S; Nelson, Eric M; Alme, Marv L
2010-12-08
We consider self-similar solutions to the gas dynamics equations. One such solution - a spherical geometry Gaussian density profile - has been analyzed in the existing literature, and a connection between it, a linear velocity profile, and a uniform specific internal energy profile has been identified. In this work, we assume the linear velocity profile to construct an entire class of self-similar sol utions in both cylindrical and spherical geometry, of which the Gaussian form is one possible member. After completing the derivation, we present some results in the context of a test problem for compressible flow codes.
Scaling in the 2D SU(3) × SU(3) spin model as a test of a new coding method for SU(3) matrices
NASA Astrophysics Data System (ADS)
Bunk, B.; Sommer, R.
1985-02-01
We present a Monte Carlo measurement of the magnetic susceptibility in the SU(3) × SU(3) spin model in two dimensions. Asymptotic scaling is verified on a 20 × 20 lattice. This laboratory is then used to test a new method for coding SU(3) variables in one (60 bit)- word of computer memory. In this approach, real numbers are truncated to fit into a 5-bit representation.
49 CFR 40.205 - How are drug test problems corrected?
Code of Federal Regulations, 2010 CFR
2010-10-01
... 49 Transportation 1 2010-10-01 2010-10-01 false How are drug test problems corrected? 40.205... WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Drug Tests § 40.205 How are drug test problems...), you must try to correct the problem promptly, if doing so is practicable. You may conduct...
49 CFR 40.205 - How are drug test problems corrected?
Code of Federal Regulations, 2011 CFR
2011-10-01
... 49 Transportation 1 2011-10-01 2011-10-01 false How are drug test problems corrected? 40.205... WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Drug Tests § 40.205 How are drug test problems...), you must try to correct the problem promptly, if doing so is practicable. You may conduct...
Ji, Yuan; Skierka, Jennifer M; Blommel, Joseph H; Moore, Brenda E; VanCuyk, Douglas L; Bruflat, Jamie K; Peterson, Lisa M; Veldhuizen, Tamra L; Fadra, Numrah; Peterson, Sandra E; Lagerstedt, Susan A; Train, Laura J; Baudhuin, Linnea M; Klee, Eric W; Ferber, Matthew J; Bielinski, Suzette J; Caraballo, Pedro J; Weinshilboum, Richard M; Black, John L
2016-05-01
Significant barriers, such as lack of professional guidelines, specialized training for interpretation of pharmacogenomics (PGx) data, and insufficient evidence to support clinical utility, prevent preemptive PGx testing from being widely clinically implemented. The current study, as a pilot project for the Right Drug, Right Dose, Right Time-Using Genomic Data to Individualize Treatment Protocol, was designed to evaluate the impact of preemptive PGx and to optimize the workflow in the clinic setting. We used an 84-gene next-generation sequencing panel that included SLCO1B1, CYP2C19, CYP2C9, and VKORC1 together with a custom-designed CYP2D6 testing cascade to genotype the 1013 subjects in laboratories approved by the Clinical Laboratory Improvement Act. Actionable PGx variants were placed in patient's electronic medical records where integrated clinical decision support rules alert providers when a relevant medication is ordered. The fraction of this cohort carrying actionable PGx variant(s) in individual genes ranged from 30% (SLCO1B1) to 79% (CYP2D6). When considering all five genes together, 99% of the subjects carried an actionable PGx variant(s) in at least one gene. Our study provides evidence in favor of preemptive PGx testing by identifying the risk of a variant being present in the population we studied. PMID:26947514
Phillips, Lawrence M; Hachamovitch, Rory; Berman, Daniel S; Iskandrian, Ami E; Min, James K; Picard, Michael H; Kwong, Raymond Y; Friedrich, Matthias G; Scherrer-Crosbie, Marielle; Hayes, Sean W; Sharir, Tali; Gosselin, Gilbert; Mazzanti, Marco; Senior, Roxy; Beanlands, Rob; Smanio, Paola; Goyal, Abhi; Al-Mallah, Mouaz; Reynolds, Harmony; Stone, Gregg W; Maron, David J; Shaw, Leslee J
2013-12-01
There is a preponderance of evidence that, in the setting of an acute coronary syndrome, an invasive approach using coronary revascularization has a morbidity and mortality benefit. However, recent stable ischemic heart disease (SIHD) randomized clinical trials testing whether the addition of coronary revascularization to guideline-directed medical therapy (GDMT) reduces death or major cardiovascular events have been negative. Based on the evidence from these trials, the primary role of GDMT as a front line medical management approach has been clearly defined in the recent SIHD clinical practice guideline; the role of prompt revascularization is less precisely defined. Based on data from observational studies, it has been hypothesized that there is a level of ischemia above which a revascularization strategy might result in benefit regarding cardiovascular events. However, eligibility for recent negative trials in SIHD has mandated at most minimal standards for ischemia. An ongoing randomized trial evaluating the effectiveness of randomization of patients to coronary angiography and revascularization as compared to no coronary angiography and GDMT in patients with moderate-severe ischemia will formally test this hypothesis. The current review will highlight the available evidence including a review of the published and ongoing SIHD trials. PMID:23963599
Phillips, Lawrence M.; Hachamovitch, Rory; Berman, Daniel S.; Iskandrian, Ami E.; Min, James K.; Picard, Michael H.; Kwong, Raymond Y.; Friedrich, Matthias G.; Scherrer-Crosbie, Marielle; Hayes, Sean W.; Sharir, Tali; Gosselin, Gilbert; Mazzanti, Marco; Senior, Roxy; Beanlands, Rob; Smanio, Paola; Goyal, Abhi; Al-Mallah, Mouaz; Reynolds, Harmony; Stone, Gregg W.; Maron, David J.; Shaw, Leslee J.
2014-01-01
There is a preponderance of evidence that, in the setting of an acute coronary syndrome, an invasive approach using coronary revascularization has a morbidity and mortality benefit. However, recent stable ischemic heart disease (SIHD) randomized clinical trials testing whether the addition of coronary revascularization to guideline-directed medical therapy (GDMT) reduces death or major cardiovascular events have been negative. Based on the evidence from these trials, the primary role of GDMT as a front line medical management approach has been clearly defined in the recent SIHD clinical practice guideline; the role of prompt revascularization is less precisely defined. Based on data from observational studies, it has been hypothesized that there is a level of ischemia above which a revascularization strategy might result in benefit regarding cardiovascular events. However, eligibility for recent negative trials in SIHD has mandated at most minimal standards for ischemia. An ongoing randomized trial evaluating the effectiveness of randomization of patients to coronary angiography and revascularization as compared to no coronary angiography and GDMT in patients with moderate-severe ischemia will formally test this hypothesis. The current review will highlight the available evidence including a review of the published and ongoing SIHD trials. PMID:23963599
Predicting non-square 2D dice probabilities
NASA Astrophysics Data System (ADS)
Pender, G. A. T.; Uhrin, M.
2014-07-01
The prediction of the final state probabilities of a general cuboid randomly thrown onto a surface is a problem that naturally arises in the minds of men and women familiar with regular cubic dice and the basic concepts of probability. Indeed, it was considered by Newton in 1664 (Newton 1967 The Mathematical Papers of Issac Newton vol I (Cambridge: Cambridge University Press) pp 60-1). In this paper we make progress on the 2D problem (which can be realized in 3D by considering a long cuboid, or alternatively a rectangular cross-sectioned dreidel). For the two-dimensional case we suggest that the ratio of the probabilities of landing on each of the two sides is given by \\frac{\\sqrt{{{k}^{2}}+{{l}^{2}}}-k}{\\sqrt{{{k}^{2}}+{{l}^{2}}}-l}\\frac{arctan \\frac{l}{k}}{arctan \\frac{k}{l}} where k and l are the lengths of the two sides. We test this theory both experimentally and computationally, and find good agreement between our theory, experimental and computational results. Our theory is known, from its derivation, to be an approximation for particularly bouncy or ‘grippy’ surfaces where the die rolls through many revolutions before settling. On real surfaces we would expect (and we observe) that the true probability ratio for a 2D die is a somewhat closer to unity than predicted by our theory. This problem may also have wider relevance in the testing of physics engines.
2004-08-01
AnisWave2D is a 2D finite-difference code for a simulating seismic wave propagation in fully anisotropic materials. The code is implemented to run in parallel over multiple processors and is fully portable. A mesh refinement algorithm has been utilized to allow the grid-spacing to be tailored to the velocity model, avoiding the over-sampling of high-velocity materials that usually occurs in fixed-grid schemes.
Leak testing of cryogenic components — problems and solutions
NASA Astrophysics Data System (ADS)
Srivastava, S. P.; Pandarkar, S. P.; Unni, T. G.; Sinha, A. K.; Mahajan, K.; Suthar, R. L.
2008-05-01
moderator pot was driving the MSLD out of range. Since it was very difficult to locate the leak by Tracer Probe Method, some other technique was ventured to solve the problem of leak location. Finally, it was possible to locate the leak by observing the change in Helium background reading of MSLD during masking/unmasking of the welded joints. This paper, in general describes the design and leak testing aspects of cryogenic components of Cold Neutron Source and in particular, the problems and solutions for leak testing of transfer lines and moderator pot.
Feautrier, D.; Smith, D.L.
1992-03-01
This report describes the development and testing of a deuterium gas target intended for use at a low-energy accelerator facility to produce neutrons for basic research and various nuclear applications. The principle source reaction is H-2(d,n)He-3. It produces a nearly mono-energetic group of neutrons. However, a lower-energy continuum neutron spectrum is produced by the H-2(d;n,p)H-2 reaction and also by deuterons which strike various components in the target assembly. The present target is designed to achieve the following objectives: (1) minimize unwanted background neutron production from the target assembly, (2) provide a relatively low level of residual long-term activity within the target components, (3) have the capacity to dissipate up to 150 watts of beam power with good target longevity, and (4) possess a relatively modest target mass in order to minimize neutron scattering from the target components. The basic physical principles that have to be considered in designing an accelerator target are discussed and the major engineering features of this particular target design are outlined. The results of initial performance tests on this target are documented and some conclusions concerning the viability of the target design are presented.
Ultrasonic 2D matrix PVDF transducer
NASA Astrophysics Data System (ADS)
Ptchelintsev, A.; Maev, R. Gr.
2000-05-01
During the past decade a substantial amount of work has been done in the area of ultrasonic imaging technology using 2D arrays. The main problems arising for the two-dimensional matrix transducers at megahertz frequencies are small size and huge count of the elements, high electrical impedance, low sensitivity, bad SNR and slower data acquisition rate. The major technological difficulty remains the high density of the interconnect. To solve these problems numerous approaches have been suggested. In the present work, a 24×24 elements (24 transmit+24 receive) matrix and a switching board were developed. The transducer consists of two 52 μm PVDF layers each representing a linear array of 24 elements placed one on the top of the other. Electrodes in these two layers are perpendicular and form the grid of 0.5×0.5 mm pitch. The layers are bonded together with the ground electrode being monolithic and located between the layers. The matrix is backed from the rear surface with an epoxy composition. During the emission, a linear element from the emitting layer generates a longitudinal wave pulse propagating inside the test object. Reflected pulses are picked-up by the receiving layer. During one transmit-receive cycle one transmit element and one receive element are selected by corresponding multiplexers. These crossed elements emulate a small element formed by their intersection. The present design presents the following advantages: minimizes number of active channels and density of the interconnect; reduces the electrical impedance of the element improving electrical matching; enables the transmit-receive mode; due to the efficient backing provides bandwidth and good time resolution; and, significantly reduces the electronics complexity. The matrix can not be used for the beam steering and focusing. Owing to this impossibility of focusing, the penetration depth is limited as well by the diffraction phenomena.
NASA Astrophysics Data System (ADS)
Jones, Alan G.; Afonso, Juan Carlos; Fullea, Javier; Salajegheh, Farshad
2014-02-01
Modeling the continental lithosphere's physical properties, especially its depth extent, must be done within a self-consistent petrological-geophysical framework; modeling using only one or two data types may easily lead to inconsistencies and erroneous interpretations. Using the LitMod approach for hypothesis testing and first-order modeling, we show how assumptions made about crustal information and the probable compositions of the lithospheric and sub-lithospheric mantle affect particular observables, particularly especially surface topographic elevation. The critical crustal parameter is density, leading to ca. 600 m error in topography for 50 kg m- 3 imprecision. The next key parameter is crustal thickness, and uncertainties in its definition lead to around ca. 4 km uncertainty in LAB for every 1 km of variation in Moho depth. Possible errors in the other assumed crustal parameters introduce a few kilometers of uncertainty in the depth to the LAB. We use Ireland as a natural laboratory to demonstrate the approach. From first-order arguments and given reasonable assumptions, a topographic elevation in the range of 50-100 m, which is the average across Ireland, requires that the lithosphere-asthenosphere boundary (LAB) beneath most of Ireland must lie in the range 90-115 km. A somewhat shallower (to 85 km) LAB is permitted, but the crust must be thinned (< 29 km) to compensate. The observations, especially topography, are inconsistent with suggestions, based on interpretation of S-to-P receiver functions, that the LAB thins from 85 km in southern Ireland to 55 km in central northern Ireland over a distance of < 150 km. Such a thin lithosphere would result in over 1000 m of uplift, and such rapid thinning by 30 km over less than 150 km would yield significant north-south variations in topographic elevation, Bouguer anomaly, and geoid height, none of which are observed. Even juxtaposing the most extreme probable depleted composition for the lithospheric mantle
Mechanical modeling of porous oxide fuel pellet A Test Problem
Nukala, Phani K; Barai, Pallab; Simunovic, Srdjan; Ott, Larry J
2009-10-01
A poro-elasto-plastic material model has been developed to capture the response of oxide fuels inside the nuclear reactors under operating conditions. Behavior of the oxide fuel and variation in void volume fraction under mechanical loading as predicted by the developed model has been reported in this article. The significant effect of void volume fraction on the overall stress distribution of the fuel pellet has also been described. An important oxide fuel issue that can have significant impact on the fuel performance is the mechanical response of oxide fuel pellet and clad system. Specifically, modeling the thermo-mechanical response of the fuel pellet in terms of its thermal expansion, mechanical deformation, swelling due to void formation and evolution, and the eventual contact of the fuel with the clad is of significant interest in understanding the fuel-clad mechanical interaction (FCMI). These phenomena are nonlinear and coupled since reduction in the fuel-clad gap affects thermal conductivity of the gap, which in turn affects temperature distribution within the fuel and the material properties of the fuel. Consequently, in order to accurately capture fuel-clad gap closure, we need to account for fuel swelling due to generation, retention, and evolution of fission gas in addition to the usual thermal expansion and mechanical deformation. Both fuel chemistry and microstructure also have a significant effect on the nucleation and growth of fission gas bubbles. Fuel-clad gap closure leading to eventual contact of the fuel with the clad introduces significant stresses in the clad, which makes thermo-mechanical response of the clad even more relevant. The overall aim of this test problem is to incorporate the above features in order to accurately capture fuel-clad mechanical interaction. Because of the complex nature of the problem, a series of test problems with increasing multi-physics coupling features, modeling accuracy, and complexity are defined with the
CYP2D7 Sequence Variation Interferes with TaqMan CYP2D6*15 and *35 Genotyping
Riffel, Amanda K.; Dehghani, Mehdi; Hartshorne, Toinette; Floyd, Kristen C.; Leeder, J. Steven; Rosenblatt, Kevin P.; Gaedigk, Andrea
2016-01-01
TaqMan™ genotyping assays are widely used to genotype CYP2D6, which encodes a major drug metabolizing enzyme. Assay design for CYP2D6 can be challenging owing to the presence of two pseudogenes, CYP2D7 and CYP2D8, structural and copy number variation and numerous single nucleotide polymorphisms (SNPs) some of which reflect the wild-type sequence of the CYP2D7 pseudogene. The aim of this study was to identify the mechanism causing false-positive CYP2D6*15 calls and remediate those by redesigning and validating alternative TaqMan genotype assays. Among 13,866 DNA samples genotyped by the CompanionDx® lab on the OpenArray platform, 70 samples were identified as heterozygotes for 137Tins, the key SNP of CYP2D6*15. However, only 15 samples were confirmed when tested with the Luminex xTAG CYP2D6 Kit and sequencing of CYP2D6-specific long range (XL)-PCR products. Genotype and gene resequencing of CYP2D6 and CYP2D7-specific XL-PCR products revealed a CC>GT dinucleotide SNP in exon 1 of CYP2D7 that reverts the sequence to CYP2D6 and allows a TaqMan assay PCR primer to bind. Because CYP2D7 also carries a Tins, a false-positive mutation signal is generated. This CYP2D7 SNP was also responsible for generating false-positive signals for rs769258 (CYP2D6*35) which is also located in exon 1. Although alternative CYP2D6*15 and *35 assays resolved the issue, we discovered a novel CYP2D6*15 subvariant in one sample that carries additional SNPs preventing detection with the alternate assay. The frequency of CYP2D6*15 was 0.1% in this ethnically diverse U.S. population sample. In addition, we also discovered linkage between the CYP2D7 CC>GT dinucleotide SNP and the 77G>A (rs28371696) SNP of CYP2D6*43. The frequency of this tentatively functional allele was 0.2%. Taken together, these findings emphasize that regardless of how careful genotyping assays are designed and evaluated before being commercially marketed, rare or unknown SNPs underneath primer and/or probe regions can impact
CYP2D7 Sequence Variation Interferes with TaqMan CYP2D6 (*) 15 and (*) 35 Genotyping.
Riffel, Amanda K; Dehghani, Mehdi; Hartshorne, Toinette; Floyd, Kristen C; Leeder, J Steven; Rosenblatt, Kevin P; Gaedigk, Andrea
2015-01-01
TaqMan™ genotyping assays are widely used to genotype CYP2D6, which encodes a major drug metabolizing enzyme. Assay design for CYP2D6 can be challenging owing to the presence of two pseudogenes, CYP2D7 and CYP2D8, structural and copy number variation and numerous single nucleotide polymorphisms (SNPs) some of which reflect the wild-type sequence of the CYP2D7 pseudogene. The aim of this study was to identify the mechanism causing false-positive CYP2D6 (*) 15 calls and remediate those by redesigning and validating alternative TaqMan genotype assays. Among 13,866 DNA samples genotyped by the CompanionDx® lab on the OpenArray platform, 70 samples were identified as heterozygotes for 137Tins, the key SNP of CYP2D6 (*) 15. However, only 15 samples were confirmed when tested with the Luminex xTAG CYP2D6 Kit and sequencing of CYP2D6-specific long range (XL)-PCR products. Genotype and gene resequencing of CYP2D6 and CYP2D7-specific XL-PCR products revealed a CC>GT dinucleotide SNP in exon 1 of CYP2D7 that reverts the sequence to CYP2D6 and allows a TaqMan assay PCR primer to bind. Because CYP2D7 also carries a Tins, a false-positive mutation signal is generated. This CYP2D7 SNP was also responsible for generating false-positive signals for rs769258 (CYP2D6 (*) 35) which is also located in exon 1. Although alternative CYP2D6 (*) 15 and (*) 35 assays resolved the issue, we discovered a novel CYP2D6 (*) 15 subvariant in one sample that carries additional SNPs preventing detection with the alternate assay. The frequency of CYP2D6 (*) 15 was 0.1% in this ethnically diverse U.S. population sample. In addition, we also discovered linkage between the CYP2D7 CC>GT dinucleotide SNP and the 77G>A (rs28371696) SNP of CYP2D6 (*) 43. The frequency of this tentatively functional allele was 0.2%. Taken together, these findings emphasize that regardless of how careful genotyping assays are designed and evaluated before being commercially marketed, rare or unknown SNPs underneath primer
49 CFR 40.271 - How are alcohol testing problems corrected?
Code of Federal Regulations, 2010 CFR
2010-10-01
... 49 Transportation 1 2010-10-01 2010-10-01 false How are alcohol testing problems corrected? 40.271... WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Alcohol Testing § 40.271 How are alcohol testing problems corrected? (a) As a BAT or STT, you have the responsibility of trying to complete successfully...
49 CFR 40.267 - What problems always cause an alcohol test to be cancelled?
Code of Federal Regulations, 2014 CFR
2014-10-01
... 49 Transportation 1 2014-10-01 2014-10-01 false What problems always cause an alcohol test to be cancelled? 40.267 Section 40.267 Transportation Office of the Secretary of Transportation PROCEDURES FOR TRANSPORTATION WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Alcohol Testing § 40.267 What problems always cause an alcohol test to...
49 CFR 40.271 - How are alcohol testing problems corrected?
Code of Federal Regulations, 2012 CFR
2012-10-01
... 49 Transportation 1 2012-10-01 2012-10-01 false How are alcohol testing problems corrected? 40.271... WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Alcohol Testing § 40.271 How are alcohol testing... alcohol test for each employee. (1) If, during or shortly after the testing process, you become aware...
49 CFR 40.271 - How are alcohol testing problems corrected?
Code of Federal Regulations, 2013 CFR
2013-10-01
... 49 Transportation 1 2013-10-01 2013-10-01 false How are alcohol testing problems corrected? 40.271... WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Alcohol Testing § 40.271 How are alcohol testing... alcohol test for each employee. (1) If, during or shortly after the testing process, you become aware...
49 CFR 40.271 - How are alcohol testing problems corrected?
Code of Federal Regulations, 2011 CFR
2011-10-01
... 49 Transportation 1 2011-10-01 2011-10-01 false How are alcohol testing problems corrected? 40.271... WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Alcohol Testing § 40.271 How are alcohol testing... alcohol test for each employee. (1) If, during or shortly after the testing process, you become aware...
Testing problem solving in turkey vultures (Cathartes aura) using the string-pulling test.
Ellison, Anne Margaret; Watson, Jane; Demers, Eric
2015-01-01
To examine problem solving in turkey vultures (Cathartes aura), six captive vultures were presented with a string-pulling task, which involved drawing a string up to access food. This test has been used to assess cognition in many bird species. A small piece of meat suspended by a string was attached to a perch. Two birds solved the problem without apparent trial-and-error learning; a third bird solved the problem after observing a successful bird, suggesting that this individual learned from the other vulture. The remaining birds failed to complete the task. The successful birds significantly reduced the time needed to solve the task from early trials compared to late trials, suggesting that they had learned to solve the problem and improved their technique. The successful vultures solved the problem in a novel way: they pulled the string through their beak with their tongue, and may have gathered the string in their crop until the food was in reach. In contrast, ravens, parrots and finches use a stepwise process; they pull the string up, tuck it under foot, and reach down to pull up another length. As scavengers, turkey vultures use their beak for tearing and ripping at carcasses, but possess large, flat, webbed feet that are ill-suited to pulling or grasping. The ability to solve this problem and the novel approach used by the turkey vultures in this study may be a result of the unique evolutionary pressures imposed on this scavenging species. PMID:25015133
Chimpanzee Problem-Solving: A Test for Comprehension.
ERIC Educational Resources Information Center
Premack, David; Woodruff, Guy
1978-01-01
Investigates a chimpanzee's capacity to recognize representations of problems and solutions, as well as its ability to perceive the relationship between each type of problem and its appropriate solutions using televised programs and photographic solutions. (HM)
NASA Astrophysics Data System (ADS)
Mayor, Louise
2016-05-01
Graphene might be the most famous example, but there are other 2D materials and compounds too. Louise Mayor explains how these atomically thin sheets can be layered together to create flexible “van der Waals heterostructures”, which could lead to a range of novel applications.
Material behavior and materials problems in TFTR (Tokamak Fusion Test Reactor)
Dylla, H.F.; Ulrickson, M.A.; Owens, D.K.; Heifetz, D.B.; Mills, B.E.; Pontau, A.E.; Wampler, W.R.; Doyle, B.L.; Lee, S.R.; Watson, R.D.; Croessmann, C.D.
1988-05-01
This paper reviews the experience with first-wall materials over a 20-month period of operation spanning 1985--1987. Experience with the axisymmetric inner wall limiter, constructed of graphite tiles, will be described including the necessary conditioning procedures needed for impurity and particle control of high power ({le}20 MW) neutral injection experiments. The thermal effects in disruptions have been quantified and no significant damage to the bumper limiter has occurred as a result of disruptions. Carbon and metal impurity redeposition effects have been quantified through surface analysis of wall samples. Estimates of the tritium retention in the graphite limiter tiles and redeposited carbon films have been made based on analysis of deuterium retention in removed graphite tiles and wall samples. New limiter structures have been designed using a 2D carbon/carbon (C/C) composite material for RF antenna protection. Laboratory tests of the important thermal, mechanical and vacuum properties of C/C materials will be described. Finally, the last series of experiments in TFTR with in-situ Zr/Al surface pumps will be described. Problems with Ar/Al embrittlement have led to the removal of the getter material from the in-torus environment. 53 refs., 8 figs., 3 tabs.
Report of the 1988 2-D Intercomparison Workshop, chapter 3
NASA Technical Reports Server (NTRS)
Jackman, Charles H.; Brasseur, Guy; Soloman, Susan; Guthrie, Paul D.; Garcia, Rolando; Yung, Yuk L.; Gray, Lesley J.; Tung, K. K.; Ko, Malcolm K. W.; Isaken, Ivar
1989-01-01
Several factors contribute to the errors encountered. With the exception of the line-by-line model, all of the models employ simplifying assumptions that place fundamental limits on their accuracy and range of validity. For example, all 2-D modeling groups use the diffusivity factor approximation. This approximation produces little error in tropospheric H2O and CO2 cooling rates, but can produce significant errors in CO2 and O3 cooling rates at the stratopause. All models suffer from fundamental uncertainties in shapes and strengths of spectral lines. Thermal flux algorithms being used in 2-D tracer tranport models produce cooling rates that differ by as much as 40 percent for the same input model atmosphere. Disagreements of this magnitude are important since the thermal cooling rates must be subtracted from the almost-equal solar heating rates to derive the net radiative heating rates and the 2-D model diabatic circulation. For much of the annual cycle, the net radiative heating rates are comparable in magnitude to the cooling rate differences described. Many of the models underestimate the cooling rates in the middle and lower stratosphere. The consequences of these errors for the net heating rates and the diabatic circulation will depend on their meridional structure, which was not tested here. Other models underestimate the cooling near 1 mbar. Suchs errors pose potential problems for future interactive ozone assessment studies, since they could produce artificially-high temperatures and increased O3 destruction at these levels. These concerns suggest that a great deal of work is needed to improve the performance of thermal cooling rate algorithms used in the 2-D tracer transport models.
49 CFR 40.199 - What problems always cause a drug test to be cancelled?
Code of Federal Regulations, 2010 CFR
2010-10-01
... 49 Transportation 1 2010-10-01 2010-10-01 false What problems always cause a drug test to be cancelled? 40.199 Section 40.199 Transportation Office of the Secretary of Transportation PROCEDURES FOR TRANSPORTATION WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Drug Tests § 40.199 What problems...
2001-01-31
This software reduces the data from two-dimensional kSA MOS program, k-Space Associates, Ann Arbor, MI. Initial MOS data is recorded without headers in 38 columns, with one row of data per acquisition per lase beam tracked. The final MOSS 2d data file is reduced, graphed, and saved in a tab-delimited column format with headers that can be plotted in any graphing software.
Tomosynthesis imaging with 2D scanning trajectories
NASA Astrophysics Data System (ADS)
Khare, Kedar; Claus, Bernhard E. H.; Eberhard, Jeffrey W.
2011-03-01
Tomosynthesis imaging in chest radiography provides volumetric information with the potential for improved diagnostic value when compared to the standard AP or LAT projections. In this paper we explore the image quality benefits of 2D scanning trajectories when coupled with advanced image reconstruction approaches. It is intuitively clear that 2D trajectories provide projection data that is more complete in terms of Radon space filling, when compared with conventional tomosynthesis using a linearly scanned source. Incorporating this additional information for obtaining improved image quality is, however, not a straightforward problem. The typical tomosynthesis reconstruction algorithms are based on direct inversion methods e.g. Filtered Backprojection (FBP) or iterative algorithms that are variants of the Algebraic Reconstruction Technique (ART). The FBP approach is fast and provides high frequency details in the image but at the same time introduces streaking artifacts degrading the image quality. The iterative methods can reduce the image artifacts by using image priors but suffer from a slow convergence rate, thereby producing images lacking high frequency details. In this paper we propose using a fast converging optimal gradient iterative scheme that has advantages of both the FBP and iterative methods in that it produces images with high frequency details while reducing the image artifacts. We show that using favorable 2D scanning trajectories along with the proposed reconstruction method has the advantage of providing improved depth information for structures such as the spine and potentially producing images with more isotropic resolution.
MAGNUM-2D computer code: user's guide
England, R.L.; Kline, N.W.; Ekblad, K.J.; Baca, R.G.
1985-01-01
Information relevant to the general use of the MAGNUM-2D computer code is presented. This computer code was developed for the purpose of modeling (i.e., simulating) the thermal and hydraulic conditions in the vicinity of a waste package emplaced in a deep geologic repository. The MAGNUM-2D computer computes (1) the temperature field surrounding the waste package as a function of the heat generation rate of the nuclear waste and thermal properties of the basalt and (2) the hydraulic head distribution and associated groundwater flow fields as a function of the temperature gradients and hydraulic properties of the basalt. MAGNUM-2D is a two-dimensional numerical model for transient or steady-state analysis of coupled heat transfer and groundwater flow in a fractured porous medium. The governing equations consist of a set of coupled, quasi-linear partial differential equations that are solved using a Galerkin finite-element technique. A Newton-Raphson algorithm is embedded in the Galerkin functional to formulate the problem in terms of the incremental changes in the dependent variables. Both triangular and quadrilateral finite elements are used to represent the continuum portions of the spatial domain. Line elements may be used to represent discrete conduits. 18 refs., 4 figs., 1 tab.
Interparticle Attraction in 2D Complex Plasmas
NASA Astrophysics Data System (ADS)
Kompaneets, Roman; Morfill, Gregor E.; Ivlev, Alexei V.
2016-03-01
Complex (dusty) plasmas allow experimental studies of various physical processes occurring in classical liquids and solids by directly observing individual microparticles. A major problem is that the interaction between microparticles is generally not molecularlike. In this Letter, we propose how to achieve a molecularlike interaction potential in laboratory 2D complex plasmas. We argue that this principal aim can be achieved by using relatively small microparticles and properly adjusting discharge parameters. If experimentally confirmed, this will make it possible to employ complex plasmas as a model system with an interaction potential resembling that of conventional liquids.
A scalable 2-D parallel sparse solver
Kothari, S.C.; Mitra, S.
1995-12-01
Scalability beyond a small number of processors, typically 32 or less, is known to be a problem for existing parallel general sparse (PGS) direct solvers. This paper presents a parallel general sparse PGS direct solver for general sparse linear systems on distributed memory machines. The algorithm is based on the well-known sequential sparse algorithm Y12M. To achieve efficient parallelization, a 2-D scattered decomposition of the sparse matrix is used. The proposed algorithm is more scalable than existing parallel sparse direct solvers. Its scalability is evaluated on a 256 processor nCUBE2s machine using Boeing/Harwell benchmark matrices.
VAM2D: Variably saturated analysis model in two dimensions
Huyakorn, P.S.; Kool, J.B.; Wu, Y.S. )
1991-10-01
This report documents a two-dimensional finite element model, VAM2D, developed to simulate water flow and solute transport in variably saturated porous media. Both flow and transport simulation can be handled concurrently or sequentially. The formulation of the governing equations and the numerical procedures used in the code are presented. The flow equation is approximated using the Galerkin finite element method. Nonlinear soil moisture characteristics and atmospheric boundary conditions (e.g., infiltration, evaporation and seepage face), are treated using Picard and Newton-Raphson iterations. Hysteresis effects and anisotropy in the unsaturated hydraulic conductivity can be taken into account if needed. The contaminant transport simulation can account for advection, hydrodynamic dispersion, linear equilibrium sorption, and first-order degradation. Transport of a single component or a multi-component decay chain can be handled. The transport equation is approximated using an upstream weighted residual method. Several test problems are presented to verify the code and demonstrate its utility. These problems range from simple one-dimensional to complex two-dimensional and axisymmetric problems. This document has been produced as a user's manual. It contains detailed information on the code structure along with instructions for input data preparation and sample input and printed output for selected test problems. Also included are instructions for job set up and restarting procedures. 44 refs., 54 figs., 24 tabs.
Nanoimprint lithography: 2D or not 2D? A review
NASA Astrophysics Data System (ADS)
Schift, Helmut
2015-11-01
Nanoimprint lithography (NIL) is more than a planar high-end technology for the patterning of wafer-like substrates. It is essentially a 3D process, because it replicates various stamp topographies by 3D displacement of material and takes advantage of the bending of stamps while the mold cavities are filled. But at the same time, it keeps all assets of a 2D technique being able to pattern thin masking layers like in photon- and electron-based traditional lithography. This review reports about 20 years of development of replication techniques at Paul Scherrer Institut, with a focus on 3D aspects of molding, which enable NIL to stay 2D, but at the same time enable 3D applications which are "more than Moore." As an example, the manufacturing of a demonstrator for backlighting applications based on thermally activated selective topography equilibration will be presented. This technique allows generating almost arbitrary sloped, convex and concave profiles in the same polymer film with dimensions in micro- and nanometer scale.
Observed-Score Equating as a Test Assembly Problem.
ERIC Educational Resources Information Center
van der Linden, Wim J.; Luecht, Richard M.
1998-01-01
Derives a set of linear conditions of item-response functions that guarantees identical observed-score distributions on two test forms. The conditions can be added as constraints to a linear programming model for test assembly. An example illustrates the use of the model for an item pool from the Law School Admissions Test (LSAT). (SLD)
Numerical Evaluation of 2D Ground States
NASA Astrophysics Data System (ADS)
Kolkovska, Natalia
2016-02-01
A ground state is defined as the positive radial solution of the multidimensional nonlinear problem
Some Reliability Problems in a Criterion-Referenced Test.
ERIC Educational Resources Information Center
Roudabush, Glenn E.; Green, Donald Ross
This paper describes the development of a criterion-referenced test. The Prescriptive Mathematics Inventory (PMI) was developed to measure 400 stated behavioral objectives. The test consists of three overlapping levels with the objectives chosen to cover 90 to 95 per cent of the mathematics curriculum nominally taught in grades 4 through 8. Each…
An empirical coverage test for the g-sample problem
Orlowski, L.A.; Grundy, W.D.; Mielke, P.W., Jr.
1991-01-01
A nonparametric g-sample empirical coverage test has recently been developed for univariate continuous data. It is based upon the empirical coverages which are spacings of multiple random samples. The test is capable of detecting any distributional differences which may exist among the parent populations, without additional assumptions beyond randomness and continuity. The test can be effective with the limited and/or unequal sample sizes most often encountered in geologic studies. A computer program for implementing this procedure, G-SECT 1, is available. ?? 1991 International Association for Mathematical Geology.
Dominant 2D magnetic turbulence in the solar wind
NASA Technical Reports Server (NTRS)
Bieber, John W.; Wanner, Wolfgang; Matthaeus, William H.
1995-01-01
There have been recent suggestions that solar wind magnetic turbulence may be a composite of slab geometry (wavevector aligned with the mean magnetic field) and 2D geometry (wavevectors perpendicular to the mean field). We report results of two new tests of this hypothesis using Helios measurements of inertial ranged magnetic spectra in the solar wind. The first test is based upon a characteristic difference between perpendicular and parallel reduced power spectra which is expected for the 2D component but not for the slab component. The second test examines the dependence of power spectrum density upon the magnetic field angle (i.e., the angle between the mean magnetic field and the radial direction), a relationship which is expected to be in opposite directions for the slab and 2D components. Both tests support the presence of a dominant (approximately 85 percent by energy) 2D component in solar wind magnetic turbulence.
Dominant 2D magnetic turbulence in the solar wind
Bieber, John W.; Wanner, Wolfgang; Matthaeus, William H.
1996-07-20
There have been recent suggestions that solar wind magnetic turbulence may be a composite of slab geometry (wavevectors aligned with the mean magnetic field) and 2D geometry (wavevectors perpendicular to the mean field). We report results of two new tests of this hypothesis using Helios measurements of mid-inertial range magnetic spectra in the solar wind. The first test is based upon a characteristic difference between reduced magnetic power spectra in the two different directions perpendicular to the mean field. Such a difference is expected for 2D geometry but not for slab geometry. The second test examines the dependence of power spectrum density upon the magnetic field angle (i.e., the angle between the mean magnetic field and the radial direction), a relationship which is expected to be in opposite directions for the slab and 2D components. Both tests support the presence of a dominant ({approx}85% by energy) 2D component in solar wind magnetic turbulence.
ERIC Educational Resources Information Center
van Gog, Tamara; Kester, Liesbeth; Dirkx, Kim; Hoogerheide, Vincent; Boerboom, Joris; Verkoeijen, Peter P. J. L.
2015-01-01
Four experiments investigated whether the testing effect also applies to the acquisition of problem-solving skills from worked examples. Experiment 1 (n?=?120) showed no beneficial effects of testing consisting of "isomorphic" problem solving or "example recall" on final test performance, which consisted of isomorphic problem…
The 2D large deformation analysis using Daubechies wavelet
NASA Astrophysics Data System (ADS)
Liu, Yanan; Qin, Fei; Liu, Yinghua; Cen, Zhangzhi
2010-01-01
In this paper, Daubechies (DB) wavelet is used for solution of 2D large deformation problems. Because the DB wavelet scaling functions are directly used as basis function, no meshes are needed in function approximation. Using the DB wavelet, the solution formulations based on total Lagrangian approach for two-dimensional large deformation problems are established. Due to the lack of Kroneker delta properties in wavelet scaling functions, Lagrange multipliers are used for imposition of boundary condition. Numerical examples of 2D large deformation problems illustrate that this method is effective and stable.
Usability Testing Finds Problems for Novice Users of Pediatric Portals
Britto, Maria T.; Jimison, Holly B.; Munafo, Jennifer Knopf; Wissman, Jennifer; Rogers, Michelle L.; Hersh, William
2009-01-01
Objective Patient portals may improve pediatric chronic disease outcomes, but few have been rigorously evaluated for usability by parents. Using scenario-based testing with think-aloud protocols, we evaluated the usability of portals for parents of children with cystic fibrosis, diabetes or arthritis. Design Sixteen parents used a prototype and test data to complete 14 tasks followed by a validated satisfaction questionnaire. Three iterations of the prototype were used. Measurements During the usability testing, we measured the time it took participants to complete or give up on each task. Sessions were videotaped and content-analyzed for common themes. Following testing, participants completed the Computer Usability Satisfaction Questionnaire which measured their opinions on the efficiency of the system, its ease of use, and the likability of the system interface. A 7-point Likert scale was used, with seven indicating the highest possible satisfaction. Results Mean task completion times ranged from 73 (± 61) seconds to locate a document to 431 (± 286) seconds to graph laboratory results. Tasks such as graphing, location of data, requesting access, and data interpretation were challenging. Satisfaction was greatest for interface pleasantness (5.9 ± 0.7) and likeability (5.8 ± 0.6) and lowest for error messages (2.3 ± 1.2) and clarity of information (4.2 ± 1.4). Overall mean satisfaction scores improved between iteration one and three. Conclusions Despite parental involvement and prior heuristic testing, scenario-based testing demonstrated difficulties in navigation, medical language complexity, error recovery, and provider-based organizational schema. While such usability testing can be expensive, the current study demonstrates that it can assist in making healthcare system interfaces for laypersons more user-friendly and potentially more functional for patients and their families. PMID:19567793
Crash test for the Copenhagen problem with oblateness
NASA Astrophysics Data System (ADS)
Zotos, Euaggelos E.
2015-05-01
The case of the planar circular restricted three-body problem where one of the two primaries is an oblate spheroid is investigated. We conduct a thorough numerical analysis on the phase space mixing by classifying initial conditions of orbits and distinguishing between three types of motion: (i) bounded, (ii) escape and (iii) collisional. The presented outcomes reveal the high complexity of this dynamical system. Furthermore, our numerical analysis shows a strong dependence of the properties of the considered escape basins with the total orbital energy, with a remarkable presence of fractal basin boundaries along all the escape regimes. Interpreting the collisional motion as leaking in the phase space we related our results to both chaotic scattering and the theory of leaking Hamiltonian systems. We also determined the escape and collisional basins and computed the corresponding escape/crash times. The highly fractal basin boundaries observed are related with high sensitivity to initial conditions thus implying an uncertainty between escape solutions which evolve to different regions of the phase space. We hope our contribution to be useful for a further understanding of the escape and crash mechanism of orbits in this version of the restricted three-body problem.
Problem-Solving Test: Real-Time Polymerase Chain Reaction
ERIC Educational Resources Information Center
Szeberenyi, Jozsef
2009-01-01
Terms to be familiar with before you start to solve the test: polymerase chain reaction, DNA amplification, electrophoresis, breast cancer, "HER2" gene, genomic DNA, "in vitro" DNA synthesis, template, primer, Taq polymerase, 5[prime][right arrow]3[prime] elongation activity, 5[prime][right arrow]3[prime] exonuclease activity, deoxyribonucleoside…
Problem-Solving Test: Submitochondrial Localization of Proteins
ERIC Educational Resources Information Center
Szeberenyi, Jozsef
2011-01-01
Mitochondria are surrounded by two membranes (outer and inner mitochondrial membrane) that separate two mitochondrial compartments (intermembrane space and matrix). Hundreds of proteins are distributed among these submitochondrial components. A simple biochemical/immunological procedure is described in this test to determine the localization of…
Problem-Solving Test: The Mechanism of Protein Synthesis
ERIC Educational Resources Information Center
Szeberenyi, Jozsef
2009-01-01
Terms to be familiar with before you start to solve the test: protein synthesis, ribosomes, amino acids, peptides, peptide bond, polypeptide chain, N- and C-terminus, hemoglobin, [alpha]- and [beta]-globin chains, radioactive labeling, [[to the third power]H] and [[to the fourteenth power]C]leucine, cytosol, differential centrifugation, density…
Language Testing in the Military: Problems, Politics and Progress
ERIC Educational Resources Information Center
Green, Rita; Wall, Dianne
2005-01-01
There appears to be little literature available -- either descriptive or research-related -- on language testing in the military. This form of specific purposes assessment affects both military personnel and civilians working within the military structure in terms of posting, promotion and remuneration, and it could be argued that it has serious…
[Problems of lung function testing in the laboratory].
Tojo, Naoko
2006-08-01
Spirometry is indispensable for the screening test of general respiratory function, and measurements of lung volume and diffusing capacity play an important role in the assessment of disease severity, functional disability, disease activity and response to treatment. Pulmonary function testing requires cooperation between the subjects and the examiner, and the results obtained depend on technical as well as personal factors. In order to diminish the variability of results and improve measurement accuracy, the Japan Respiratory Society published the first guidelines on the standardization of spirometry and diffusing capacity for both technical and clinical staff in 2004. It is therefore essential to distribute the guidelines to both laboratory personnel and general physicians. Furthermore, training workshops are mandatory to improve their understanding of the basics of lung function testing. Recently, there has been increasing interest in noninvasive methods of lung function testing without requiring the patient's cooperation during spontaneous breathing. Three alternative techniques, i.e. the negative expiratory pressure (NEP) method to detect expiratory flow limitation, impulse oscillation system (IOS) to measure respiratory system resistance (Rrs) and reactance (Xrs), and interruption resistance (Rint) to measure respiratory resistance have been introduced. Further study is required to determine the advantage of these methods. PMID:16989403
Common Problems of Mobile Applications for Foreign Language Testing
ERIC Educational Resources Information Center
Garcia Laborda, Jesus; Magal-Royo, Teresa; Lopez, Jose Luis Gimenez
2011-01-01
As the use of mobile learning educational applications has become more common anywhere in the world, new concerns have appeared in the classroom, human interaction in software engineering and ergonomics. new tests of foreign languages for a number of purposes have become more and more common recently. However, studies interrelating language tests…
Differential Validity: A Problem with Tests or Criteria?
ERIC Educational Resources Information Center
Hollmann, Thomas D.
The evidence used in condemning a test as racially biased is usually a validity coefficient for one racial group that is significantly different from that of another racial group. However, both variables in the calculation of a validity coefficient should be examined to determine where the bias lies. A study was conducted to investigate the…
Problem-Solving Test: Expression Cloning of the Erythropoietin Receptor
ERIC Educational Resources Information Center
Szeberenyi, Jozsef
2008-01-01
Terms to be familiar with before you start to solve the test: cytokines, cytokine receptors, cDNA library, cDNA synthesis, poly(A)[superscript +] RNA, primer, template, reverse transcriptase, restriction endonucleases, cohesive ends, expression vector, promoter, Shine-Dalgarno sequence, poly(A) signal, DNA helicase, DNA ligase, topoisomerases,…
A new inversion method for (T2, D) 2D NMR logging and fluid typing
NASA Astrophysics Data System (ADS)
Tan, Maojin; Zou, Youlong; Zhou, Cancan
2013-02-01
One-dimensional nuclear magnetic resonance (1D NMR) logging technology has some significant limitations in fluid typing. However, not only can two-dimensional nuclear magnetic resonance (2D NMR) provide some accurate porosity parameters, but it can also identify fluids more accurately than 1D NMR. In this paper, based on the relaxation mechanism of (T2, D) 2D NMR in a gradient magnetic field, a hybrid inversion method that combines least-squares-based QR decomposition (LSQR) and truncated singular value decomposition (TSVD) is examined in the 2D NMR inversion of various fluid models. The forward modeling and inversion tests are performed in detail with different acquisition parameters, such as magnetic field gradients (G) and echo spacing (TE) groups. The simulated results are discussed and described in detail, the influence of the above-mentioned observation parameters on the inversion accuracy is investigated and analyzed, and the observation parameters in multi-TE activation are optimized. Furthermore, the hybrid inversion can be applied to quantitatively determine the fluid saturation. To study the effects of noise level on the hybrid method and inversion results, the numerical simulation experiments are performed using different signal-to-noise-ratios (SNRs), and the effect of different SNRs on fluid typing using three fluid models are discussed and analyzed in detail.
Crash test for the restricted three-body problem.
Nagler, Jan
2005-02-01
The restricted three-body problem serves to investigate the chaotic behavior of a small body under the gravitational influence of two heavy primary bodies. We analyze numerically the phase space mixing of bounded motion, escape, and crash in this simple model of (chaotic) celestial mechanics. The presented extensive numerical analysis reveals a high degree of complexity. We extend the recently presented findings for the Copenhagen case of equal main masses to the general case of different primary body masses. Collisions of the small body onto the primaries are comparatively frequent, and their probability displays a scale-free dependence on the size of the primaries as shown for the Copenhagen case. Interpreting the crash as leaking in phase space the results are related to both chaotic scattering and the theory of leaking Hamiltonian systems. PMID:15783407
Development and Implementation of Radiation-Hydrodynamics Verification Test Problems
Marcath, Matthew J.; Wang, Matthew Y.; Ramsey, Scott D.
2012-08-22
Analytic solutions to the radiation-hydrodynamic equations are useful for verifying any large-scale numerical simulation software that solves the same set of equations. The one-dimensional, spherically symmetric Coggeshall No.9 and No.11 analytic solutions, cell-averaged over a uniform-grid have been developed to analyze the corresponding solutions from the Los Alamos National Laboratory Eulerian Applications Project radiation-hydrodynamics code xRAGE. These Coggeshall solutions have been shown to be independent of heat conduction, providing a unique opportunity for comparison with xRAGE solutions with and without the heat conduction module. Solution convergence was analyzed based on radial step size. Since no shocks are involved in either problem and the solutions are smooth, second-order convergence was expected for both cases. The global L1 errors were used to estimate the convergence rates with and without the heat conduction module implemented.
[Problem-solving in immunohematology: direct compatibility laboratory test ].
Mannessier, L; Roubinet, F; Chiaroni, J
2001-12-01
Cross-matching between the serum of a patient and the red blood cells to be transfused is most important for the prevention of hemolytic transfusion reactions in allo-immunized or new-born patients found positive with direct antiglobulin test. Cross-matching is a time-consuming and complex laboratory test. In order to obtain valid results, it is necessary to abide by some technical rules detailed in this article. The choice of the blood units to be cross-matched depends on the patient's clinical story and on the specificity of anti-erythrocyte antibodies present in the serum. The identification and the management of most frequent difficulties met by using the cross-match technique are discussed hereby. PMID:11802611
Significance testing of rules in rule-based models of human problem solving
NASA Technical Reports Server (NTRS)
Lewis, C. M.; Hammer, J. M.
1986-01-01
Rule-based models of human problem solving have typically not been tested for statistical significance. Three methods of testing rules - analysis of variance, randomization, and contingency tables - are presented. Advantages and disadvantages of the methods are also described.
NASA Technical Reports Server (NTRS)
Scharton, Terry D.
1995-01-01
The intent of this paper is to make a case for developing and conducting vibration tests which are both realistic and practical (a question of tailoring versus standards). Tests are essential for finding things overlooked in the analyses. The best test is often the most realistic test which can be conducted within the cost and budget constraints. Some standards are essential, but the author believes more in the individual's ingenuity to solve a specific problem than in the application of standards which reduce problems (and technology) to their lowest common denominator. Force limited vibration tests and base-drive modal tests are two examples of realistic, but practical testing approaches. Since both of these approaches are relatively new, a number of interesting research problems exist, and these are emphasized herein.
Proof test of the computer program BUCKY for plasticity problems
NASA Technical Reports Server (NTRS)
Smith, James P.
1994-01-01
A theoretical equation describing the elastic-plastic deformation of a cantilever beam subject to a constant pressure is developed. The theoretical result is compared numerically to the computer program BUCKY for the case of an elastic-perfectly plastic specimen. It is shown that the theoretical and numerical results compare favorably in the plastic range. Comparisons are made to another research code to further validate the BUCKY results. This paper serves as a quality test for the computer program BUCKY developed at NASA Johnson Space Center.
Sparse radar imaging using 2D compressed sensing
NASA Astrophysics Data System (ADS)
Hou, Qingkai; Liu, Yang; Chen, Zengping; Su, Shaoying
2014-10-01
Radar imaging is an ill-posed linear inverse problem and compressed sensing (CS) has been proved to have tremendous potential in this field. This paper surveys the theory of radar imaging and a conclusion is drawn that the processing of ISAR imaging can be denoted mathematically as a problem of 2D sparse decomposition. Based on CS, we propose a novel measuring strategy for ISAR imaging radar and utilize random sub-sampling in both range and azimuth dimensions, which will reduce the amount of sampling data tremendously. In order to handle 2D reconstructing problem, the ordinary solution is converting the 2D problem into 1D by Kronecker product, which will increase the size of dictionary and computational cost sharply. In this paper, we introduce the 2D-SL0 algorithm into the reconstruction of imaging. It is proved that 2D-SL0 can achieve equivalent result as other 1D reconstructing methods, but the computational complexity and memory usage is reduced significantly. Moreover, we will state the results of simulating experiments and prove the effectiveness and feasibility of our method.
ERIC Educational Resources Information Center
Toronto Board of Education (Ontario). Research Dept.
In addition to a review of the Differential Aptitude Tests (DAT), a number of other aptitude tests are examined. They are: (1) Flanagan Aptitude Classification Tests, (2) Holzinger-Crowder Uni-Factor Tests, (3) Employee Aptitude Survey, (4) Revised Minnesota Paper Form Board Test, (5) Minnesota Clerical Test, and (6) Turse Clerical Aptitudes Test.…
Molecular testing in oncology: problems, pitfalls and progress.
O'Brien, Cathal P; Taylor, Sarah E; O'Leary, John J; Finn, Stephen P
2014-03-01
Recent advances in the understanding of the molecular basis of cancer and the development of molecular diagnostics based on this knowledge have done much to progress the fields of oncology and pathology. Technological developments such as Next Generation Sequencing (NGS) and multiplex assays have made feasible the widespread adoption of molecular diagnostics for clinical use. While these developments and advances carry much promise, there are pitfalls to implementing this testing. Choosing appropriate biomarkers is a vital first step for clinical use and being able to understand the complex relationship between predictive and prognostic biomarkers is a crucial component of this. Testing for standard of care biomarkers is not straightforward, one must choose carefully between clinical trial assays, assays that analyse the same biological phenomenon or surrogate biomarkers. Sample heterogeneity and population specific difference is assays may skew results and must be controlled for at the assay design stage. At a technical level, NGS has the potential to revolutionise laboratory practice and approaches to cancer treatment. However, use of this technology requires careful planning and implementation if one is to avoid technical and ethical quagmires. Finally, with FDA regulation of companion diagnostics one may be limited to therapy specific assays. PMID:24472389
Sex Differences and Self-Reported Attention Problems During Baseline Concussion Testing.
Brooks, Brian L; Iverson, Grant L; Atkins, Joseph E; Zafonte, Ross; Berkner, Paul D
2016-01-01
Amateur athletic programs often use computerized cognitive testing as part of their concussion management programs. There is evidence that athletes with preexisting attention problems will have worse cognitive performance and more symptoms at baseline testing. The purpose of this study was to examine whether attention problems affect assessments differently for male and female athletes. Participants were drawn from a database that included 6,840 adolescents from Maine who completed Immediate Postconcussion Assessment and Cognitive Testing (ImPACT) at baseline (primary outcome measure). The final sample included 249 boys and 100 girls with self-reported attention problems. Each participant was individually matched for sex, age, number of past concussions, and sport to a control participant (249 boys, 100 girls). Boys with attention problems had worse reaction time than boys without attention problems. Girls with attention problems had worse visual-motor speed than girls without attention problems. Boys with attention problems reported more total symptoms, including more cognitive-sensory and sleep-arousal symptoms, compared with boys without attention problems. Girls with attention problems reported more cognitive-sensory, sleep-arousal, and affective symptoms than girls without attention problems. When considering the assessment, management, and outcome from concussions in adolescent athletes, it is important to consider both sex and preinjury attention problems regarding cognitive test results and symptom reporting. PMID:25923339
Assessing corrosion problems in photovoltaic cells via electrochemical stress testing
NASA Technical Reports Server (NTRS)
Shalaby, H.
1985-01-01
A series of accelerated electrochemical experiments to study the degradation properties of polyvinylbutyral-encapsulated silicon solar cells has been carried out. The cells' electrical performance with silk screen-silver and nickel-solder contacts was evaluated. The degradation mechanism was shown to be electrochemical corrosion of the cell contacts; metallization elements migrate into the encapsulating material, which acts as an ionic conducting medium. The corrosion products form a conductive path which results in a gradual loss of the insulation characteristics of the encapsulant. The precipitation of corrosion products in the encapsulant also contributes to its discoloration which in turn leads to a reduction in its transparency and the consequent optical loss. Delamination of the encapsulating layers could be attributed to electrochemical gas evolution reactions. The usefulness of the testing technique in qualitatively establishing a reliability difference between metallizations and antireflection coating types is demonstrated.
Extension and application of the Preissmann slot model to 2D transient mixed flows
NASA Astrophysics Data System (ADS)
Maranzoni, Andrea; Dazzi, Susanna; Aureli, Francesca; Mignosa, Paolo
2015-08-01
This paper presents an extension of the Preissmann slot concept for the modeling of highly transient two-dimensional (2D) mixed flows. The classic conservative formulation of the 2D shallow water equations for free surface flows is adapted by assuming that two fictitious vertical slots, aligned along the two Cartesian plane directions and normally intersecting, are added on the ceiling of each integration element. Accordingly, transitions between free surface and pressurized flow can be handled in a natural and straightforward way by using the same set of governing equations. The opportunity of coupling free surface and pressurized flows is actually useful not only in one-dimensional (1D) problems concerning sewer systems but also for modeling 2D flooding phenomena in which the pressurization of bridges, culverts, or other crossing hydraulic structures can be expected. Numerical simulations are performed by using a shock-capturing MUSCL-Hancock finite volume scheme combined with the FORCE (First-Order Centred) solver for the evaluation of the numerical fluxes. The validation of the mathematical model is accomplished on the basis of both exact solutions of 1D discontinuous initial value problems and reference radial solutions of idealized test cases with cylindrical symmetry. Furthermore, the capability of the model to deal with practical field-scale applications is assessed by simulating the transit of a bore under an arch bridge. Numerical results show that the proposed model is suitable for the prediction of highly transient 2D mixed flows.
An Approach for Addressing the Multiple Testing Problem in Social Policy Impact Evaluations
ERIC Educational Resources Information Center
Schochet, Peter Z.
2009-01-01
In social policy evaluations, the multiple testing problem occurs due to the many hypothesis tests that are typically conducted across multiple outcomes and subgroups, which can lead to spurious impact findings. This article discusses a framework for addressing this problem that balances Types I and II errors. The framework involves specifying…
ERIC Educational Resources Information Center
Hill, Kennedy T.
1983-01-01
Reviews a 20-year program of research on motivation and test performance, concluding that test anxiety and test-taking skill deficits are distorting factors in efforts to test student aptitude, achievement, and competency. (FL)
ERIC Educational Resources Information Center
Keating, Xiaofen Deng
2003-01-01
This paper aims to examine current nationwide youth fitness test programs, address problems embedded in the programs, and possible solutions. The current Fitnessgram, President's Challenge, and YMCA youth fitness test programs were selected to represent nationwide youth fitness test programs. Sponsors of the nationwide youth fitness test programs…
Formal analysis, hardness, and algorithms for extracting internal structure of test-based problems.
Jaśkowski, Wojciech; Krawiec, Krzysztof
2011-01-01
Problems in which some elementary entities interact with each other are common in computational intelligence. This scenario, typical for coevolving artificial life agents, learning strategies for games, and machine learning from examples, can be formalized as a test-based problem and conveniently embedded in the common conceptual framework of coevolution. In test-based problems, candidate solutions are evaluated on a number of test cases (agents, opponents, examples). It has been recently shown that every test of such problem can be regarded as a separate objective, and the whole problem as multi-objective optimization. Research on reducing the number of such objectives while preserving the relations between candidate solutions and tests led to the notions of underlying objectives and internal problem structure, which can be formalized as a coordinate system that spatially arranges candidate solutions and tests. The coordinate system that spans the minimal number of axes determines the so-called dimension of a problem and, being an inherent property of every problem, is of particular interest. In this study, we investigate in-depth the formalism of a coordinate system and its properties, relate them to properties of partially ordered sets, and design an exact algorithm for finding a minimal coordinate system. We also prove that this problem is NP-hard and come up with a heuristic which is superior to the best algorithm proposed so far. Finally, we apply the algorithms to three abstract problems and demonstrate that the dimension of the problem is typically much lower than the number of tests, and for some problems converges to the intrinsic parameter of the problem--its a priori dimension. PMID:21815770
Inverse problems in the design, modeling and testing of engineering systems
NASA Technical Reports Server (NTRS)
Alifanov, Oleg M.
1991-01-01
Formulations, classification, areas of application, and approaches to solving different inverse problems are considered for the design of structures, modeling, and experimental data processing. Problems in the practical implementation of theoretical-experimental methods based on solving inverse problems are analyzed in order to identify mathematical models of physical processes, aid in input data preparation for design parameter optimization, help in design parameter optimization itself, and to model experiments, large-scale tests, and real tests of engineering systems.
A faster method for 3D/2D medical image registration--a simulation study.
Birkfellner, Wolfgang; Wirth, Joachim; Burgstaller, Wolfgang; Baumann, Bernard; Staedele, Harald; Hammer, Beat; Gellrich, Niels Claudius; Jacob, Augustinus Ludwig; Regazzoni, Pietro; Messmer, Peter
2003-08-21
3D/2D patient-to-computed-tomography (CT) registration is a method to determine a transformation that maps two coordinate systems by comparing a projection image rendered from CT to a real projection image. Iterative variation of the CT's position between rendering steps finally leads to exact registration. Applications include exact patient positioning in radiation therapy, calibration of surgical robots, and pose estimation in computer-aided surgery. One of the problems associated with 3D/2D registration is the fact that finding a registration includes solving a minimization problem in six degrees of freedom (dof) in motion. This results in considerable time requirements since for each iteration step at least one volume rendering has to be computed. We show that by choosing an appropriate world coordinate system and by applying a 2D/2D registration method in each iteration step, the number of iterations can be grossly reduced from n6 to n5. Here, n is the number of discrete variations around a given coordinate. Depending on the configuration of the optimization algorithm, this reduces the total number of iterations necessary to at least 1/3 of it's original value. The method was implemented and extensively tested on simulated x-ray images of a tibia, a pelvis and a skull base. When using one projective image and a discrete full parameter space search for solving the optimization problem, average accuracy was found to be 1.0 +/- 0.6(degrees) and 4.1 +/- 1.9 (mm) for a registration in six parameters, and 1.0 +/- 0.7(degrees) and 4.2 +/- 1.6 (mm) when using the 5 + 1 dof method described in this paper. Time requirements were reduced by a factor 3.1. We conclude that this hardware-independent optimization of 3D/2D registration is a step towards increasing the acceptance of this promising method for a wide number of clinical applications. PMID:12974581
NKG2D ligands as therapeutic targets
Spear, Paul; Wu, Ming-Ru; Sentman, Marie-Louise; Sentman, Charles L.
2013-01-01
The Natural Killer Group 2D (NKG2D) receptor plays an important role in protecting the host from infections and cancer. By recognizing ligands induced on infected or tumor cells, NKG2D modulates lymphocyte activation and promotes immunity to eliminate ligand-expressing cells. Because these ligands are not widely expressed on healthy adult tissue, NKG2D ligands may present a useful target for immunotherapeutic approaches in cancer. Novel therapies targeting NKG2D ligands for the treatment of cancer have shown preclinical success and are poised to enter into clinical trials. In this review, the NKG2D receptor and its ligands are discussed in the context of cancer, infection, and autoimmunity. In addition, therapies targeting NKG2D ligands in cancer are also reviewed. PMID:23833565
Canard configured aircraft with 2-D nozzle
NASA Technical Reports Server (NTRS)
Child, R. D.; Henderson, W. P.
1978-01-01
A closely-coupled canard fighter with vectorable two-dimensional nozzle was designed for enhanced transonic maneuvering. The HiMAT maneuver goal of a sustained 8g turn at a free-stream Mach number of 0.9 and 30,000 feet was the primary design consideration. The aerodynamic design process was initiated with a linear theory optimization minimizing the zero percent suction drag including jet effects and refined with three-dimensional nonlinear potential flow techniques. Allowances were made for mutual interference and viscous effects. The design process to arrive at the resultant configuration is described, and the design of a powered 2-D nozzle model to be tested in the LRC 16-foot Propulsion Wind Tunnel is shown.
2D Electrostatic Actuation of Microshutter Arrays
NASA Technical Reports Server (NTRS)
Burns, Devin E.; Oh, Lance H.; Li, Mary J.; Jones, Justin S.; Kelly, Daniel P.; Zheng, Yun; Kutyrev, Alexander S.; Moseley, Samuel H.
2015-01-01
An electrostatically actuated microshutter array consisting of rotational microshutters (shutters that rotate about a torsion bar) were designed and fabricated through the use of models and experiments. Design iterations focused on minimizing the torsional stiffness of the microshutters, while maintaining their structural integrity. Mechanical and electromechanical test systems were constructed to measure the static and dynamic behavior of the microshutters. The torsional stiffness was reduced by a factor of four over initial designs without sacrificing durability. Analysis of the resonant behavior of the microshutter arrays demonstrates that the first resonant mode is a torsional mode occurring around 3000 Hz. At low vacuum pressures, this resonant mode can be used to significantly reduce the drive voltage necessary for actuation requiring as little as 25V. 2D electrostatic latching and addressing was demonstrated using both a resonant and pulsed addressing scheme.
2D Electrostatic Actuation of Microshutter Arrays
NASA Technical Reports Server (NTRS)
Burns, Devin E.; Oh, Lance H.; Li, Mary J.; Kelly, Daniel P.; Kutyrev, Alexander S.; Moseley, Samuel H.
2015-01-01
Electrostatically actuated microshutter arrays consisting of rotational microshutters (shutters that rotate about a torsion bar) were designed and fabricated through the use of models and experiments. Design iterations focused on minimizing the torsional stiffness of the microshutters, while maintaining their structural integrity. Mechanical and electromechanical test systems were constructed to measure the static and dynamic behavior of the microshutters. The torsional stiffness was reduced by a factor of four over initial designs without sacrificing durability. Analysis of the resonant behavior of the microshutters demonstrates that the first resonant mode is a torsional mode occurring around 3000 Hz. At low vacuum pressures, this resonant mode can be used to significantly reduce the drive voltage necessary for actuation requiring as little as 25V. 2D electrostatic latching and addressing was demonstrated using both a resonant and pulsed addressing scheme.
Beta/gamma test problems for ITS. [Integrated Tiger Series (ITS)
Mei, G.T.
1993-01-01
The Integrated Tiger Series of Coupled Electron/Photon Monte Carlo Transport Codes (ITS 3.0, PC Version) was used at Oak Ridge National Laboratory (ORNL) to compare with and extend the experimental findings of the beta/gamma response of selected health physics instruments. In order to assure that ITS gives correct results, several beta/gamma problems have been tested. ITS was used to simulate these problems numerically, and results for each were compared to the problem's experimental or analytical results. ITS successfully predicted the experimental or analytical results of all tested problems within the statistical uncertainty inherent in the Monte Carlo method.
Some Problems of Computer-Aided Testing and "Interview-Like Tests"
ERIC Educational Resources Information Center
Smoline, D.V.
2008-01-01
Computer-based testing--is an effective teacher's tool, intended to optimize course goals and assessment techniques in a comparatively short time. However, this is accomplished only if we deal with high-quality tests. It is strange, but despite the 100-year history of Testing Theory (see, Anastasi, A., Urbina, S. (1997). Psychological testing.…
49 CFR 40.267 - What problems always cause an alcohol test to be cancelled?
Code of Federal Regulations, 2013 CFR
2013-10-01
... the case of a screening test conducted on a saliva ASD or a breath tube ASD: (1) The STT or BAT reads... 49 Transportation 1 2013-10-01 2013-10-01 false What problems always cause an alcohol test to be... always cause an alcohol test to be cancelled? As an employer, a BAT, or an STT, you must cancel...
49 CFR 40.267 - What problems always cause an alcohol test to be cancelled?
Code of Federal Regulations, 2010 CFR
2010-10-01
... the case of a screening test conducted on a saliva ASD or a breath tube ASD: (1) The STT or BAT reads... 49 Transportation 1 2010-10-01 2010-10-01 false What problems always cause an alcohol test to be... always cause an alcohol test to be cancelled? As an employer, a BAT, or an STT, you must cancel...
49 CFR 40.267 - What problems always cause an alcohol test to be cancelled?
Code of Federal Regulations, 2011 CFR
2011-10-01
... the case of a screening test conducted on a saliva ASD or a breath tube ASD: (1) The STT or BAT reads... 49 Transportation 1 2011-10-01 2011-10-01 false What problems always cause an alcohol test to be... always cause an alcohol test to be cancelled? As an employer, a BAT, or an STT, you must cancel...
On Regularity Criteria for the 2D Generalized MHD System
NASA Astrophysics Data System (ADS)
Jiang, Zaihong; Wang, Yanan; Zhou, Yong
2016-06-01
This paper deals with the problem of regularity criteria for the 2D generalized MHD system with fractional dissipative terms {-Λ^{2α}u} for the velocity field and {-Λ^{2β}b} for the magnetic field respectively. Various regularity criteria are established to guarantee smoothness of solutions. It turns out that our regularity criteria imply previous global existence results naturally.
Dispersionless 2D Toda hierarchy, Hurwitz numbers and Riemann theorem
NASA Astrophysics Data System (ADS)
Natanzon, Sergey M.
2016-01-01
We describe all formal symmetric solutions of dispersionless 2D Toda hierarchy. This classification we use for solving of two classical problems: 1) The calculation of conformal mapping of an arbitrary simply connected domain to the standard disk; 2) Calculation of 2- Hurwitz numbers of genus 0.
2D signature for detection and identification of drugs
NASA Astrophysics Data System (ADS)
Trofimov, Vyacheslav A.; Varentsova, Svetlana A.; Shen, Jingling; Zhang, Cunlin; Zhou, Qingli; Shi, Yulei
2011-06-01
The method of spectral dynamics analysis (SDA-method) is used for obtaining the2D THz signature of drugs. This signature is used for the detection and identification of drugs with similar Fourier spectra by transmitted THz signal. We discuss the efficiency of SDA method for the identification problem of pure methamphetamine (MA), methylenedioxyamphetamine (MDA), 3, 4-methylenedioxymethamphetamine (MDMA) and Ketamine.
NASA Technical Reports Server (NTRS)
Salmon, R. F.; Imbrogno, S.
1976-01-01
The importance of measuring accurate air and fuel flows as well as the importance of obtaining accurate exhaust pollutant measurements were emphasized. Some of the problems and the corrective actions taken to incorporate fixes and/or modifications were identified.
2d PDE Linear Symmetric Matrix Solver
1983-10-01
ICCG2 (Incomplete Cholesky factorized Conjugate Gradient algorithm for 2d symmetric problems) was developed to solve a linear symmetric matrix system arising from a 9-point discretization of two-dimensional elliptic and parabolic partial differential equations found in plasma physics applications, such as resistive MHD, spatial diffusive transport, and phase space transport (Fokker-Planck equation) problems. These problems share the common feature of being stiff and requiring implicit solution techniques. When these parabolic or elliptic PDE''s are discretized withmore » finite-difference or finite-element methods,the resulting matrix system is frequently of block-tridiagonal form. To use ICCG2, the discretization of the two-dimensional partial differential equation and its boundary conditions must result in a block-tridiagonal supermatrix composed of elementary tridiagonal matrices. The incomplete Cholesky conjugate gradient algorithm is used to solve the linear symmetric matrix equation. Loops are arranged to vectorize on the Cray1 with the CFT compiler, wherever possible. Recursive loops, which cannot be vectorized, are written for optimum scalar speed. For matrices lacking symmetry, ILUCG2 should be used. Similar methods in three dimensions are available in ICCG3 and ILUCG3. A general source containing extensions and macros, which must be processed by a pre-compiler to obtain the standard FORTRAN source, is provided along with the standard FORTRAN source because it is believed to be more readable. The pre-compiler is not included, but pre-compilation may be performed by a text editor as described in the UCRL-88746 Preprint.« less
2d PDE Linear Asymmetric Matrix Solver
1983-10-01
ILUCG2 (Incomplete LU factorized Conjugate Gradient algorithm for 2d problems) was developed to solve a linear asymmetric matrix system arising from a 9-point discretization of two-dimensional elliptic and parabolic partial differential equations found in plasma physics applications, such as plasma diffusion, equilibria, and phase space transport (Fokker-Planck equation) problems. These equations share the common feature of being stiff and requiring implicit solution techniques. When these parabolic or elliptic PDE''s are discretized with finite-difference or finite-elementmore » methods, the resulting matrix system is frequently of block-tridiagonal form. To use ILUCG2, the discretization of the two-dimensional partial differential equation and its boundary conditions must result in a block-tridiagonal supermatrix composed of elementary tridiagonal matrices. A generalization of the incomplete Cholesky conjugate gradient algorithm is used to solve the matrix equation. Loops are arranged to vectorize on the Cray1 with the CFT compiler, wherever possible. Recursive loops, which cannot be vectorized, are written for optimum scalar speed. For problems having a symmetric matrix ICCG2 should be used since it runs up to four times faster and uses approximately 30% less storage. Similar methods in three dimensions are available in ICCG3 and ILUCG3. A general source, containing extensions and macros, which must be processed by a pre-compiler to obtain the standard FORTRAN source, is provided along with the standard FORTRAN source because it is believed to be more readable. The pre-compiler is not included, but pre-compilation may be performed by a text editor as described in the UCRL-88746 Preprint.« less
A Test of the Testing Effect: Acquiring Problem-Solving Skills from Worked Examples
ERIC Educational Resources Information Center
van Gog, Tamara; Kester, Liesbeth
2012-01-01
The "testing effect" refers to the finding that after an initial study opportunity, testing is more effective for long-term retention than restudying. The testing effect seems robust and is a finding from the field of cognitive science that has important implications for education. However, it is unclear whether this effect also applies to the…
Prospects and Problems for a National Test: Some Reflections of a Test Author.
ERIC Educational Resources Information Center
Hogan, Thomas P.
Reflections on the proposal for creation and implementation of a national test are presented from the perspective of a test author. The most readily identified characteristic of the proposed national test is the intensity of debate surrounding it. Another easily identified characteristic is the anticipated effect. While proponents expect higher…
Perspectives for spintronics in 2D materials
NASA Astrophysics Data System (ADS)
Han, Wei
2016-03-01
The past decade has been especially creative for spintronics since the (re)discovery of various two dimensional (2D) materials. Due to the unusual physical characteristics, 2D materials have provided new platforms to probe the spin interaction with other degrees of freedom for electrons, as well as to be used for novel spintronics applications. This review briefly presents the most important recent and ongoing research for spintronics in 2D materials.
Likelihood Methods for Testing Group Problem Solving Models with Censored Data.
ERIC Educational Resources Information Center
Regal, Ronald R.; Larntz, Kinley
1978-01-01
Models relating individual and group problem solving solution times under the condition of limited time (time limit censoring) are presented. Maximum likelihood estimation of parameters and a goodness of fit test are presented. (Author/JKS)
Testing foreign language impact on engineering students' scientific problem-solving performance
NASA Astrophysics Data System (ADS)
Tatzl, Dietmar; Messnarz, Bernd
2013-12-01
This article investigates the influence of English as the examination language on the solution of physics and science problems by non-native speakers in tertiary engineering education. For that purpose, a statistically significant total number of 96 students in four year groups from freshman to senior level participated in a testing experiment in the Degree Programme of Aviation at the FH JOANNEUM University of Applied Sciences, Graz, Austria. Half of each test group were given a set of 12 physics problems described in German, the other half received the same set of problems described in English. It was the goal to test linguistic reading comprehension necessary for scientific problem solving instead of physics knowledge as such. The results imply that written undergraduate English-medium engineering tests and examinations may not require additional examination time or language-specific aids for students who have reached university-entrance proficiency in English as a foreign language.
An inverse design method for 2D airfoil
NASA Astrophysics Data System (ADS)
Liang, Zhi-Yong; Cui, Peng; Zhang, Gen-Bao
2010-03-01
The computational method for aerodynamic design of aircraft is applied more universally than before, in which the design of an airfoil is a hot problem. The forward problem is discussed by most relative papers, but inverse method is more useful in practical designs. In this paper, the inverse design of 2D airfoil was investigated. A finite element method based on the variational principle was used for carrying out. Through the simulation, it was shown that the method was fit for the design.
ERIC Educational Resources Information Center
Bennett, Randy Elliot; Morley, Mary; Quardt, Dennis
2000-01-01
Describes three open-ended response types that could broaden the conception of mathematical problem solving used in computerized admissions tests: (1) mathematical expression (ME); (2) generating examples (GE); and (3) and graphical modeling (GM). Illustrates how combining ME, GE, and GM can form extended constructed response problems. (SLD)
Insel, T R; Goodwin, F K
1983-12-01
Diagnostic tests in medicine must satisfy certain validity and accuracy criteria to be clinically useful. In psychiatry, the validity of a diagnostic procedure might be tested independently against clinical diagnosis, treatment response, and family history criteria; a strong relationship to any of the three would suggest clinical usefulness. Predictive value theory provides a model for such a test. The dexamethasone suppression test (DST) has a lower predictive value for major depressive disorder than most conventional laboratory tests used for diagnosis in medicine. In spite of promising early reports, the DST is not predictive of treatment response nor does it appear to identify genetic subtypes of depression. Although no diagnostic laboratory test is currently powerful enough for routine clinical use in psychiatry, laboratory tests may prove useful in predicting relapse and in continuing research on the psychobiology of mental disorders. PMID:6642460
Quantum process tomography by 2D fluorescence spectroscopy
Pachón, Leonardo A.; Marcus, Andrew H.; Aspuru-Guzik, Alán
2015-06-07
Reconstruction of the dynamics (quantum process tomography) of the single-exciton manifold in energy transfer systems is proposed here on the basis of two-dimensional fluorescence spectroscopy (2D-FS) with phase-modulation. The quantum-process-tomography protocol introduced here benefits from, e.g., the sensitivity enhancement ascribed to 2D-FS. Although the isotropically averaged spectroscopic signals depend on the quantum yield parameter Γ of the doubly excited-exciton manifold, it is shown that the reconstruction of the dynamics is insensitive to this parameter. Applications to foundational and applied problems, as well as further extensions, are discussed.
Realdi, Giuseppe; Previato, Lorenzo; Vitturi, Nicola
2008-07-01
The leading function of the physician is the clinical reasoning, which involves appropriate investigation of the problems of the patient, formulation of a diagnostic suspect based on the patient's symptoms and signs, gathering of additional relevant information, to select necessary tests and administration of the most suitable therapy. The problems of the patient are expressed by symptoms or signs or abnormal test results, requested for a variety of reasons. The entire scientific, as well as diagnostic approach, is based on three steps: to stumble in a problem; to try a solution through a hypothesis; to disprove or to prove the hypothesis by a process of criticism. Clinicians use the information obtained from the history and physical examination to estimate initial (or pre-test) probability and then use the results from tests and other diagnostic procedures to modify this probability until the post-test probability is such that the suspected diagnosis is either confirmed or ruled out. When the pre-test probability of disease is high, tests characterized by high specificity will be preferred, in order to confirm the diagnostic suspect. When the pre-test probability of disease is low, a test with high sensitivity is advisable to exclude the hypothetical disease. The above mentioned process of decision making has been transferred to a problem oriented medical record that is currently employed in our Clinic. PMID:18420030
49 CFR 40.203 - What problems cause a drug test to be cancelled unless they are corrected?
Code of Federal Regulations, 2010 CFR
2010-10-01
... 49 Transportation 1 2010-10-01 2010-10-01 false What problems cause a drug test to be cancelled... PROCEDURES FOR TRANSPORTATION WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Drug Tests § 40.203 What problems cause a drug test to be cancelled unless they are corrected? (a) As the MRO, when...
49 CFR 40.269 - What problems cause an alcohol test to be cancelled unless they are corrected?
Code of Federal Regulations, 2010 CFR
2010-10-01
... 49 Transportation 1 2010-10-01 2010-10-01 false What problems cause an alcohol test to be... Transportation PROCEDURES FOR TRANSPORTATION WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Alcohol Testing § 40.269 What problems cause an alcohol test to be cancelled unless they are corrected? As a...
49 CFR 40.203 - What problems cause a drug test to be cancelled unless they are corrected?
Code of Federal Regulations, 2011 CFR
2011-10-01
... 49 Transportation 1 2011-10-01 2011-10-01 false What problems cause a drug test to be cancelled... PROCEDURES FOR TRANSPORTATION WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Drug Tests § 40.203 What problems cause a drug test to be cancelled unless they are corrected? (a) As the MRO, when...
49 CFR 40.269 - What problems cause an alcohol test to be cancelled unless they are corrected?
Code of Federal Regulations, 2011 CFR
2011-10-01
... 49 Transportation 1 2011-10-01 2011-10-01 false What problems cause an alcohol test to be... Transportation PROCEDURES FOR TRANSPORTATION WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Alcohol Testing § 40.269 What problems cause an alcohol test to be cancelled unless they are corrected? As a...
Annotated Bibliography of EDGE2D Use
J.D. Strachan and G. Corrigan
2005-06-24
This annotated bibliography is intended to help EDGE2D users, and particularly new users, find existing published literature that has used EDGE2D. Our idea is that a person can find existing studies which may relate to his intended use, as well as gain ideas about other possible applications by scanning the attached tables.
Staring 2-D hadamard transform spectral imager
Gentry, Stephen M.; Wehlburg, Christine M.; Wehlburg, Joseph C.; Smith, Mark W.; Smith, Jody L.
2006-02-07
A staring imaging system inputs a 2D spatial image containing multi-frequency spectral information. This image is encoded in one dimension of the image with a cyclic Hadamarid S-matrix. The resulting image is detecting with a spatial 2D detector; and a computer applies a Hadamard transform to recover the encoded image.
Raybould, Alan
2006-01-01
Environmental risk assessments can provide high confidence of minimal risk by testing theories, "risk hypotheses", that predict the likelihood of unacceptable harmful events. The creation of risk hypotheses and a plan to test them is called problem formulation. Effective problem formulation seeks to maximize the possibility of detecting effects that indicate potential risk; if such effects are not detected, minimal risk is indicated with high confidence. Two important implications are that artificial test conditions can increase confidence, whereas prescriptive data requirements can reduce confidence (increase uncertainty) if they constrain problem formulation. Poor problem formulation can increase environmental risk because it leads to the collection of superfluous data that may delay or prevent the introduction of environmentally beneficial products. PMID:17445509
Problem-Solving Test: RNA and Protein Synthesis in Bacteriophage-Infected "E. coli" Cells
ERIC Educational Resources Information Center
Szeberenyi, Jozsef
2008-01-01
The classic experiment presented in this problem-solving test was designed to identify the template molecules of translation by analyzing the synthesis of phage proteins in "Escherichia coli" cells infected with bacteriophage T4. The work described in this test led to one of the most seminal discoveries of early molecular biology: it dealt a…
Use of laboratory and field testing to identify potential production problems in the Troll field
Hartley, R.; Jadid, M.B.
1989-02-01
The areal extent of the oil found in Troll made it clear at a very early stage in the field's appraisal that subsea wells would be required if the oil were developed. Owing to cooling in the subsea flowline, subsea wells can be expected to pose more production chemistry problems than would be expected with conventional platform wells. Consequently, a number of laboratory tests were carried out during the appraisal campaign to identify problems to be expected with scaling, foaming, emulsification, wax deposition, and hydrates. Dehydration and wax deposition tests were also carried out offshore during appraisal-well testing. These tests are described, together with the methods subsequently adopted to minimize future production problems.
Almost but not quite 2D, Non-linear Bayesian Inversion of CSEM Data
NASA Astrophysics Data System (ADS)
Ray, A.; Key, K.; Bodin, T.
2013-12-01
efficiently evaluate the forward response using 1D profiles extracted from the model at the common-midpoints of the EM source-receiver pairs. Since the 1D approximation is locally valid at different midpoint locations, the computation time is far lower than is required by a full 2D or 3D simulation. We have applied this method to both synthetic and real CSEM survey data from the Scarborough gas field on the Northwest shelf of Australia, resulting in a spatially variable quantification of resistivity and its uncertainty in 2D. This Bayesian approach results in a large database of 2D models that comprise a posterior probability distribution, which we can subset to test various hypotheses about the range of model structures compatible with the data. For example, we can subset the model distributions to examine the hypothesis that a resistive reservoir extends overs a certain spatial extent. Depending on how this conditions other parts of the model space, light can be shed on the geological viability of the hypothesis. Since tackling spatially variable uncertainty and trade-offs in 2D and 3D is a challenging research problem, the insights gained from this work may prove valuable for subsequent full 2D and 3D Bayesian inversions.
Light field morphing using 2D features.
Wang, Lifeng; Lin, Stephen; Lee, Seungyong; Guo, Baining; Shum, Heung-Yeung
2005-01-01
We present a 2D feature-based technique for morphing 3D objects represented by light fields. Existing light field morphing methods require the user to specify corresponding 3D feature elements to guide morph computation. Since slight errors in 3D specification can lead to significant morphing artifacts, we propose a scheme based on 2D feature elements that is less sensitive to imprecise marking of features. First, 2D features are specified by the user in a number of key views in the source and target light fields. Then the two light fields are warped view by view as guided by the corresponding 2D features. Finally, the two warped light fields are blended together to yield the desired light field morph. Two key issues in light field morphing are feature specification and warping of light field rays. For feature specification, we introduce a user interface for delineating 2D features in key views of a light field, which are automatically interpolated to other views. For ray warping, we describe a 2D technique that accounts for visibility changes and present a comparison to the ideal morphing of light fields. Light field morphing based on 2D features makes it simple to incorporate previous image morphing techniques such as nonuniform blending, as well as to morph between an image and a light field. PMID:15631126
2D materials for nanophotonic devices
NASA Astrophysics Data System (ADS)
Xu, Renjing; Yang, Jiong; Zhang, Shuang; Pei, Jiajie; Lu, Yuerui
2015-12-01
Two-dimensional (2D) materials have become very important building blocks for electronic, photonic, and phononic devices. The 2D material family has four key members, including the metallic graphene, transition metal dichalcogenide (TMD) layered semiconductors, semiconducting black phosphorous, and the insulating h-BN. Owing to the strong quantum confinements and defect-free surfaces, these atomically thin layers have offered us perfect platforms to investigate the interactions among photons, electrons and phonons. The unique interactions in these 2D materials are very important for both scientific research and application engineering. In this talk, I would like to briefly summarize and highlight the key findings, opportunities and challenges in this field. Next, I will introduce/highlight our recent achievements. We demonstrated atomically thin micro-lens and gratings using 2D MoS2, which is the thinnest optical component around the world. These devices are based on our discovery that the elastic light-matter interactions in highindex 2D materials is very strong. Also, I would like to introduce a new two-dimensional material phosphorene. Phosphorene has strongly anisotropic optical response, which creates 1D excitons in a 2D system. The strong confinement in phosphorene also enables the ultra-high trion (charged exciton) binding energies, which have been successfully measured in our experiments. Finally, I will briefly talk about the potential applications of 2D materials in energy harvesting.
Inertial solvation in femtosecond 2D spectra
NASA Astrophysics Data System (ADS)
Hybl, John; Albrecht Ferro, Allison; Farrow, Darcie; Jonas, David
2001-03-01
We have used 2D Fourier transform spectroscopy to investigate polar solvation. 2D spectroscopy can reveal molecular lineshapes beneath ensemble averaged spectra and freeze molecular motions to give an undistorted picture of the microscopic dynamics of polar solvation. The transition from "inhomogeneous" to "homogeneous" 2D spectra is governed by both vibrational relaxation and solvent motion. Therefore, the time dependence of the 2D spectrum directly reflects the total response of the solvent-solute system. IR144, a cyanine dye with a dipole moment change upon electronic excitation, was used to probe inertial solvation in methanol and propylene carbonate. Since the static Stokes' shift of IR144 in each of these solvents is similar, differences in the 2D spectra result from solvation dynamics. Initial results indicate that the larger propylene carbonate responds more slowly than methanol, but appear to be inconsistent with rotational estimates of the inertial response. To disentangle intra-molecular vibrations from solvent motion, the 2D spectra of IR144 will be compared to the time-dependent 2D spectra of the structurally related nonpolar cyanine dye HDITCP.
ERIC Educational Resources Information Center
Breitung, H. A.; And Others
1974-01-01
New placement procedure at Humboldt University includes interviews and placement tests. Interviews reveal response ability, tempo, pronunciation, comprehension, etc. The 60-minute test that follows is described and results discussed, as well as difficulty level and grading. Results: better grouping of students, less shifting, better work. (Text is…
Towards 2D Bayesian Tomography of Receiver Functions
NASA Astrophysics Data System (ADS)
Ray, A.; Bodin, T.; Key, K.
2014-12-01
Receiver function analysis is a powerful tool widely used to isolate and interpret receiver-side structure effects in teleseismic records. The idea is to deconvolve the vertical component from the horizontal components to produce a time series, thus eliminating the influence of the source and distant path effects. The receiver function is usually migrated and directly interpreted by visual inspection. However, deconvolution is a numerically unstable procedure that needs to be stabilized, and the solution depends on the choice of regularization parameters (e.g. water level and the width of a low pass filter). Since the solution is blurred with multiple reflections from the subsurface that produce apparent discontinuities, qualitative interpretation of receiver functions is subjective. Alternatively, waveforms can be directly inverted for a 1D S-wave velocity model beneath the receiver. An inversion procedure is more quantitative, as a forward model will take into account all possible reflections and conversions. If cast in a Bayesian framework, an inversion also enables one to assess model uncertainties and quantify parameter trade-offs. However, seismologists have preferred migration techniques as they are easier to implement, computationally cheaper, and allow construction of 2D or 3D sections. Inversions have been limited thus far to the 1D case. In this work we present a method for inversion of converted waveforms measured at a number of aligned stations. The unknown model is a 2D vertical cross section parameterized with a variable number of discontinuities, although the forward model used to compute synthetics under individual stations is 1D. Body waves are inverted jointly with surface wave dispersion measurements to reduce the range of possible solutions. The problem is solved with a fully non linear Bayesian inversion scheme where the posterior velocity distribution is sampled with a Markov Chain Monte Carlo Algorithm. Our approach uses the 'trans
ELLIPT2D: A Flexible Finite Element Code Written Python
Pletzer, A.; Mollis, J.C.
2001-03-22
The use of the Python scripting language for scientific applications and in particular to solve partial differential equations is explored. It is shown that Python's rich data structure and object-oriented features can be exploited to write programs that are not only significantly more concise than their counter parts written in Fortran, C or C++, but are also numerically efficient. To illustrate this, a two-dimensional finite element code (ELLIPT2D) has been written. ELLIPT2D provides a flexible and easy-to-use framework for solving a large class of second-order elliptic problems. The program allows for structured or unstructured meshes. All functions defining the elliptic operator are user supplied and so are the boundary conditions, which can be of Dirichlet, Neumann or Robbins type. ELLIPT2D makes extensive use of dictionaries (hash tables) as a way to represent sparse matrices.Other key features of the Python language that have been widely used include: operator over loading, error handling, array slicing, and the Tkinter module for building graphical use interfaces. As an example of the utility of ELLIPT2D, a nonlinear solution of the Grad-Shafranov equation is computed using a Newton iterative scheme. A second application focuses on a solution of the toroidal Laplace equation coupled to a magnetohydrodynamic stability code, a problem arising in the context of magnetic fusion research.
NASA Astrophysics Data System (ADS)
Pasternack, Gregory B.; Gilbert, Andrew T.; Wheaton, Joseph M.; Buckland, Evan M.
2006-08-01
SummaryResource managers, scientists, government regulators, and stakeholders are considering sophisticated numerical models for managing complex environmental problems. In this study, observations from a river-rehabilitation experiment involving gravel augmentation and spawning habitat enhancement were used to assess sources and magnitudes of error in depth, velocity, and shear velocity predictions made at the 1-m scale with a commercial two-dimensional (depth-averaged) model. Error in 2D model depth prediction averaged 21%. This error was attributable to topographic survey resolution, which at 1 pt per 1.14 m 2, was inadequate to resolve small humps and depressions influencing point measurements. Error in 2D model velocity prediction averaged 29%. More than half of this error was attributable to depth prediction error. Despite depth and velocity error, 56% of tested 2D model predictions of shear velocity were within the 95% confidence limit of the best field-based estimation method. Ninety percent of the error in shear velocity prediction was explained by velocity prediction error. Multiple field-based estimates of shear velocity differed by up to 160%, so the lower error for the 2D model's predictions suggests such models are at least as accurate as field measurement. 2D models enable detailed, spatially distributed estimates compared to the small number measurable in a field campaign of comparable cost. They also can be used for design evaluation. Although such numerical models are limited to channel types adhering to model assumptions and yield predictions only accurate to ˜20-30%, they can provide a useful tool for river-rehabilitation design and assessment, including spatially diverse habitat heterogeneity as well as for pre- and post-project appraisal.
NASA Technical Reports Server (NTRS)
Bromley, L. K.; Travis, A. D.
1980-01-01
The compatibility and performance of the Shuttle communications system must be certified prior to operational missions. For this purpose, NASA has established the Electronics Systems Test Laboratory (ESTL) at the Johnson Space Center. This paper discusses the Shuttle communications system compatibility and performance testing being performed in the ESTL. The ESTL system verification test philosophy, including capabilities, procedures, and unique testing equipment are summarized. Summaries of the significant results of compatibility and performance tests of the Orbiter/Space-flight Tracking and Data Network, Orbiter/Air Force Remote Tracking Station, Orbiter/Tracking and Data Relay Satellite System and Orbiter/Shuttle Launch Support System interfaces are presented. The ESTL's unique ability to locate potential communication problems and participate in the resolution of these problems are discussed in detail.
Design of the LRP airfoil series using 2D CFD
NASA Astrophysics Data System (ADS)
Zahle, Frederik; Bak, Christian; Sørensen, Niels N.; Vronsky, Tomas; Gaudern, Nicholas
2014-06-01
This paper describes the design and wind tunnel testing of a high-Reynolds number, high lift airfoil series designed for wind turbines. The airfoils were designed using direct gradient- based numerical multi-point optimization based on a Bezier parameterization of the shape, coupled to the 2D Navier-Stokes flow solver EllipSys2D. The resulting airfoils, the LRP2-30 and LRP2-36, achieve both higher operational lift coefficients and higher lift to drag ratios compared to the equivalent FFA-W3 airfoils.
Evaluation of 2D ceramic matrix composites in aeroconvective environments
NASA Technical Reports Server (NTRS)
Riccitiello, Salvatore R.; Love, Wendell L.; Balter-Peterson, Aliza
1992-01-01
An evaluation is conducted of a novel ceramic-matrix composite (CMC) material system for use in the aeroconvective-heating environments encountered by the nose caps and wing leading edges of such aerospace vehicles as the Space Shuttle, during orbit-insertion and reentry from LEO. These CMCs are composed of an SiC matrix that is reinforced with Nicalon, Nextel, or carbon refractory fibers in a 2D architecture. The test program conducted for the 2D CMCs gave attention to their subsurface oxidation.
A New 2D-Transport, 1D-Diffusion Approximation of the Boltzmann Transport equation
Larsen, Edward
2013-06-17
The work performed in this project consisted of the derivation, implementation, and testing of a new, computationally advantageous approximation to the 3D Boltz- mann transport equation. The solution of the Boltzmann equation is the neutron flux in nuclear reactor cores and shields, but solving this equation is difficult and costly. The new “2D/1D” approximation takes advantage of a special geometric feature of typical 3D reactors to approximate the neutron transport physics in a specific (ax- ial) direction, but not in the other two (radial) directions. The resulting equation is much less expensive to solve computationally, and its solutions are expected to be sufficiently accurate for many practical problems. In this project we formulated the new equation, discretized it using standard methods, developed a stable itera- tion scheme for solving the equation, implemented the new numerical scheme in the MPACT code, and tested the method on several realistic problems. All the hoped- for features of this new approximation were seen. For large, difficult problems, the resulting 2D/1D solution is highly accurate, and is calculated about 100 times faster than a 3D discrete ordinates simulation.
Ginsparg, P.
1991-01-01
These are introductory lectures for a general audience that give an overview of the subject of matrix models and their application to random surfaces, 2d gravity, and string theory. They are intentionally 1.5 years out of date.
Ginsparg, P.
1991-12-31
These are introductory lectures for a general audience that give an overview of the subject of matrix models and their application to random surfaces, 2d gravity, and string theory. They are intentionally 1.5 years out of date.
2D electronic materials for army applications
NASA Astrophysics Data System (ADS)
O'Regan, Terrance; Perconti, Philip
2015-05-01
The record electronic properties achieved in monolayer graphene and related 2D materials such as molybdenum disulfide and hexagonal boron nitride show promise for revolutionary high-speed and low-power electronic devices. Heterogeneous 2D-stacked materials may create enabling technology for future communication and computation applications to meet soldier requirements. For instance, transparent, flexible and even wearable systems may become feasible. With soldier and squad level electronic power demands increasing, the Army is committed to developing and harnessing graphene-like 2D materials for compact low size-weight-and-power-cost (SWAP-C) systems. This paper will review developments in 2D electronic materials at the Army Research Laboratory over the last five years and discuss directions for future army applications.
Chemical Approaches to 2D Materials.
Samorì, Paolo; Palermo, Vincenzo; Feng, Xinliang
2016-08-01
Chemistry plays an ever-increasing role in the production, functionalization, processing and applications of graphene and other 2D materials. This special issue highlights a selection of enlightening chemical approaches to 2D materials, which nicely reflect the breadth of the field and convey the excitement of the individuals involved in it, who are trying to translate graphene and related materials from the laboratory into a real, high-impact technology. PMID:27478083
Confusion and clairvoyance: some remarks on the composite hypothesis testing problem
NASA Astrophysics Data System (ADS)
Theiler, James
2012-06-01
This paper discusses issues related to the inherent ambiguity of the composite hypothesis testing problem, a problem that is central to the detection of target signals in cluttered backgrounds. In particular, the paper examines the recently proposed method of continuum fusion (which, because it combines an ensemble of clairvoyant detectors, might also be called clairvoyant fusion), and its relationship to other strategies for composite hypothesis testing. A specific example involving the affine subspace model adds to the confusion by illustrating irreconcilable differences between Bayesian and non-Bayesian approaches to target detection.
Barth, A; Küfferle, B
2001-11-01
Concretism is considered an important aspect of schizophrenic thought disorder. Traditionally it is measured using the method of proverb interpretation, in which metaphoric proverbs are presented with the request that the subject tell its meaning. Interpretations are recorded and scored on concretistic tendencies. However, this method has two problems: its reliability is doubtful and it is rather complicated to perform. In this paper, a new version of a multiple choice proverb test is presented which can solve these problems in a reliable and economic manner. Using the new test, it is has been shown that schizophrenic patients have greater deficits in proverb interpretation than depressive patients. PMID:11758092
Simultaneous 2D Strain Sensing Using Polymer Planar Bragg Gratings
Rosenberger, Manuel; Eisenbeil, Waltraud; Schmauss, Bernhard; Hellmann, Ralf
2015-01-01
We demonstrate the application of polymer planar Bragg gratings for multi-axial strain sensing and particularly highlight simultaneous 2D strain measurement. A polymer planar Bragg grating (PPBG) fabricated with a single writing step in bulk polymethylmethacrylate is used for measuring both tensile and compressive strain at various angles. It is shown that the sensitivity of the PPBG strongly depends on the angle between the optical waveguide into which the grating is inscribed and the direction along which the mechanical load is applied. Additionally, a 2D PPBG fabricated by writing two Bragg gratings angularly displaced from each other into a single polymer platelet is bonded to a stainless steel plate. The two reflected wavelengths exhibit different sensitivities while tested toward tensile and compressive strain. These characteristics make 2D PPBG suitable for measuring multi-axial tensile and compressive strain. PMID:25686313
Simultaneous 2D strain sensing using polymer planar Bragg gratings.
Rosenberger, Manuel; Eisenbeil, Waltraud; Schmauss, Bernhard; Hellmann, Ralf
2015-01-01
We demonstrate the application of polymer planar Bragg gratings for multi-axial strain sensing and particularly highlight simultaneous 2D strain measurement. A polymer planar Bragg grating (PPBG) fabricated with a single writing step in bulk polymethylmethacrylate is used for measuring both tensile and compressive strain at various angles. It is shown that the sensitivity of the PPBG strongly depends on the angle between the optical waveguide into which the grating is inscribed and the direction along which the mechanical load is applied. Additionally, a 2D PPBG fabricated by writing two Bragg gratings angularly displaced from each other into a single polymer platelet is bonded to a stainless steel plate. The two reflected wavelengths exhibit different sensitivities while tested toward tensile and compressive strain. These characteristics make 2D PPBG suitable for measuring multi-axial tensile and compressive strain. PMID:25686313
2D FEM Heat Transfer & E&M Field Code
1992-04-02
TOPAZ and TOPAZ2D are two-dimensional implicit finite element computer codes for heat transfer analysis. TOPAZ2D can also be used to solve electrostatic and magnetostatic problems. The programs solve for the steady-state or transient temperature or electrostatic and magnetostatic potential field on two-dimensional planar or axisymmetric geometries. Material properties may be temperature or potential-dependent and either isotropic or orthotropic. A variety of time and temperature-dependent boundary conditions can be specified including temperature, flux, convection, and radiation.more » By implementing the user subroutine feature, users can model chemical reaction kinetics and allow for any type of functional representation of boundary conditions and internal heat generation. The programs can solve problems of diffuse and specular band radiation in an enclosure coupled with conduction in the material surrounding the enclosure. Additional features include thermal contact resistance across an interface, bulk fluids, phase change, and energy balances.« less
2D FEM Heat Transfer & E&M Field Code
1992-04-02
TOPAZ and TOPAZ2D are two-dimensional implicit finite element computer codes for heat transfer analysis. TOPAZ2D can also be used to solve electrostatic and magnetostatic problems. The programs solve for the steady-state or transient temperature or electrostatic and magnetostatic potential field on two-dimensional planar or axisymmetric geometries. Material properties may be temperature or potential-dependent and either isotropic or orthotropic. A variety of time and temperature-dependent boundary conditions can be specified including temperature, flux, convection, and radiation. By implementing the user subroutine feature, users can model chemical reaction kinetics and allow for any type of functional representation of boundary conditions and internal heat generation. The programs can solve problems of diffuse and specular band radiation in an enclosure coupled with conduction in the material surrounding the enclosure. Additional features include thermal contact resistance across an interface, bulk fluids, phase change, and energy balances.
A multigroup radiation diffusion test problem: Comparison of code results with analytic solution
Shestakov, A I; Harte, J A; Bolstad, J H; Offner, S R
2006-12-21
We consider a 1D, slab-symmetric test problem for the multigroup radiation diffusion and matter energy balance equations. The test simulates diffusion of energy from a hot central region. Opacities vary with the cube of the frequency and radiation emission is given by a Wien spectrum. We compare results from two LLNL codes, Raptor and Lasnex, with tabular data that define the analytic solution.
ERIC Educational Resources Information Center
Dandreaux, Danielle M.; Frick, Paul J.
2009-01-01
This study tested several theoretically important differences between youth with a childhood-onset and youth with an adolescent-onset to their severe conduct problems. Seventy-eight pre-adjudicated adolescent boys (ranging in age from 11 to 18) housed in two short-term detention facilities and one outpatient program for youth at risk for…
ERIC Educational Resources Information Center
Educational Testing Service, Princeton, NJ.
Three topics were discussed at this conference: (1) the supply and identification of high level talent; (2) problems of evaluation in general education; and (3) the development of tests for the measurement of non-intellectural functions. Tony Oxtoby and John T. Dailey discussed military and civilian personnel research on talent identification and…
Testing Foreign Language Impact on Engineering Students' Scientific Problem-Solving Performance
ERIC Educational Resources Information Center
Tatzl, Dietmar; Messnarz, Bernd
2013-01-01
This article investigates the influence of English as the examination language on the solution of physics and science problems by non-native speakers in tertiary engineering education. For that purpose, a statistically significant total number of 96 students in four year groups from freshman to senior level participated in a testing experiment in…
ERIC Educational Resources Information Center
Smith, Mike U.
Both teachers and students alike acknowledge that genetics and genetics problem-solving are extremely difficult to learn and to teach. Therefore, a number of recommendations for teaching college genetics are offered. Although few of these ideas have as yet been tested in controlled experiments, they are supported by research and experience and may…
3D multiple-point statistics simulation using 2D training images
NASA Astrophysics Data System (ADS)
Comunian, A.; Renard, P.; Straubhaar, J.
2012-03-01
One of the main issues in the application of multiple-point statistics (MPS) to the simulation of three-dimensional (3D) blocks is the lack of a suitable 3D training image. In this work, we compare three methods of overcoming this issue using information coming from bidimensional (2D) training images. One approach is based on the aggregation of probabilities. The other approaches are novel. One relies on merging the lists obtained using the impala algorithm from diverse 2D training images, creating a list of compatible data events that is then used for the MPS simulation. The other (s2Dcd) is based on sequential simulations of 2D slices constrained by the conditioning data computed at the previous simulation steps. These three methods are tested on the reproduction of two 3D images that are used as references, and on a real case study where two training images of sedimentary structures are considered. The tests show that it is possible to obtain 3D MPS simulations with at least two 2D training images. The simulations obtained, in particular those obtained with the s2Dcd method, are close to the references, according to a number of comparison criteria. The CPU time required to simulate with the method s2Dcd is from two to four orders of magnitude smaller than the one required by a MPS simulation performed using a 3D training image, while the results obtained are comparable. This computational efficiency and the possibility of using MPS for 3D simulation without the need for a 3D training image facilitates the inclusion of MPS in Monte Carlo, uncertainty evaluation, and stochastic inverse problems frameworks.
Emotional Intelligence and Problem Solving Strategy: Comparative Study Basedon "Tower of Hanoi" Test
Arefnasab, Zahra; Zare, Hosein; Babamahmoodi, Abdolreza
2012-01-01
Objective: The aim of this study was to compare problem solving strategies between peoples with high and low emotional intelligence (EI). Methods: This study is a cross sectional descriptive study.The sample groups include senior BS& BA between 20-30 years old into two with high and low emotional intelligence, each group had 30 subjects.Data was analyzed with non-parametric chi square test for main dependent variable (problem solving strategies) and accessory dependent variables(manner of starting and fulfillmentof the test).The Independent two group T-test was used for analyzing other accessory dependent variables(Number of errors and total time used for fulfillment of the test). Results: There was a significant difference between two groups in “number of errors” (t=-3.67,p=0) and “total time used for fulfillment of the test”(-6.17,p=0) and there was significant relation between EI and “problem solving strategies” (χ2=25.71, p<0.01) and (Cramer's v = 0.65, p<0.01) .Also there was significant relation between EI and “fulfillment of test” (χ2=20.31, p<0.01) and (φ=0.58, p<0.01). But the relation between EI and "manner of starting the test" was not significant (χ2=1.11, p=0.29). Subjects with high EI used more “insightful” strategy and subjects with low EI used more “trial- error” strategy. The first group completed the test more rapidlyand with fewer errors, compared with the second group. In addition the first group was more successful in performing the test than the second one. Conclusion: People with high EI significantly solve problems better than people with lowEI. PMID:24644484
Finite Element Analysis of 2-D Elastic Contacts Involving FGMs
NASA Astrophysics Data System (ADS)
Abhilash, M. N.; Murthy, H.
2014-05-01
The response of elastic indenters in contact with Functionally Graded Material (FGM) coated homogeneous elastic half space has been presented in the current paper. Finite element analysis has been used due to its ability to handle complex geometry, material, and boundary conditions. Indenters of different typical surface profiles have been considered and the problem has been idealized as a two-dimensional (2D) plane strain problem considering only normal loads. Initially, indenters were considered to be rigid and the results were validated with the solutions presented in the literature. The analysis has then been extended to the case of elastic indenters on FGM-coated half spaces and the results are discussed.
A parallel splitting wavelet method for 2D conservation laws
NASA Astrophysics Data System (ADS)
Schmidt, Alex A.; Kozakevicius, Alice J.; Jakobsson, Stefan
2016-06-01
The current work presents a parallel formulation using the MPI protocol for an adaptive high order finite difference scheme to solve 2D conservation laws. Adaptivity is achieved at each time iteration by the application of an interpolating wavelet transform in each space dimension. High order approximations for the numerical fluxes are computed by ENO and WENO schemes. Since time evolution is made by a TVD Runge-Kutta space splitting scheme, the problem is naturally suitable for parallelization. Numerical simulations and speedup results are presented for Euler equations in gas dynamics problems.
Rowley-Neale, Samuel J; Fearn, Jamie M; Brownson, Dale A C; Smith, Graham C; Ji, Xiaobo; Banks, Craig E
2016-08-21
Two-dimensional molybdenum disulphide nanosheets (2D-MoS2) have proven to be an effective electrocatalyst, with particular attention being focused on their use towards increasing the efficiency of the reactions associated with hydrogen fuel cells. Whilst the majority of research has focused on the Hydrogen Evolution Reaction (HER), herein we explore the use of 2D-MoS2 as a potential electrocatalyst for the much less researched Oxygen Reduction Reaction (ORR). We stray from literature conventions and perform experiments in 0.1 M H2SO4 acidic electrolyte for the first time, evaluating the electrochemical performance of the ORR with 2D-MoS2 electrically wired/immobilised upon several carbon based electrodes (namely; Boron Doped Diamond (BDD), Edge Plane Pyrolytic Graphite (EPPG), Glassy Carbon (GC) and Screen-Printed Electrodes (SPE)) whilst exploring a range of 2D-MoS2 coverages/masses. Consequently, the findings of this study are highly applicable to real world fuel cell applications. We show that significant improvements in ORR activity can be achieved through the careful selection of the underlying/supporting carbon materials that electrically wire the 2D-MoS2 and utilisation of an optimal mass of 2D-MoS2. The ORR onset is observed to be reduced to ca. +0.10 V for EPPG, GC and SPEs at 2D-MoS2 (1524 ng cm(-2) modification), which is far closer to Pt at +0.46 V compared to bare/unmodified EPPG, GC and SPE counterparts. This report is the first to demonstrate such beneficial electrochemical responses in acidic conditions using a 2D-MoS2 based electrocatalyst material on a carbon-based substrate (SPEs in this case). Investigation of the beneficial reaction mechanism reveals the ORR to occur via a 4 electron process in specific conditions; elsewhere a 2 electron process is observed. This work offers valuable insights for those wishing to design, fabricate and/or electrochemically test 2D-nanosheet materials towards the ORR. PMID:27448174
Simulator test to study hot-flow problems related to a gas cooled reactor
NASA Technical Reports Server (NTRS)
Poole, J. W.; Freeman, M. P.; Doak, K. W.; Thorpe, M. L.
1973-01-01
An advance study of materials, fuel injection, and hot flow problems related to the gas core nuclear rocket is reported. The first task was to test a previously constructed induction heated plasma GCNR simulator above 300 kW. A number of tests are reported operating in the range of 300 kW at 10,000 cps. A second simulator was designed but not constructed for cold-hot visualization studies using louvered walls. A third task was a paper investigation of practical uranium feed systems, including a detailed discussion of related problems. The last assignment resulted in two designs for plasma nozzle test devices that could be operated at 200 atm on hydrogen.
NASA Technical Reports Server (NTRS)
Gelinas, R. J.; Doss, S. K.; Vajk, J. P.; Djomehri, J.; Miller, K.
1983-01-01
The mathematical background regarding the moving finite element (MFE) method of Miller and Miller (1981) is discussed, taking into account a general system of partial differential equations (PDE) and the amenability of the MFE method in two dimensions to code modularization and to semiautomatic user-construction of numerous PDE systems for both Dirichlet and zero-Neumann boundary conditions. A description of test problem results is presented, giving attention to aspects of single square wave propagation, and a solution of the heat equation.
Assessment of the 2D MOC solver in MPACT: Michigan parallel characteristics transport code
Collins, B.; Kochunas, B.; Downar, T.
2013-07-01
MPACT (Michigan Parallel Characteristics Transport Code) is a new reactor analysis tool being developed by researchers at the University of Michigan as an advanced pin-resolved transport capability within VERA (Virtual Environment for Reactor Analysis). VERA is the end-user reactor simulation tool being developed by the Consortium for the Advanced Simulation of Light Water Reactors (CASL). The MPACT development project is itself unique for the way it is changing how students perform research to achieve the instructional and research goals of an academic institution, while providing immediate value to the industry. One of the major computational pieces in MPACT is the 2D MOC solver. It is critical that the 2D MOC solver provide an efficient, accurate, and robust solution over a broad range of reactor operating conditions. The C5G7 benchmark is first used to test the accuracy of the method with a fixed set of cross-sections. The VERA Core Physics Progression Problems are then used to compare the accuracy of both the 2D transport solver and also the cross-section treatments. (authors)
Code of Federal Regulations, 2010 CFR
2010-10-01
... 49 Transportation 1 2010-10-01 2010-10-01 false What problem requires corrective action but does... of Transportation PROCEDURES FOR TRANSPORTATION WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Drug Tests § 40.208 What problem requires corrective action but does not result in...
Code of Federal Regulations, 2011 CFR
2011-10-01
... 49 Transportation 1 2011-10-01 2011-10-01 false What problem requires corrective action but does... of Transportation PROCEDURES FOR TRANSPORTATION WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Drug Tests § 40.208 What problem requires corrective action but does not result in...
Optical modulators with 2D layered materials
NASA Astrophysics Data System (ADS)
Sun, Zhipei; Martinez, Amos; Wang, Feng
2016-04-01
Light modulation is an essential operation in photonics and optoelectronics. With existing and emerging technologies increasingly demanding compact, efficient, fast and broadband optical modulators, high-performance light modulation solutions are becoming indispensable. The recent realization that 2D layered materials could modulate light with superior performance has prompted intense research and significant advances, paving the way for realistic applications. In this Review, we cover the state of the art of optical modulators based on 2D materials, including graphene, transition metal dichalcogenides and black phosphorus. We discuss recent advances employing hybrid structures, such as 2D heterostructures, plasmonic structures, and silicon and fibre integrated structures. We also take a look at the future perspectives and discuss the potential of yet relatively unexplored mechanisms, such as magneto-optic and acousto-optic modulation.
Large Area Synthesis of 2D Materials
NASA Astrophysics Data System (ADS)
Vogel, Eric
Transition metal dichalcogenides (TMDs) have generated significant interest for numerous applications including sensors, flexible electronics, heterostructures and optoelectronics due to their interesting, thickness-dependent properties. Despite recent progress, the synthesis of high-quality and highly uniform TMDs on a large scale is still a challenge. In this talk, synthesis routes for WSe2 and MoS2 that achieve monolayer thickness uniformity across large area substrates with electrical properties equivalent to geological crystals will be described. Controlled doping of 2D semiconductors is also critically required. However, methods established for conventional semiconductors, such as ion implantation, are not easily applicable to 2D materials because of their atomically thin structure. Redox-active molecular dopants will be demonstrated which provide large changes in carrier density and workfunction through the choice of dopant, treatment time, and the solution concentration. Finally, several applications of these large-area, uniform 2D materials will be described including heterostructures, biosensors and strain sensors.
2D microwave imaging reflectometer electronics
Spear, A. G.; Domier, C. W. Hu, X.; Muscatello, C. M.; Ren, X.; Luhmann, N. C.; Tobias, B. J.
2014-11-15
A 2D microwave imaging reflectometer system has been developed to visualize electron density fluctuations on the DIII-D tokamak. Simultaneously illuminated at four probe frequencies, large aperture optics image reflections from four density-dependent cutoff surfaces in the plasma over an extended region of the DIII-D plasma. Localized density fluctuations in the vicinity of the plasma cutoff surfaces modulate the plasma reflections, yielding a 2D image of electron density fluctuations. Details are presented of the receiver down conversion electronics that generate the in-phase (I) and quadrature (Q) reflectometer signals from which 2D density fluctuation data are obtained. Also presented are details on the control system and backplane used to manage the electronics as well as an introduction to the computer based control program.
2D microwave imaging reflectometer electronics
NASA Astrophysics Data System (ADS)
Spear, A. G.; Domier, C. W.; Hu, X.; Muscatello, C. M.; Ren, X.; Tobias, B. J.; Luhmann, N. C.
2014-11-01
A 2D microwave imaging reflectometer system has been developed to visualize electron density fluctuations on the DIII-D tokamak. Simultaneously illuminated at four probe frequencies, large aperture optics image reflections from four density-dependent cutoff surfaces in the plasma over an extended region of the DIII-D plasma. Localized density fluctuations in the vicinity of the plasma cutoff surfaces modulate the plasma reflections, yielding a 2D image of electron density fluctuations. Details are presented of the receiver down conversion electronics that generate the in-phase (I) and quadrature (Q) reflectometer signals from which 2D density fluctuation data are obtained. Also presented are details on the control system and backplane used to manage the electronics as well as an introduction to the computer based control program.
2D microwave imaging reflectometer electronics.
Spear, A G; Domier, C W; Hu, X; Muscatello, C M; Ren, X; Tobias, B J; Luhmann, N C
2014-11-01
A 2D microwave imaging reflectometer system has been developed to visualize electron density fluctuations on the DIII-D tokamak. Simultaneously illuminated at four probe frequencies, large aperture optics image reflections from four density-dependent cutoff surfaces in the plasma over an extended region of the DIII-D plasma. Localized density fluctuations in the vicinity of the plasma cutoff surfaces modulate the plasma reflections, yielding a 2D image of electron density fluctuations. Details are presented of the receiver down conversion electronics that generate the in-phase (I) and quadrature (Q) reflectometer signals from which 2D density fluctuation data are obtained. Also presented are details on the control system and backplane used to manage the electronics as well as an introduction to the computer based control program. PMID:25430247
2D-Crystal-Based Functional Inks.
Bonaccorso, Francesco; Bartolotta, Antonino; Coleman, Jonathan N; Backes, Claudia
2016-08-01
The possibility to produce and process graphene, related 2D crystals, and heterostructures in the liquid phase makes them promising materials for an ever-growing class of applications as composite materials, sensors, in flexible optoelectronics, and energy storage and conversion. In particular, the ability to formulate functional inks with on-demand rheological and morphological properties, i.e., lateral size and thickness of the dispersed 2D crystals, is a step forward toward the development of industrial-scale, reliable, inexpensive printing/coating processes, a boost for the full exploitation of such nanomaterials. Here, the exfoliation strategies of graphite and other layered crystals are reviewed, along with the advances in the sorting of lateral size and thickness of the exfoliated sheets together with the formulation of functional inks and the current development of printing/coating processes of interest for the realization of 2D-crystal-based devices. PMID:27273554
Infinitesimal-area 2D radiative analysis using parametric surface representation, through NURBS
Daun, K.J.; Hollands, K.G.T.
1999-07-01
The use of form factors in the treatment of radiant enclosures requires that the radiosity and surface properties be treated as uniform over finite areas. This restriction can be relaxed by applying an infinitesimal-area analysis, where the radiant exchange is taken to be between infinitesimal areas, rather than finite areas. This paper presents a generic infinitesimal-area formulation that can be applied to two-dimensional enclosure problems. (Previous infinitesimal-area analyses have largely been restricted to specific, one-dimensional problems.) Specifically, the paper shows how the analytical expression for the kernel of the integral equation can be obtained without human intervention, once the enclosure surface has been defined parametrically. This can be accomplished by using a computer algebra package or by using NURBS algorithms, which are the industry standard for the geometrical representations used in CAD-CAM codes. Once the kernel has been obtained by this formalism, the 2D integral equation can be set up and solved numerically. The result is a single general-purpose infinitesimal-area analysis code that can proceed from surface specification to solution. The authors have implemented this 2D code and tested it on 1D problems, whose solutions have been given in the literature, obtaining agreement commensurate with the accuracy of the published solutions.
The 2D lingual appliance system.
Cacciafesta, Vittorio
2013-09-01
The two-dimensional (2D) lingual bracket system represents a valuable treatment option for adult patients seeking a completely invisible orthodontic appliance. The ease of direct or simplified indirect bonding of 2D lingual brackets in combination with low friction mechanics makes it possible to achieve a good functional and aesthetic occlusion, even in the presence of a severe malocclusion. The use of a self-ligating bracket significantly reduces chair-side time for the orthodontist, and the low-profile bracket design greatly improves patient comfort. PMID:24005953
Inkjet printing of 2D layered materials.
Li, Jiantong; Lemme, Max C; Östling, Mikael
2014-11-10
Inkjet printing of 2D layered materials, such as graphene and MoS2, has attracted great interests for emerging electronics. However, incompatible rheology, low concentration, severe aggregation and toxicity of solvents constitute critical challenges which hamper the manufacturing efficiency and product quality. Here, we introduce a simple and general technology concept (distillation-assisted solvent exchange) to efficiently overcome these challenges. By implementing the concept, we have demonstrated excellent jetting performance, ideal printing patterns and a variety of promising applications for inkjet printing of 2D layered materials. PMID:25169938
Measurement of 2D birefringence distribution
NASA Astrophysics Data System (ADS)
Noguchi, Masato; Ishikawa, Tsuyoshi; Ohno, Masahiro; Tachihara, Satoru
1992-10-01
A new measuring method of 2-D birefringence distribution has been developed. It has not been an easy job to get a birefringence distribution in an optical element with conventional ellipsometry because of its lack of scanning means. Finding an analogy between the rotating analyzer method in ellipsometry and the phase-shifting method in recently developed digital interferometry, we have applied the phase-shifting algorithm to ellipsometry, and have developed a new method that makes the measurement of 2-D birefringence distribution easy and possible. The system contains few moving parts, assuring reliability, and measures a large area of a sample at one time, making the measuring time very short.
Potential role of CYP2D6 in the central nervous system
Cheng, Jie; Zhen, Yueying; Miksys, Sharon; Beyoğlu, Diren; Krausz, Kristopher W.; Tyndale, Rachel F.; Yu, Aiming; Idle, Jeffrey R.; Gonzalez, Frank J.
2013-01-01
Cytochrome P450 2D6 (CYP2D6) is a pivotal enzyme responsible for a major human drug oxidation polymorphism in human populations. Distribution of CYP2D6 in brain and its role in serotonin metabolism suggest this CYP2D6 may have a function in central nervous system. To establish an efficient and accurate platform for the study of CYP2D6 in vivo, a transgenic human CYP2D6 (Tg-2D6) model was generated by transgenesis in wild-type C57BL/6 (WT) mice using a P1 phage artificial chromosome clone containing the complete human CYP2D locus, including CYP2D6 gene and 5’- and 3’- flanking sequences. Human CYP2D6 was expressed not only in the liver, but also in brain. The abundance of serotonin and 5-hydroxyindoleacetic acid in brain of Tg-2D6 is higher than in WT mice either basal levels or after harmaline induction. Metabolomics of brain homogenate and cerebrospinal fluid revealed a significant up-regulation of l-carnitine, acetyl-l-carnitine, pantothenic acid, dCDP, anandamide, N-acetylglucosaminylamine, and a down-regulation of stearoyl-l-carnitine in Tg-2D6 mice compared with WT mice. Anxiety tests indicate Tg-2D6 mice have a higher capability to adapt to anxiety. Overall, these findings indicate that the Tg-2D6 mouse model may serve as a valuable in vivo tool to determine CYP2D6-involved neurophysiological metabolism and function. PMID:23614566
Active exterior cloaking for the 2D Laplace and Helmholtz equations.
Vasquez, Fernando Guevara; Milton, Graeme W; Onofrei, Daniel
2009-08-14
A new cloaking method is presented for 2D quasistatics and the 2D Helmholtz equation that we speculate extends to other linear wave equations. For 2D quasistatics it is proven how a single active exterior cloaking device can be used to shield an object from surrounding fields, yet produce very small scattered fields. The problem is reduced to finding a polynomial which is close to 1 in a disk and close to 0 in another disk, and such a polynomial is constructed. For the 2D Helmholtz equation it is numerically shown that three exterior cloaking devices placed around the object suffice to hide it. PMID:19792644
Optimal design of 2D digital filters based on neural networks
NASA Astrophysics Data System (ADS)
Wang, Xiao-hua; He, Yi-gang; Zheng, Zhe-zhao; Zhang, Xu-hong
2005-02-01
Two-dimensional (2-D) digital filters are widely useful in image processing and other 2-D digital signal processing fields,but designing 2-D filters is much more difficult than designing one-dimensional (1-D) ones.In this paper, a new design approach for designing linear-phase 2-D digital filters is described,which is based on a new neural networks algorithm (NNA).By using the symmetry of the given 2-D magnitude specification,a compact express for the magnitude response of a linear-phase 2-D finite impulse response (FIR) filter is derived.Consequently,the optimal problem of designing linear-phase 2-D FIR digital filters is turned to approximate the desired 2-D magnitude response by using the compact express.To solve the problem,a new NNA is presented based on minimizing the mean-squared error,and the convergence theorem is presented and proved to ensure the designed 2-D filter stable.Three design examples are also given to illustrate the effectiveness of the NNA-based design approach.
Parallel stitching of 2D materials
Ling, Xi; Wu, Lijun; Lin, Yuxuan; Ma, Qiong; Wang, Ziqiang; Song, Yi; Yu, Lili; Huang, Shengxi; Fang, Wenjing; Zhang, Xu; et al
2016-01-27
Diverse parallel stitched 2D heterostructures, including metal–semiconductor, semiconductor–semiconductor, and insulator–semiconductor, are synthesized directly through selective “sowing” of aromatic molecules as the seeds in the chemical vapor deposition (CVD) method. Lastly, the methodology enables the large-scale fabrication of lateral heterostructures, which offers tremendous potential for its application in integrated circuits.
Parallel Stitching of 2D Materials.
Ling, Xi; Lin, Yuxuan; Ma, Qiong; Wang, Ziqiang; Song, Yi; Yu, Lili; Huang, Shengxi; Fang, Wenjing; Zhang, Xu; Hsu, Allen L; Bie, Yaqing; Lee, Yi-Hsien; Zhu, Yimei; Wu, Lijun; Li, Ju; Jarillo-Herrero, Pablo; Dresselhaus, Mildred; Palacios, Tomás; Kong, Jing
2016-03-01
Diverse parallel stitched 2D heterostructures, including metal-semiconductor, semiconductor-semiconductor, and insulator-semiconductor, are synthesized directly through selective "sowing" of aromatic molecules as the seeds in the chemical vapor deposition (CVD) method. The methodology enables the large-scale fabrication of lateral heterostructures, which offers tremendous potential for its application in integrated circuits. PMID:26813882
Baby universes in 2d quantum gravity
NASA Astrophysics Data System (ADS)
Ambjørn, Jan; Jain, Sanjay; Thorleifsson, Gudmar
1993-06-01
We investigate the fractal structure of 2d quantum gravity, both for pure gravity and for gravity coupled to multiple gaussian fields and for gravity coupled to Ising spins. The roughness of the surfaces is described in terms of baby universes and using numerical simulations we measure their distribution which is related to the string susceptibility exponent γstring.
NASA Astrophysics Data System (ADS)
Di Fiore, V.; Cavuoto, G.; Tarallo, D.; Punzo, M.; Evangelista, L.
2016-05-01
A joint analysis of down-hole (DH) and multichannel analysis of surface waves (MASW) measurements offers a complete evaluation of shear wave velocity profiles, especially for sites where a strong lateral variability is expected, such as archeological sites. In this complex stratigraphic setting, the high "subsoil anisotropy" (i.e., sharp lithological changes due to the presence of anthropogenic backfill deposits and/or buried man-made structures) implies a different role for DH and MASW tests. This paper discusses some results of a broad experimental program conducted on the Palatine Hill, one of the most ancient areas of the city of Rome (Italy). The experiments were part of a project on seismic microzoning and consisted of 20 MASW and 11 DH tests. The main objective of this study was to examine the difficulties related to the interpretation of the DH and MASW tests and the reliability limits inherent in the application of the noninvasive method in complex stratigraphic settings. As is well known, DH tests provide good determinations of shear wave velocities (Vs) for different lithologies and man-made materials, whereas MASW tests provide average values for the subsoil volume investigated. The data obtained from each method with blind tests were compared and were correlated to site-specific subsurface conditions, including lateral variability. Differences between punctual (DH) and global (MASW) Vs measurements are discussed, quantifying the errors by synthetic comparison and by site response analyses. This study demonstrates that, for archeological sites, VS profiles obtained from the DH and MASW methods differ by more than 15 %. However, the local site effect showed comparable results in terms of natural frequencies, whereas the resolution of the inverted shear wave velocity was influenced by the fundamental mode of propagation.
NASA High-Speed 2D Photogrammetric Measurement System
NASA Technical Reports Server (NTRS)
Dismond, Harriett R.
2012-01-01
The object of this report is to provide users of the NASA high-speed 2D photogrammetric measurement system with procedures required to obtain drop-model trajectory and impact data for full-scale and sub-scale models. This guide focuses on use of the system for vertical drop testing at the NASA Langley Landing and Impact Research (LandIR) Facility.
Combination Of Thermography And Pressure Tests To Combat Air Leakage Problems In Building Enclosures
NASA Astrophysics Data System (ADS)
Spruin, W. G.
1987-05-01
Uncontrolled air leakage in a building enclosure is the main component of space heating and cooling costs. In Atlantic Canada, Public Works Canada has combined thermography and pressure testing to identify design and construction problems in new construction and to identify specific areas of air leakage in existing housing stock. A study case shows how thermography and pressure testing has been utilized to locate and compare specific areas of air leakage in a residence before and after air sealing. The study provides both quantitative and qualitative evidence of how air sealing increases the air tightness in building enclosures.
NASA Astrophysics Data System (ADS)
Bolotina, I.; Bulavinov, A.; Pinchuk, R.; Salchak, Y.
2016-04-01
The paper considers the problems of ultrasonic nondestructive testing of products intended for mechanical engineering. The functional and electronic circuits of an ultrasonic tomograph are presented. The function of signal radiation from the clocked multielement apparatus is described, the cross-functional flowchart of the prototype of a US tomograph is considered. The development trends of ultrasonic tomography for near-term outlook are demonstrated.
The transonic Reynolds number problem. [limitations of transonic aerodynamic test facilities
NASA Technical Reports Server (NTRS)
Jones, J. L.
1977-01-01
Problems in modeling the complex interacting flow fields in the transonic speed regime are reviewed. The limitations of wind tunnel test capabilities are identified, and options for resolving the deficiency are examined. The evolution of the National Transonic Facility, and the various needs for research investigations to be done there are discussed. The relative priorities that should be given within and across subdisciplines for guidance in planning for the most effective use of the facility are considered.
NASA Astrophysics Data System (ADS)
Rowley-Neale, Samuel J.; Fearn, Jamie M.; Brownson, Dale A. C.; Smith, Graham C.; Ji, Xiaobo; Banks, Craig E.
2016-08-01
Two-dimensional molybdenum disulphide nanosheets (2D-MoS2) have proven to be an effective electrocatalyst, with particular attention being focused on their use towards increasing the efficiency of the reactions associated with hydrogen fuel cells. Whilst the majority of research has focused on the Hydrogen Evolution Reaction (HER), herein we explore the use of 2D-MoS2 as a potential electrocatalyst for the much less researched Oxygen Reduction Reaction (ORR). We stray from literature conventions and perform experiments in 0.1 M H2SO4 acidic electrolyte for the first time, evaluating the electrochemical performance of the ORR with 2D-MoS2 electrically wired/immobilised upon several carbon based electrodes (namely; Boron Doped Diamond (BDD), Edge Plane Pyrolytic Graphite (EPPG), Glassy Carbon (GC) and Screen-Printed Electrodes (SPE)) whilst exploring a range of 2D-MoS2 coverages/masses. Consequently, the findings of this study are highly applicable to real world fuel cell applications. We show that significant improvements in ORR activity can be achieved through the careful selection of the underlying/supporting carbon materials that electrically wire the 2D-MoS2 and utilisation of an optimal mass of 2D-MoS2. The ORR onset is observed to be reduced to ca. +0.10 V for EPPG, GC and SPEs at 2D-MoS2 (1524 ng cm-2 modification), which is far closer to Pt at +0.46 V compared to bare/unmodified EPPG, GC and SPE counterparts. This report is the first to demonstrate such beneficial electrochemical responses in acidic conditions using a 2D-MoS2 based electrocatalyst material on a carbon-based substrate (SPEs in this case). Investigation of the beneficial reaction mechanism reveals the ORR to occur via a 4 electron process in specific conditions; elsewhere a 2 electron process is observed. This work offers valuable insights for those wishing to design, fabricate and/or electrochemically test 2D-nanosheet materials towards the ORR.Two-dimensional molybdenum disulphide nanosheets
Application of 2D Non-Graphene Materials and 2D Oxide Nanostructures for Biosensing Technology
Shavanova, Kateryna; Bakakina, Yulia; Burkova, Inna; Shtepliuk, Ivan; Viter, Roman; Ubelis, Arnolds; Beni, Valerio; Starodub, Nickolaj; Yakimova, Rositsa; Khranovskyy, Volodymyr
2016-01-01
The discovery of graphene and its unique properties has inspired researchers to try to invent other two-dimensional (2D) materials. After considerable research effort, a distinct “beyond graphene” domain has been established, comprising the library of non-graphene 2D materials. It is significant that some 2D non-graphene materials possess solid advantages over their predecessor, such as having a direct band gap, and therefore are highly promising for a number of applications. These applications are not limited to nano- and opto-electronics, but have a strong potential in biosensing technologies, as one example. However, since most of the 2D non-graphene materials have been newly discovered, most of the research efforts are concentrated on material synthesis and the investigation of the properties of the material. Applications of 2D non-graphene materials are still at the embryonic stage, and the integration of 2D non-graphene materials into devices is scarcely reported. However, in recent years, numerous reports have blossomed about 2D material-based biosensors, evidencing the growing potential of 2D non-graphene materials for biosensing applications. This review highlights the recent progress in research on the potential of using 2D non-graphene materials and similar oxide nanostructures for different types of biosensors (optical and electrochemical). A wide range of biological targets, such as glucose, dopamine, cortisol, DNA, IgG, bisphenol, ascorbic acid, cytochrome and estradiol, has been reported to be successfully detected by biosensors with transducers made of 2D non-graphene materials. PMID:26861346
Application of 2D Non-Graphene Materials and 2D Oxide Nanostructures for Biosensing Technology.
Shavanova, Kateryna; Bakakina, Yulia; Burkova, Inna; Shtepliuk, Ivan; Viter, Roman; Ubelis, Arnolds; Beni, Valerio; Starodub, Nickolaj; Yakimova, Rositsa; Khranovskyy, Volodymyr
2016-01-01
The discovery of graphene and its unique properties has inspired researchers to try to invent other two-dimensional (2D) materials. After considerable research effort, a distinct "beyond graphene" domain has been established, comprising the library of non-graphene 2D materials. It is significant that some 2D non-graphene materials possess solid advantages over their predecessor, such as having a direct band gap, and therefore are highly promising for a number of applications. These applications are not limited to nano- and opto-electronics, but have a strong potential in biosensing technologies, as one example. However, since most of the 2D non-graphene materials have been newly discovered, most of the research efforts are concentrated on material synthesis and the investigation of the properties of the material. Applications of 2D non-graphene materials are still at the embryonic stage, and the integration of 2D non-graphene materials into devices is scarcely reported. However, in recent years, numerous reports have blossomed about 2D material-based biosensors, evidencing the growing potential of 2D non-graphene materials for biosensing applications. This review highlights the recent progress in research on the potential of using 2D non-graphene materials and similar oxide nanostructures for different types of biosensors (optical and electrochemical). A wide range of biological targets, such as glucose, dopamine, cortisol, DNA, IgG, bisphenol, ascorbic acid, cytochrome and estradiol, has been reported to be successfully detected by biosensors with transducers made of 2D non-graphene materials. PMID:26861346
A scanning-mode 2D shear wave imaging (s2D-SWI) system for ultrasound elastography.
Qiu, Weibao; Wang, Congzhi; Li, Yongchuan; Zhou, Juan; Yang, Ge; Xiao, Yang; Feng, Ge; Jin, Qiaofeng; Mu, Peitian; Qian, Ming; Zheng, Hairong
2015-09-01
Ultrasound elastography is widely used for the non-invasive measurement of tissue elasticity properties. Shear wave imaging (SWI) is a quantitative method for assessing tissue stiffness. SWI has been demonstrated to be less operator dependent than quasi-static elastography, and has the ability to acquire quantitative elasticity information in contrast with acoustic radiation force impulse (ARFI) imaging. However, traditional SWI implementations cannot acquire two dimensional (2D) quantitative images of the tissue elasticity distribution. This study proposes and evaluates a scanning-mode 2D SWI (s2D-SWI) system. The hardware and image processing algorithms are presented in detail. Programmable devices are used to support flexible control of the system and the image processing algorithms. An analytic signal based cross-correlation method and a Radon transformation based shear wave speed determination method are proposed, which can be implemented using parallel computation. Imaging of tissue mimicking phantoms, and in vitro, and in vivo imaging test are conducted to demonstrate the performance of the proposed system. The s2D-SWI system represents a new choice for the quantitative mapping of tissue elasticity, and has great potential for implementation in commercial ultrasound scanners. PMID:26025508
McDermott, K B; Roediger, H L
1996-03-01
Three experiments examined whether a conceptual implicit memory test (specifically, category instance generation) would exhibit repetition effects similar to those found in free recall. The transfer appropriate processing account of dissociations among memory tests led us to predict that the tests would show parallel effects; this prediction was based upon the theory's assumption that conceptual tests will behave similarly as a function of various independent variables. In Experiment 1, conceptual repetition (i.e., following a target word [e.g., puzzles] with an associate [e.g., jigsaw]) did not enhance priming on the instance generation test relative to the condition of simply presenting the target word once, although this manipulation did affect free recall. In Experiment 2, conceptual repetition was achieved by following a picture with its corresponding word (or vice versa). In this case, there was an effect of conceptual repetition on free recall but no reliable effect on category instance generation or category cued recall. In addition, we obtained a picture superiority effect in free recall but not in category instance generation. In the third experiment, when the same study sequence was used as in Experiment 1, but with instructions that encouraged relational processing, priming on the category instance generation task was enhanced by conceptual repetition. Results demonstrate that conceptual memory tests can be dissociated and present problems for Roediger's (1990) transfer appropriate processing account of dissociations between explicit and implicit tests. PMID:8653098
A new algorithm for generating highly accurate benchmark solutions to transport test problems
Azmy, Y.Y.
1997-06-01
We present a new algorithm for solving the neutron transport equation in its discrete-variable form. The new algorithm is based on computing the full matrix relating the scalar flux spatial moments in all cells to the fixed neutron source spatial moments, foregoing the need to compute the angular flux spatial moments, and thereby eliminating the need for sweeping the spatial mesh in each discrete-angular direction. The matrix equation is solved exactly in test cases, producing a solution vector that is free from iteration convergence error, and subject only to truncation and roundoff errors. Our algorithm is designed to provide method developers with a quick and simple solution scheme to test their new methods on difficult test problems without the need to develop sophisticated solution techniques, e.g. acceleration, before establishing the worthiness of their innovation. We demonstrate the utility of the new algorithm by applying it to the Arbitrarily High Order Transport Nodal (AHOT-N) method, and using it to solve two of Burre`s Suite of Test Problems (BSTP). Our results provide highly accurate benchmark solutions, that can be distributed electronically and used to verify the pointwise accuracy of other solution methods and algorithms.
Static & Dynamic Response of 2D Solids
1996-07-15
NIKE2D is an implicit finite-element code for analyzing the finite deformation, static and dynamic response of two-dimensional, axisymmetric, plane strain, and plane stress solids. The code is fully vectorized and available on several computing platforms. A number of material models are incorporated to simulate a wide range of material behavior including elasto-placicity, anisotropy, creep, thermal effects, and rate dependence. Slideline algorithms model gaps and sliding along material interfaces, including interface friction, penetration and single surfacemore » contact. Interactive-graphics and rezoning is included for analyses with large mesh distortions. In addition to quasi-Newton and arc-length procedures, adaptive algorithms can be defined to solve the implicit equations using the solution language ISLAND. Each of these capabilities and more make NIKE2D a robust analysis tool.« less
Stochastic Inversion of 2D Magnetotelluric Data
2010-07-01
The algorithm is developed to invert 2D magnetotelluric (MT) data based on sharp boundary parametrization using a Bayesian framework. Within the algorithm, we consider the locations and the resistivity of regions formed by the interfaces are as unknowns. We use a parallel, adaptive finite-element algorithm to forward simulate frequency-domain MT responses of 2D conductivity structure. Those unknown parameters are spatially correlated and are described by a geostatistical model. The joint posterior probability distribution function ismore » explored by Markov Chain Monte Carlo (MCMC) sampling methods. The developed stochastic model is effective for estimating the interface locations and resistivity. Most importantly, it provides details uncertainty information on each unknown parameter. Hardware requirements: PC, Supercomputer, Multi-platform, Workstation; Software requirements C and Fortan; Operation Systems/version is Linux/Unix or Windows« less
Stochastic Inversion of 2D Magnetotelluric Data
Chen, Jinsong
2010-07-01
The algorithm is developed to invert 2D magnetotelluric (MT) data based on sharp boundary parametrization using a Bayesian framework. Within the algorithm, we consider the locations and the resistivity of regions formed by the interfaces are as unknowns. We use a parallel, adaptive finite-element algorithm to forward simulate frequency-domain MT responses of 2D conductivity structure. Those unknown parameters are spatially correlated and are described by a geostatistical model. The joint posterior probability distribution function is explored by Markov Chain Monte Carlo (MCMC) sampling methods. The developed stochastic model is effective for estimating the interface locations and resistivity. Most importantly, it provides details uncertainty information on each unknown parameter. Hardware requirements: PC, Supercomputer, Multi-platform, Workstation; Software requirements C and Fortan; Operation Systems/version is Linux/Unix or Windows
Explicit 2-D Hydrodynamic FEM Program
1996-08-07
DYNA2D* is a vectorized, explicit, two-dimensional, axisymmetric and plane strain finite element program for analyzing the large deformation dynamic and hydrodynamic response of inelastic solids. DYNA2D* contains 13 material models and 9 equations of state (EOS) to cover a wide range of material behavior. The material models implemented in all machine versions are: elastic, orthotropic elastic, kinematic/isotropic elastic plasticity, thermoelastoplastic, soil and crushable foam, linear viscoelastic, rubber, high explosive burn, isotropic elastic-plastic, temperature-dependent elastic-plastic. Themore » isotropic and temperature-dependent elastic-plastic models determine only the deviatoric stresses. Pressure is determined by one of 9 equations of state including linear polynomial, JWL high explosive, Sack Tuesday high explosive, Gruneisen, ratio of polynomials, linear polynomial with energy deposition, ignition and growth of reaction in HE, tabulated compaction, and tabulated.« less
Schottky diodes from 2D germanane
NASA Astrophysics Data System (ADS)
Sahoo, Nanda Gopal; Esteves, Richard J.; Punetha, Vinay Deep; Pestov, Dmitry; Arachchige, Indika U.; McLeskey, James T.
2016-07-01
We report on the fabrication and characterization of a Schottky diode made using 2D germanane (hydrogenated germanene). When compared to germanium, the 2D structure has higher electron mobility, an optimal band-gap, and exceptional stability making germanane an outstanding candidate for a variety of opto-electronic devices. One-atom-thick sheets of hydrogenated puckered germanium atoms have been synthesized from a CaGe2 framework via intercalation and characterized by XRD, Raman, and FTIR techniques. The material was then used to fabricate Schottky diodes by suspending the germanane in benzonitrile and drop-casting it onto interdigitated metal electrodes. The devices demonstrate significant rectifying behavior and the outstanding potential of this material.
TOPAZ2D validation status report, August 1990
Davis, B.
1990-08-01
Analytic solutions to two heat transfer problems were used to partially evaluate the performance TOPAZ, and LLNL finite element heat transfer code. The two benchmark analytic solutions were for: 2D steady state slab, with constant properties, constant uniform temperature boundary conditions on three sides, and constant temperature distribution according to a sine function on the fourth side; 1D transient non-linear, with temperature dependent conductivity and specific heat (varying such that the thermal diffusivity remained constant), constant heat flux on the front face and adiabatic conditions on the other face. The TOPAZ solution converged to the analytic solution in both the transient and the steady state problem. Consistent mass matrix type of analysis yielded best performance for the transient problem, in the late-time response; but notable unnatural anomalies were observed in the early-time temperature response at nodal locations near the front face. 5 refs., 22 figs.
Layer Engineering of 2D Semiconductor Junctions.
He, Yongmin; Sobhani, Ali; Lei, Sidong; Zhang, Zhuhua; Gong, Yongji; Jin, Zehua; Zhou, Wu; Yang, Yingchao; Zhang, Yuan; Wang, Xifan; Yakobson, Boris; Vajtai, Robert; Halas, Naomi J; Li, Bo; Xie, Erqing; Ajayan, Pulickel
2016-07-01
A new concept for junction fabrication by connecting multiple regions with varying layer thicknesses, based on the thickness dependence, is demonstrated. This type of junction is only possible in super-thin-layered 2D materials, and exhibits similar characteristics as p-n junctions. Rectification and photovoltaic effects are observed in chemically homogeneous MoSe2 junctions between domains of different thicknesses. PMID:27136275
NASA Astrophysics Data System (ADS)
Smith, Greg; Lankshear, Allan
1998-07-01
2dF is a multi-object instrument mounted at prime focus at the AAT capable of spectroscopic analysis of 400 objects in a single 2 degree field. It also prepares a second 2 degree 400 object field while the first field is being observed. At its heart is a high precision robotic positioner that places individual fiber end magnetic buttons on one of two field plates. The button gripper is carried on orthogonal gantries powered by linear synchronous motors and contains a TV camera which precisely locates backlit buttons to allow placement in user defined locations to 10 (mu) accuracy. Fiducial points on both plates can also be observed by the camera to allow repeated checks on positioning accuracy. Field plates rotate to follow apparent sky rotation. The spectrographs both analyze light from the 200 observing fibers each and back- illuminate the 400 fibers being re-positioned during the observing run. The 2dF fiber position and spectrograph system is a large and complex instrument located at the prime focus of the Anglo Australian Telescope. The mechanical design has departed somewhat from the earlier concepts of Gray et al, but still reflects the audacity of those first ideas. The positioner is capable of positioning 400 fibers on a field plate while another 400 fibers on another plate are observing at the focus of the telescope and feeding the twin spectrographs. When first proposed it must have seemed like ingenuity unfettered by caution. Yet now it works, and works wonderfully well. 2dF is a system which functions as the result of the combined and coordinated efforts of the astronomers, the mechanical designers and tradespeople, the electronic designers, the programmers, the support staff at the telescope, and the manufacturing subcontractors. The mechanical design of the 2dF positioner and spectrographs was carried out by the mechanical engineering staff of the AAO and the majority of the manufacture was carried out in the AAO workshops.
Realistic and efficient 2D crack simulation
NASA Astrophysics Data System (ADS)
Yadegar, Jacob; Liu, Xiaoqing; Singh, Abhishek
2010-04-01
Although numerical algorithms for 2D crack simulation have been studied in Modeling and Simulation (M&S) and computer graphics for decades, realism and computational efficiency are still major challenges. In this paper, we introduce a high-fidelity, scalable, adaptive and efficient/runtime 2D crack/fracture simulation system by applying the mathematically elegant Peano-Cesaro triangular meshing/remeshing technique to model the generation of shards/fragments. The recursive fractal sweep associated with the Peano-Cesaro triangulation provides efficient local multi-resolution refinement to any level-of-detail. The generated binary decomposition tree also provides efficient neighbor retrieval mechanism used for mesh element splitting and merging with minimal memory requirements essential for realistic 2D fragment formation. Upon load impact/contact/penetration, a number of factors including impact angle, impact energy, and material properties are all taken into account to produce the criteria of crack initialization, propagation, and termination leading to realistic fractal-like rubble/fragments formation. The aforementioned parameters are used as variables of probabilistic models of cracks/shards formation, making the proposed solution highly adaptive by allowing machine learning mechanisms learn the optimal values for the variables/parameters based on prior benchmark data generated by off-line physics based simulation solutions that produce accurate fractures/shards though at highly non-real time paste. Crack/fracture simulation has been conducted on various load impacts with different initial locations at various impulse scales. The simulation results demonstrate that the proposed system has the capability to realistically and efficiently simulate 2D crack phenomena (such as window shattering and shards generation) with diverse potentials in military and civil M&S applications such as training and mission planning.
Compact 2-D graphical representation of DNA
NASA Astrophysics Data System (ADS)
Randić, Milan; Vračko, Marjan; Zupan, Jure; Novič, Marjana
2003-05-01
We present a novel 2-D graphical representation for DNA sequences which has an important advantage over the existing graphical representations of DNA in being very compact. It is based on: (1) use of binary labels for the four nucleic acid bases, and (2) use of the 'worm' curve as template on which binary codes are placed. The approach is illustrated on DNA sequences of the first exon of human β-globin and gorilla β-globin.
2D materials: Graphene and others
NASA Astrophysics Data System (ADS)
Bansal, Suneev Anil; Singh, Amrinder Pal; Kumar, Suresh
2016-05-01
Present report reviews the recent advancements in new atomically thick 2D materials. Materials covered in this review are Graphene, Silicene, Germanene, Boron Nitride (BN) and Transition metal chalcogenides (TMC). These materials show extraordinary mechanical, electronic and optical properties which make them suitable candidates for future applications. Apart from unique properties, tune-ability of highly desirable properties of these materials is also an important area to be emphasized on.
A HUPO test sample study reveals common problems in mass spectrometry-based proteomics.
Bell, Alexander W; Deutsch, Eric W; Au, Catherine E; Kearney, Robert E; Beavis, Ron; Sechi, Salvatore; Nilsson, Tommy; Bergeron, John J M
2009-06-01
We performed a test sample study to try to identify errors leading to irreproducibility, including incompleteness of peptide sampling, in liquid chromatography-mass spectrometry-based proteomics. We distributed an equimolar test sample, comprising 20 highly purified recombinant human proteins, to 27 laboratories. Each protein contained one or more unique tryptic peptides of 1,250 Da to test for ion selection and sampling in the mass spectrometer. Of the 27 labs, members of only 7 labs initially reported all 20 proteins correctly, and members of only 1 lab reported all tryptic peptides of 1,250 Da. Centralized analysis of the raw data, however, revealed that all 20 proteins and most of the 1,250 Da peptides had been detected in all 27 labs. Our centralized analysis determined missed identifications (false negatives), environmental contamination, database matching and curation of protein identifications as sources of problems. Improved search engines and databases are needed for mass spectrometry-based proteomics. PMID:19448641
Engineering light outcoupling in 2D materials.
Lien, Der-Hsien; Kang, Jeong Seuk; Amani, Matin; Chen, Kevin; Tosun, Mahmut; Wang, Hsin-Ping; Roy, Tania; Eggleston, Michael S; Wu, Ming C; Dubey, Madan; Lee, Si-Chen; He, Jr-Hau; Javey, Ali
2015-02-11
When light is incident on 2D transition metal dichalcogenides (TMDCs), it engages in multiple reflections within underlying substrates, producing interferences that lead to enhancement or attenuation of the incoming and outgoing strength of light. Here, we report a simple method to engineer the light outcoupling in semiconducting TMDCs by modulating their dielectric surroundings. We show that by modulating the thicknesses of underlying substrates and capping layers, the interference caused by substrate can significantly enhance the light absorption and emission of WSe2, resulting in a ∼11 times increase in Raman signal and a ∼30 times increase in the photoluminescence (PL) intensity of WSe2. On the basis of the interference model, we also propose a strategy to control the photonic and optoelectronic properties of thin-layer WSe2. This work demonstrates the utilization of outcoupling engineering in 2D materials and offers a new route toward the realization of novel optoelectronic devices, such as 2D LEDs and solar cells. PMID:25602462
Hunting down magnetic monopoles in 2D topological insulators?
NASA Astrophysics Data System (ADS)
He, Xugang; Cmpmsd At Bnl Team
Contrary to the existence of electric charge, magnetic monopole does not exist in nature. It is thus extraordinary to find that magnetic monopoles can be pictured conceptually in topological insulators. For 2D topological insulators, the topological invariant corresponds to the total flux of an effective magnetic field (the Berry curvature) over the reciprocal space. Upon wrapping the 2D reciprocal space into a compact manifold as a torus, the non-zero total flux can be considered to originate from magnetic monopoles with quantized charge. We will first illustrate the intrinsic difficulty via extending a 2D problem to a 3D reciprocal space, and then demonstrate that analytical continuation to the complex momentum space offers a natural solution in which 1) the magnetic monopoles emerge naturally in pairs each forming a string above and below the real axis possessing opposite charge, and 2) the total charge below the real axis gives exactly the topological invariant. In essence, the robustness of the topology is mapped to the robustness of the total charge in the lower complex plan, a mapping intriguing even mathematically. Finally, we will illustrate the evolution across the topological phase transition, providing a natural description of the metallic nature in the phase boundary, and offering a clear explanation why a change of global topology can be induced via a local change in reciprocal space. Work supported by US DOE BES DE-AC02-98CH10886.
The k-sample problem in a multi-state model and testing transition probability matrices.
Tattar, Prabhanjan N; Vaman, H J
2014-07-01
The choice of multi-state models is natural in analysis of survival data, e.g., when the subjects in a study pass through different states like 'healthy', 'in a state of remission', 'relapse' or 'dead' in a health related quality of life study. Competing risks is another common instance of the use of multi-state models. Statistical inference for such event history data can be carried out by assuming a stochastic process model. Under such a setting, comparison of the event history data generated by two different treatments calls for testing equality of the corresponding transition probability matrices. The present paper proposes solution to this class of problems by assuming a non-homogeneous Markov process to describe the transitions among the health states. A class of test statistics are derived for comparison of [Formula: see text] treatments by using a 'weight process'. This class, in particular, yields generalisations of the log-rank, Gehan, Peto-Peto and Harrington-Fleming tests. For an intrinsic comparison of the treatments, the 'leave-one-out' jackknife method is employed for identifying influential observations. The proposed methods are then used to develop the Kolmogorov-Smirnov type supremum tests corresponding to the various extended tests. To demonstrate the usefulness of the test procedures developed, a simulation study was carried out and an application to the Trial V data provided by International Breast Cancer Study Group is discussed. PMID:23722306
Recent update of the RPLUS2D/3D codes
NASA Technical Reports Server (NTRS)
Tsai, Y.-L. Peter
1991-01-01
The development of the RPLUS2D/3D codes is summarized. These codes utilize LU algorithms to solve chemical non-equilibrium flows in a body-fitted coordinate system. The motivation behind the development of these codes is the need to numerically predict chemical non-equilibrium flows for the National AeroSpace Plane Program. Recent improvements include vectorization method, blocking algorithms for geometric flexibility, out-of-core storage for large-size problems, and an LU-SW/UP combination for CPU-time efficiency and solution quality.
2D superconductivity by ionic gating
NASA Astrophysics Data System (ADS)
Iwasa, Yoshi
2D superconductivity is attracting a renewed interest due to the discoveries of new highly crystalline 2D superconductors in the past decade. Superconductivity at the oxide interfaces triggered by LaAlO3/SrTiO3 has become one of the promising routes for creation of new 2D superconductors. Also, the MBE grown metallic monolayers including FeSe are also offering a new platform of 2D superconductors. In the last two years, there appear a variety of monolayer/bilayer superconductors fabricated by CVD or mechanical exfoliation. Among these, electric field induced superconductivity by electric double layer transistor (EDLT) is a unique platform of 2D superconductivity, because of its ability of high density charge accumulation, and also because of the versatility in terms of materials, stemming from oxides to organics and layered chalcogenides. In this presentation, the following issues of electric filed induced superconductivity will be addressed; (1) Tunable carrier density, (2) Weak pinning, (3) Absence of inversion symmetry. (1) Since the sheet carrier density is quasi-continuously tunable from 0 to the order of 1014 cm-2, one is able to establish an electronic phase diagram of superconductivity, which will be compared with that of bulk superconductors. (2) The thickness of superconductivity can be estimated as 2 - 10 nm, dependent on materials, and is much smaller than the in-plane coherence length. Such a thin but low resistance at normal state results in extremely weak pinning beyond the dirty Boson model in the amorphous metallic films. (3) Due to the electric filed, the inversion symmetry is inherently broken in EDLT. This feature appears in the enhancement of Pauli limit of the upper critical field for the in-plane magnetic fields. In transition metal dichalcogenide with a substantial spin-orbit interactions, we were able to confirm the stabilization of Cooper pair due to its spin-valley locking. This work has been supported by Grant-in-Aid for Specially
Interpretation of Magnetic Phase Anomalies over 2D Tabular Bodies
NASA Astrophysics Data System (ADS)
Subrahmanyam, M.
2016-05-01
In this study, phase angle (inverse tangent of the ratio of the horizontal to vertical gradients of magnetic anomalies) profile over two-dimensional tabular bodies has been subjected to detailed analysis for determining the source parameters. Distances between certain characteristic positions on this phase curve are related to the parameters of two-dimensional tabular magnetic sources. In this paper, I have derived the mathematical expressions for these relations. It has been demonstrated here that for locating the origin of the 2D tabular source, knowledge on the type of the model (contact, sheet, dyke, and fault) is not necessary. A procedure is evolved to determine the location, depth, width and magnetization angle of the 2D sources from the mathematical expressions. The method is tested on real field data. The effect of the overlapping bodies is also discussed with two synthetic examples. The interpretation technique is developed for contact, sheet, dike and inclined fault bodies.
FPCAS2D user's guide, version 1.0
NASA Astrophysics Data System (ADS)
Bakhle, Milind A.
1994-12-01
The FPCAS2D computer code has been developed for aeroelastic stability analysis of bladed disks such as those in fans, compressors, turbines, propellers, or propfans. The aerodynamic analysis used in this code is based on the unsteady two-dimensional full potential equation which is solved for a cascade of blades. The structural analysis is based on a two degree-of-freedom rigid typical section model for each blade. Detailed explanations of the aerodynamic analysis, the numerical algorithms, and the aeroelastic analysis are not given in this report. This guide can be used to assist in the preparation of the input data required by the FPCAS2D code. A complete description of the input data is provided in this report. In addition, four test cases, including inputs and outputs, are provided.
TOPAZ2D heat transfer code users manual and thermal property data base
NASA Astrophysics Data System (ADS)
Shapiro, A. B.; Edwards, A. L.
1990-05-01
TOPAZ2D is a two dimensional implicit finite element computer code for heat transfer analysis. This user's manual provides information on the structure of a TOPAZ2D input file. Also included is a material thermal property data base. This manual is supplemented with The TOPAZ2D Theoretical Manual and the TOPAZ2D Verification Manual. TOPAZ2D has been implemented on the CRAY, SUN, and VAX computers. TOPAZ2D can be used to solve for the steady state or transient temperature field on two dimensional planar or axisymmetric geometries. Material properties may be temperature dependent and either isotropic or orthotropic. A variety of time and temperature dependent boundary conditions can be specified including temperature, flux, convection, and radiation. Time or temperature dependent internal heat generation can be defined locally be element or globally by material. TOPAZ2D can solve problems of diffuse and specular band radiation in an enclosure coupled with conduction in material surrounding the enclosure. Additional features include thermally controlled reactive chemical mixtures, thermal contact resistance across an interface, bulk fluid flow, phase change, and energy balances. Thermal stresses can be calculated using the solid mechanics code NIKE2D which reads the temperature state data calculated by TOPAZ2D. A three dimensional version of the code, TOPAZ3D is available.
NASA Technical Reports Server (NTRS)
Antoniewicz, Robert F.; Duke, Eugene L.; Menon, P. K. A.
1991-01-01
The design of nonlinear controllers has relied on the use of detailed aerodynamic and engine models that must be associated with the control law in the flight system implementation. Many of these controllers were applied to vehicle flight path control problems and have attempted to combine both inner- and outer-loop control functions in a single controller. An approach to the nonlinear trajectory control problem is presented. This approach uses linearizing transformations with measurement feedback to eliminate the need for detailed aircraft models in outer-loop control applications. By applying this approach and separating the inner-loop and outer-loop functions two things were achieved: (1) the need for incorporating detailed aerodynamic models in the controller is obviated; and (2) the controller is more easily incorporated into existing aircraft flight control systems. An implementation of the controller is discussed, and this controller is tested on a six degree-of-freedom F-15 simulation and in flight on an F-15 aircraft. Simulation data are presented which validates this approach over a large portion of the F-15 flight envelope. Proof of this concept is provided by flight-test data that closely matches simulation results. Flight-test data are also presented.
ERIC Educational Resources Information Center
Bakker, Martin P.; Ormel, Johan; Verhulst, Frank C.; Oldehinkel, Albertine J.
2012-01-01
This study tested whether childhood family instability is associated with mental health problems during adolescence through continued family instability and/or through a preadolescent onset of mental health problems. This test use data from a prospective population cohort of 2,230 Dutch adolescents ("M" age = 11.09, "SD" = 0.56 at the initial…
FVCOM validation experiments: Comparisons with ROMS for three idealized barotropic test problems
NASA Astrophysics Data System (ADS)
Huang, Haosheng; Chen, Changsheng; Cowles, Geoffrey W.; Winant, Clinton D.; Beardsley, Robert C.; Hedstrom, Kate S.; Haidvogel, Dale B.
2008-07-01
The unstructured-grid Finite-Volume Coastal Ocean Model (FVCOM) is evaluated using three idealized benchmark test problems: the Rossby equatorial soliton, the hydraulic jump, and the three-dimensional barotropic wind-driven basin. These test cases examine the properties of numerical dispersion and damping, the performance of the nonlinear advection scheme for supercritical flow conditions, and the accuracy of the implicit vertical viscosity scheme in barotropic settings, respectively. It is demonstrated that FVCOM provides overall a second-order spatial accuracy for the vertically averaged equations (i.e., external mode), and with increasing grid resolution the model-computed solutions show a fast convergence toward the analytic solutions regardless of the particular triangulation method. Examples are provided to illustrate the ability of FVCOM to facilitate local grid refinement and speed up computation. Comparisons are also made between FVCOM and the structured-grid Regional Ocean Modeling System (ROMS) for these test cases. For the linear problem in a simple rectangular domain, i.e., the wind-driven basin case, the performance of the two models is quite similar. For the nonlinear case, such as the Rossby equatorial soliton, the second-order advection scheme used in FVCOM is almost as accurate as the fourth-order advection scheme implemented in ROMS if the horizontal resolution is relatively high. FVCOM has taken advantage of the new development in computational fluid dynamics in resolving flow problems containing discontinuities. One salient feature illustrated by the three-dimensional barotropic wind-driven basin case is that FVCOM and ROMS simulations show different responses to the refinement of grid size in the horizontal and in the vertical.
SNARK09 - a software package for reconstruction of 2D images from 1D projections.
Klukowska, Joanna; Davidi, Ran; Herman, Gabor T
2013-06-01
The problem of reconstruction of slices and volumes from 1D and 2D projections has arisen in a large number of scientific fields (including computerized tomography, electron microscopy, X-ray microscopy, radiology, radio astronomy and holography). Many different methods (algorithms) have been suggested for its solution. In this paper we present a software package, SNARK09, for reconstruction of 2D images from their 1D projections. In the area of image reconstruction, researchers often desire to compare two or more reconstruction techniques and assess their relative merits. SNARK09 provides a uniform framework to implement algorithms and evaluate their performance. It has been designed to treat both parallel and divergent projection geometries and can either create test data (with or without noise) for use by reconstruction algorithms or use data collected by another software or a physical device. A number of frequently-used classical reconstruction algorithms are incorporated. The package provides a means for easy incorporation of new algorithms for their testing, comparison and evaluation. It comes with tools for statistical analysis of the results and ten worked examples. PMID:23414602
GBL-2D Version 1.0: a 2D geometry boolean library.
McBride, Cory L. (Elemental Technologies, American Fort, UT); Schmidt, Rodney Cannon; Yarberry, Victor R.; Meyers, Ray J.
2006-11-01
This report describes version 1.0 of GBL-2D, a geometric Boolean library for 2D objects. The library is written in C++ and consists of a set of classes and routines. The classes primarily represent geometric data and relationships. Classes are provided for 2D points, lines, arcs, edge uses, loops, surfaces and mask sets. The routines contain algorithms for geometric Boolean operations and utility functions. Routines are provided that incorporate the Boolean operations: Union(OR), XOR, Intersection and Difference. A variety of additional analytical geometry routines and routines for importing and exporting the data in various file formats are also provided. The GBL-2D library was originally developed as a geometric modeling engine for use with a separate software tool, called SummitView [1], that manipulates the 2D mask sets created by designers of Micro-Electro-Mechanical Systems (MEMS). However, many other practical applications for this type of software can be envisioned because the need to perform 2D Boolean operations can arise in many contexts.
ERIC Educational Resources Information Center
Almquist, Alan J.; Cronin, John E.
This is one of several study guides on contemporary problems produced by the American Association for the Advancement of Science with support of the National Science Foundation. This guide focuses on the origin of man. Part I, The Biochemical Evidence for Human Evolution, contains four sections: (1) Introduction; (2) Macromolecular Data; (3)…
Periodically sheared 2D Yukawa systems
Kovács, Anikó Zsuzsa; Hartmann, Peter; Donkó, Zoltán
2015-10-15
We present non-equilibrium molecular dynamics simulation studies on the dynamic (complex) shear viscosity of a 2D Yukawa system. We have identified a non-monotonic frequency dependence of the viscosity at high frequencies and shear rates, an energy absorption maximum (local resonance) at the Einstein frequency of the system at medium shear rates, an enhanced collective wave activity, when the excitation is near the plateau frequency of the longitudinal wave dispersion, and the emergence of significant configurational anisotropy at small frequencies and high shear rates.
ENERGY LANDSCAPE OF 2D FLUID FORMS
Y. JIANG; ET AL
2000-04-01
The equilibrium states of 2D non-coarsening fluid foams, which consist of bubbles with fixed areas, correspond to local minima of the total perimeter. (1) The authors find an approximate value of the global minimum, and determine directly from an image how far a foam is from its ground state. (2) For (small) area disorder, small bubbles tend to sort inwards and large bubbles outwards. (3) Topological charges of the same sign repel while charges of opposite sign attract. (4) They discuss boundary conditions and the uniqueness of the pattern for fixed topology.
2D stepping drive for hyperspectral systems
NASA Astrophysics Data System (ADS)
Endrödy, Csaba; Mehner, Hannes; Grewe, Adrian; Sinzinger, Stefan; Hoffmann, Martin
2015-07-01
We present the design, fabrication and characterization of a compact 2D stepping microdrive for pinhole array positioning. The miniaturized solution enables a highly integrated compact hyperspectral imaging system. Based on the geometry of the pinhole array, an inch-worm drive with electrostatic actuators was designed resulting in a compact (1 cm2) positioning system featuring a step size of about 15 µm in a 170 µm displacement range. The high payload (20 mg) as required for the pinhole array and the compact system design exceed the known electrostatic inch-worm-based microdrives.
Region-based Statistical Analysis of 2D PAGE Images
Li, Feng; Seillier-Moiseiwitsch, Françoise; Korostyshevskiy, Valeriy R.
2011-01-01
A new comprehensive procedure for statistical analysis of two-dimensional polyacrylamide gel electrophoresis (2D PAGE) images is proposed, including protein region quantification, normalization and statistical analysis. Protein regions are defined by the master watershed map that is obtained from the mean gel. By working with these protein regions, the approach bypasses the current bottleneck in the analysis of 2D PAGE images: it does not require spot matching. Background correction is implemented in each protein region by local segmentation. Two-dimensional locally weighted smoothing (LOESS) is proposed to remove any systematic bias after quantification of protein regions. Proteins are separated into mutually independent sets based on detected correlations, and a multivariate analysis is used on each set to detect the group effect. A strategy for multiple hypothesis testing based on this multivariate approach combined with the usual Benjamini-Hochberg FDR procedure is formulated and applied to the differential analysis of 2D PAGE images. Each step in the analytical protocol is shown by using an actual dataset. The effectiveness of the proposed methodology is shown using simulated gels in comparison with the commercial software packages PDQuest and Dymension. We also introduce a new procedure for simulating gel images. PMID:21850152
Code of Federal Regulations, 2010 CFR
2010-10-01
... 49 Transportation 1 2010-10-01 2010-10-01 false What problems always cause a drug test to be... TESTING PROGRAMS Problems in Drug Tests § 40.201 What problems always cause a drug test to be cancelled... laboratory reports that any of the following problems have occurred. You must inform the DER that the...
Code of Federal Regulations, 2011 CFR
2011-10-01
... 49 Transportation 1 2011-10-01 2011-10-01 false What problems always cause a drug test to be... TESTING PROGRAMS Problems in Drug Tests § 40.201 What problems always cause a drug test to be cancelled... laboratory reports that any of the following problems have occurred. You must inform the DER that the...
Gurev, Viatcheslav; Arens, Sander; Augustin, Christoph M.; Baron, Lukas; Blake, Robert; Bradley, Chris; Castro, Sebastian; Crozier, Andrew; Favino, Marco; Fastl, Thomas E.; Fritz, Thomas; Gao, Hao; Gizzi, Alessio; Griffith, Boyce E.; Hurtado, Daniel E.; Krause, Rolf; Luo, Xiaoyu; Nash, Martyn P.; Pezzuto, Simone; Plank, Gernot; Rossi, Simone; Ruprecht, Daniel; Seemann, Gunnar; Smith, Nicolas P.; Sundnes, Joakim; Rice, J. Jeremy; Trayanova, Natalia; Wang, Dafang; Jenny Wang, Zhinuo; Niederer, Steven A.
2015-01-01
Models of cardiac mechanics are increasingly used to investigate cardiac physiology. These models are characterized by a high level of complexity, including the particular anisotropic material properties of biological tissue and the actively contracting material. A large number of independent simulation codes have been developed, but a consistent way of verifying the accuracy and replicability of simulations is lacking. To aid in the verification of current and future cardiac mechanics solvers, this study provides three benchmark problems for cardiac mechanics. These benchmark problems test the ability to accurately simulate pressure-type forces that depend on the deformed objects geometry, anisotropic and spatially varying material properties similar to those seen in the left ventricle and active contractile forces. The benchmark was solved by 11 different groups to generate consensus solutions, with typical differences in higher-resolution solutions at approximately 0.5%, and consistent results between linear, quadratic and cubic finite elements as well as different approaches to simulating incompressible materials. Online tools and solutions are made available to allow these tests to be effectively used in verification of future cardiac mechanics software. PMID:26807042
Pre-test CFD Calculations for a Bypass Flow Standard Problem
Rich Johnson
2011-11-01
The bypass flow in a prismatic high temperature gas-cooled reactor (HTGR) is the flow that occurs between adjacent graphite blocks. Gaps exist between blocks due to variances in their manufacture and installation and because of the expansion and shrinkage of the blocks from heating and irradiation. Although the temperature of fuel compacts and graphite is sensitive to the presence of bypass flow, there is great uncertainty in the level and effects of the bypass flow. The Next Generation Nuclear Plant (NGNP) program at the Idaho National Laboratory has undertaken to produce experimental data of isothermal bypass flow between three adjacent graphite blocks. These data are intended to provide validation for computational fluid dynamic (CFD) analyses of the bypass flow. Such validation data sets are called Standard Problems in the nuclear safety analysis field. Details of the experimental apparatus as well as several pre-test calculations of the bypass flow are provided. Pre-test calculations are useful in examining the nature of the flow and to see if there are any problems associated with the flow and its measurement. The apparatus is designed to be able to provide three different gap widths in the vertical direction (the direction of the normal coolant flow) and two gap widths in the horizontal direction. It is expected that the vertical bypass flow will range from laminar to transitional to turbulent flow for the different gap widths that will be available.
WFR-2D: an analytical model for PWAS-generated 2D ultrasonic guided wave propagation
NASA Astrophysics Data System (ADS)
Shen, Yanfeng; Giurgiutiu, Victor
2014-03-01
This paper presents WaveFormRevealer 2-D (WFR-2D), an analytical predictive tool for the simulation of 2-D ultrasonic guided wave propagation and interaction with damage. The design of structural health monitoring (SHM) systems and self-aware smart structures requires the exploration of a wide range of parameters to achieve best detection and quantification of certain types of damage. Such need for parameter exploration on sensor dimension, location, guided wave characteristics (mode type, frequency, wavelength, etc.) can be best satisfied with analytical models which are fast and efficient. The analytical model was constructed based on the exact 2-D Lamb wave solution using Bessel and Hankel functions. Damage effects were inserted in the model by considering the damage as a secondary wave source with complex-valued directivity scattering coefficients containing both amplitude and phase information from wave-damage interaction. The analytical procedure was coded with MATLAB, and a predictive simulation tool called WaveFormRevealer 2-D was developed. The wave-damage interaction coefficients (WDICs) were extracted from harmonic analysis of local finite element model (FEM) with artificial non-reflective boundaries (NRB). The WFR-2D analytical simulation results were compared and verified with full scale multiphysics finite element models and experiments with scanning laser vibrometer. First, Lamb wave propagation in a pristine aluminum plate was simulated with WFR-2D, compared with finite element results, and verified by experiments. Then, an inhomogeneity was machined into the plate to represent damage. Analytical modeling was carried out, and verified by finite element simulation and experiments. This paper finishes with conclusions and suggestions for future work.
Fully automated 2D-3D registration and verification.
Varnavas, Andreas; Carrell, Tom; Penney, Graeme
2015-12-01
Clinical application of 2D-3D registration technology often requires a significant amount of human interaction during initialisation and result verification. This is one of the main barriers to more widespread clinical use of this technology. We propose novel techniques for automated initial pose estimation of the 3D data and verification of the registration result, and show how these techniques can be combined to enable fully automated 2D-3D registration, particularly in the case of a vertebra based system. The initialisation method is based on preoperative computation of 2D templates over a wide range of 3D poses. These templates are used to apply the Generalised Hough Transform to the intraoperative 2D image and the sought 3D pose is selected with the combined use of the generated accumulator arrays and a Gradient Difference Similarity Measure. On the verification side, two algorithms are proposed: one using normalised features based on the similarity value and the other based on the pose agreement between multiple vertebra based registrations. The proposed methods are employed here for CT to fluoroscopy registration and are trained and tested with data from 31 clinical procedures with 417 low dose, i.e. low quality, high noise interventional fluoroscopy images. When similarity value based verification is used, the fully automated system achieves a 95.73% correct registration rate, whereas a no registration result is produced for the remaining 4.27% of cases (i.e. incorrect registration rate is 0%). The system also automatically detects input images outside its operating range. PMID:26387052
Statistical analysis of quiet stance sway in 2-D
DiZio, Paul; Lackner, James R.
2014-01-01
Subjects exposed to a rotating environment that perturbs their postural sway show adaptive changes in their voluntary spatially directed postural motion to restore accurate movement paths but do not exhibit any obvious learning during passive stance. We have found, however, that a variable known to characterize the degree of stochasticity in quiet stance can also reveal subtle learning phenomena in passive stance. We extended Chow and Collins (Phys Rev E 52(1):909–912, 1995) one-dimensional pinned-polymer model (PPM) to two dimensions (2-D) and then evaluated the model’s ability to make analytical predictions for 2-D quiet stance. To test the model, we tracked center of mass and centers of foot pressures, and compared and contrasted stance sway for the anterior–posterior versus medio-lateral directions before, during, and after exposure to rotation at 10 rpm. Sway of the body during rotation generated Coriolis forces that acted perpendicular to the direction of sway. We found significant adaptive changes for three characteristic features of the mean square displacement (MSD) function: the exponent of the power law defined at short time scales, the proportionality constant of the power law, and the saturation plateau value defined at longer time scales. The exponent of the power law of MSD at a short time scale lies within the bounds predicted by the 2-D PPM. The change in MSD during exposure to rotation also had a power-law exponent in the range predicted by the theoretical model. We discuss the Coriolis force paradigm for studying postural and movement control and the applicability of the PPM model in 2-D for studying postural adaptation. PMID:24477760
Statistical analysis of quiet stance sway in 2-D.
Bakshi, Avijit; DiZio, Paul; Lackner, James R
2014-04-01
Subjects exposed to a rotating environment that perturbs their postural sway show adaptive changes in their voluntary spatially directed postural motion to restore accurate movement paths but do not exhibit any obvious learning during passive stance. We have found, however, that a variable known to characterize the degree of stochasticity in quiet stance can also reveal subtle learning phenomena in passive stance. We extended Chow and Collins (Phys Rev E 52(1):909-912, 1995) one-dimensional pinned-polymer model (PPM) to two dimensions (2-D) and then evaluated the model's ability to make analytical predictions for 2-D quiet stance. To test the model, we tracked center of mass and centers of foot pressures, and compared and contrasted stance sway for the anterior-posterior versus medio-lateral directions before, during, and after exposure to rotation at 10 rpm. Sway of the body during rotation generated Coriolis forces that acted perpendicular to the direction of sway. We found significant adaptive changes for three characteristic features of the mean square displacement (MSD) function: the exponent of the power law defined at short time scales, the proportionality constant of the power law, and the saturation plateau value defined at longer time scales. The exponent of the power law of MSD at a short time scale lies within the bounds predicted by the 2-D PPM. The change in MSD during exposure to rotation also had a power-law exponent in the range predicted by the theoretical model. We discuss the Coriolis force paradigm for studying postural and movement control and the applicability of the PPM model in 2-D for studying postural adaptation. PMID:24477760
Microwave Assisted 2D Materials Exfoliation
NASA Astrophysics Data System (ADS)
Wang, Yanbin
Two-dimensional materials have emerged as extremely important materials with applications ranging from energy and environmental science to electronics and biology. Here we report our discovery of a universal, ultrafast, green, solvo-thermal technology for producing excellent-quality, few-layered nanosheets in liquid phase from well-known 2D materials such as such hexagonal boron nitride (h-BN), graphite, and MoS2. We start by mixing the uniform bulk-layered material with a common organic solvent that matches its surface energy to reduce the van der Waals attractive interactions between the layers; next, the solutions are heated in a commercial microwave oven to overcome the energy barrier between bulk and few-layers states. We discovered the minutes-long rapid exfoliation process is highly temperature dependent, which requires precise thermal management to obtain high-quality inks. We hypothesize a possible mechanism of this proposed solvo-thermal process; our theory confirms the basis of this novel technique for exfoliation of high-quality, layered 2D materials by using an as yet unknown role of the solvent.
Photocurrent spectroscopy of 2D materials
NASA Astrophysics Data System (ADS)
Cobden, David
Confocal photocurrent measurements provide a powerful means of studying many aspects of the optoelectronic and electrical properties of a 2D device or material. At a diffraction-limited point they can provide a detailed absorption spectrum, and they can probe local symmetry, ultrafast relaxation rates and processes, electron-electron interaction strengths, and transport coefficients. We illustrate this with several examples, once being the photo-Nernst effect. In gapless 2D materials, such as graphene, in a perpendicular magnetic field a photocurrent antisymmetric in the field is generated near to the free edges, with opposite sign at opposite edges. Its origin is the transverse thermoelectric current associated with the laser-induced electron temperature gradient. This effect provides an unambiguous demonstration of the Shockley-Ramo nature of long-range photocurrent generation in gapless materials. It also provides a means of investigating quasiparticle properties. For example, in the case of graphene on hBN, it can be used to probe the Lifshitz transition that occurs due to the minibands formed by the Moire superlattice. We also observe and discuss photocurrent generated in other semimetallic (WTe2) and semiconducting (WSe2) monolayers. Work supported by DoE BES and NSF EFRI grants.
Multienzyme Inkjet Printed 2D Arrays.
Gdor, Efrat; Shemesh, Shay; Magdassi, Shlomo; Mandler, Daniel
2015-08-19
The use of printing to produce 2D arrays is well established, and should be relatively facile to adapt for the purpose of printing biomaterials; however, very few studies have been published using enzyme solutions as inks. Among the printing technologies, inkjet printing is highly suitable for printing biomaterials and specifically enzymes, as it offers many advantages. Formulation of the inkjet inks is relatively simple and can be adjusted to a variety of biomaterials, while providing nonharmful environment to the enzymes. Here we demonstrate the applicability of inkjet printing for patterning multiple enzymes in a predefined array in a very straightforward, noncontact method. Specifically, various arrays of the enzymes glucose oxidase (GOx), invertase (INV) and horseradish peroxidase (HP) were printed on aminated glass surfaces, followed by immobilization using glutardialdehyde after printing. Scanning electrochemical microscopy (SECM) was used for imaging the printed patterns and to ascertain the enzyme activity. The successful formation of 2D arrays consisting of enzymes was explored as a means of developing the first surface confined enzyme based logic gates. Principally, XOR and AND gates, each consisting of two enzymes as the Boolean operators, were assembled, and their operation was studied by SECM. PMID:26214072
Modeling of Gap Closure in Uranium-Zirconium Alloy Metal Fuel - A Test Problem
Simunovic, Srdjan; Ott, Larry J; Gorti, Sarma B; Nukala, Phani K; Radhakrishnan, Balasubramaniam; Turner, John A
2009-10-01
Uranium based binary and ternary alloy fuel is a possible candidate for advanced fast spectrum reactors with long refueling intervals and reduced liner heat rating [1]. An important metal fuel issue that can impact the fuel performance is the fuel-cladding gap closure, and fuel axial growth. The dimensional change in the fuel during irradiation is due to a superposition of the thermal expansion of the fuel due to heating, volumetric changes due to possible phase transformations that occur during heating and the swelling due to fission gas retention. The volumetric changes due to phase transformation depend both on the thermodynamics of the alloy system and the kinetics of phase change reactions that occur at the operating temperature. The nucleation and growth of fission gas bubbles that contributes to fuel swelling is also influenced by the local fuel chemistry and the microstructure. Once the fuel expands and contacts the clad, expansion in the radial direction is constrained by the clad, and the overall deformation of the fuel clad assembly depends upon the dynamics of the contact problem. The neutronics portion of the problem is also inherently coupled with microstructural evolution in terms of constituent redistribution and phase transformation. Because of the complex nature of the problem, a series of test problems have been defined with increasing complexity with the objective of capturing the fuel-clad interaction in complex fuels subjected to a wide range of irradiation and temperature conditions. The abstract, if short, is inserted here before the introduction section. If the abstract is long, it should be inserted with the front material and page numbered as such, then this page would begin with the introduction section.
Targeting multiple types of tumors using NKG2D-coated iron oxide nanoparticles
Wu, Ming-Ru; Cook, W. James; Zhang, Tong; Sentman, Charles L.
2015-01-01
Iron oxide nanoparticles (IONPs) hold great potential for cancer therapy. Actively targeting IONPs to tumor cells can further increase therapeutic efficacy and decrease off-target side effects. To target tumor cells, a natural killer (NK) cell activating receptor, NKG2D, was utilized to develop pan-tumor targeting IONPs. NKG2D ligands are expressed on many tumor types and its ligands are not found on most normal tissues under steady state conditions. The data showed that mouse and human fragment crystallizable (Fc) -fusion NKG2D (Fc-NKG2D) coated IONPs (NKG2D/NPs) can target multiple NKG2D ligand positive tumor types in vitro in a dose dependent manner by magnetic cell sorting. Tumor targeting effect was robust even under a very low tumor cell to normal cell ratio and targeting efficiency correlated with NKG2D ligand expression level on tumor cells. Furthermore, the magnetic separation platform utilized to test NKG2D/NP specificity has the potential to be developed into high throughput screening strategies to identify ideal fusion proteins or antibodies for targeting IONPs. In conclusion, NKG2D/NPs can be used to target multiple tumor types and magnetic separation platform can facilitate the proof-of-concept phase of tumor targeting IONP development. PMID:25371538
Targeting multiple types of tumors using NKG2D-coated iron oxide nanoparticles
NASA Astrophysics Data System (ADS)
Wu, Ming-Ru; Cook, W. James; Zhang, Tong; Sentman, Charles L.
2014-11-01
Iron oxide nanoparticles (IONPs) hold great potential for cancer therapy. Actively targeting IONPs to tumor cells can further increase therapeutic efficacy and decrease off-target side effects. To target tumor cells, a natural killer (NK) cell activating receptor, NKG2D, was utilized to develop pan-tumor targeting IONPs. NKG2D ligands are expressed on many tumor types and its ligands are not found on most normal tissues under steady state conditions. The data showed that mouse and human fragment crystallizable (Fc)-fusion NKG2D (Fc-NKG2D) coated IONPs (NKG2D/NPs) can target multiple NKG2D ligand positive tumor types in vitro in a dose dependent manner by magnetic cell sorting. Tumor targeting effect was robust even under a very low tumor cell to normal cell ratio and targeting efficiency correlated with NKG2D ligand expression level on tumor cells. Furthermore, the magnetic separation platform utilized to test NKG2D/NP specificity has the potential to be developed into high throughput screening strategies to identify ideal fusion proteins or antibodies for targeting IONPs. In conclusion, NKG2D/NPs can be used to target multiple tumor types and magnetic separation platform can facilitate the proof-of-concept phase of tumor targeting IONP development.
Interface adhesion between 2D materials and elastomers measured by buckle delamination
NASA Astrophysics Data System (ADS)
Brennan, Christopher; Lu, Nanshu
2015-03-01
A major application for 2D materials is creating electronic devices, including flexible and wearable devices. These applications require complicated fabrication processes where 2D materials are either mechanically exfoliated or grown via chemical vapor deposition and then transferred to a host substrate. Both processes require intimate knowledge of the interactions between the 2D material and the substrate to allow for a controllable transfer. Although adhesion between 2D materials and stiff substrates such as silicon and copper have been measured by bulge or peeling tests, adhesion between 2D materials and soft polymer substrates are hard to measure by conventional methods. Here we propose a simple way of measuring the adhesion between 2D materials and soft, stretchable elastomers using mature continuum mechanics equations. By creating buckle delamination in 2D atomic layers and measuring the buckle profile using an atomic force microscope, we can readily extract 2D-elastomer adhesion energy. Here we look at the adhesion of MoS2 and graphene to PDMS. The measured adhesion values are found insensitive to the applied strains in the substrate and are one order smaller than 2D-silicon oxide adhesion which is mainly attributed substrate surface roughness differences.
ERIC Educational Resources Information Center
Culhane, Terry, Ed.; And Others
Papers include: "ESP Tests: The Problem of Student Background Discipline (J. C. Alderson, A. H. Urquhart); "The Lexical Richness of L2 Written Productions and the Validity of Vocabulary Tests" (P. J. L. Arnaud); "The Construction of a Modular Thematic Task Bound English Language Achievement Test Battery for Kuwait" (M. I. Baghdady); "The Fill-In…
NASA Astrophysics Data System (ADS)
Lacava, C.; Carrol, L.; Bozzola, A.; Marchetti, R.; Minzioni, P.; Cristiani, I.; Fournier, M.; Bernabe, S.; Gerace, D.; Andreani, L. C.
2016-03-01
We present the characterization of Silicon-on-insulator (SOI) photonic-crystal based 2D grating-couplers (2D-GCs) fabricated by CEA-Leti in the frame of the FP7 Fabulous project, which is dedicated to the realization of devices and systems for low-cost and high-performance passives-optical-networks. On the analyzed samples different test structures are present, including 2D-GC connected to another 2D-GC by different waveguides (in a Mach-Zehnder like configuration), and 2D-GC connected to two separate 2D-GCs, so as to allow a complete assessment of different parameters. Measurements were carried out using a tunable laser source operating in the extended telecom bandwidth and a fiber-based polarization controlling system at the input of device-under-test. The measured data yielded an overall fiber-to-fiber loss of 7.5 dB for the structure composed by an input 2D-GC connected to two identical 2D-GCs. This value was obtained at the peak wavelength of the grating, and the 3-dB bandwidth of the 2D-GC was assessed to be 43 nm. Assuming that the waveguide losses are negligible, so as to make a worst-case analysis, the coupling efficiency of the single 2D-GC results to be equal to -3.75 dB, constituting, to the best of our knowledge, the lowest value ever reported for a fully CMOS compatible 2D-GC. It is worth noting that both the obtained values are in good agreement with those expected by the numerical simulations performed using full 3D analysis by Lumerical FDTD-solutions.
Breakdown of wave diffusion in 2D due to loops.
Haney, Matthew; Snieder, Roel
2003-08-29
The validity of the diffusion approximation for the intensity of multiply scattered waves is tested with numerical simulations in a strongly scattering 2D medium of finite extent. We show that the diffusion equation underestimates the intensity and attribute this to both the neglect of recurrent scattering paths and interference within diffusion theory. We present a theory to quantify this discrepancy based on counting all possible scattering paths between point scatterers. Interference phenomena, due to loop paths, are incorporated in a way similar to coherent backscattering. PMID:14525183
Automatic differentiation of the TACO2D finite element code using ADIFOR
Carle, A.; Fagan, M.
1996-04-01
The need for sensitivities in particular applications is becoming increasingly important in problems such as optimal design or control. In this study, the authors use ADIFOR to generate derivative code for TACO2D, a finite element heat transfer code. The study of TACO2D indicates that ADIFOR-generated derivatives yield accurate derivatives at a fraction of the time requirements of finite difference approximations, and space requirements proportional to the number of variables. The primary focus on TACO2D was for the design of chemical vapor deposition reactors.
A Novel Crosstalk Suppression Method of the 2-D Networked Resistive Sensor Array
Wu, Jianfeng; Wang, Lei; Li, Jianqing; Song, Aiguo
2014-01-01
The 2-D resistive sensor array in the row–column fashion suffered from the crosstalk problem for parasitic parallel paths. Firstly, we proposed an Improved Isolated Drive Feedback Circuit with Compensation (IIDFCC) based on the voltage feedback method to suppress the crosstalk. In this method, a compensated resistor was specially used to reduce the crosstalk caused by the column multiplexer resistors and the adjacent row elements. Then, a mathematical equivalent resistance expression of the element being tested (EBT) of this circuit was analytically derived and verified by the circuit simulations. The simulation results show that the measurement method can greatly reduce the influence on the EBT caused by parasitic parallel paths for the multiplexers' channel resistor and the adjacent elements. PMID:25046011
Graphene suspensions for 2D printing
NASA Astrophysics Data System (ADS)
Soots, R. A.; Yakimchuk, E. A.; Nebogatikova, N. A.; Kotin, I. A.; Antonova, I. V.
2016-04-01
It is shown that, by processing a graphite suspension in ethanol or water by ultrasound and centrifuging, it is possible to obtain particles with thicknesses within 1-6 nm and, in the most interesting cases, 1-1.5 nm. Analogous treatment of a graphite suspension in organic solvent yields eventually thicker particles (up to 6-10 nm thick) even upon long-term treatment. Using the proposed ink based on graphene and aqueous ethanol with ethylcellulose and terpineol additives for 2D printing, thin (~5 nm thick) films with sheet resistance upon annealing ~30 MΩ/□ were obtained. With the ink based on aqueous graphene suspension, the sheet resistance was ~5-12 kΩ/□ for 6- to 15-nm-thick layers with a carrier mobility of ~30-50 cm2/(V s).
NASA Astrophysics Data System (ADS)
Torgoev, Almaz; Havenith, Hans-Balder
2016-01-01
A 2D elasto-dynamic modelling of the pure topographic seismic response is performed for six models with a total length of around 23.0 km. These models are reconstructed from the real topographic settings of the landslide-prone slopes situated in the Mailuu-Suu River Valley, Southern Kyrgyzstan. The main studied parameter is the Arias Intensity (Ia, m/sec), which is applied in the GIS-based Newmark method to regionally map the seismically-induced landslide susceptibility. This method maps the Ia values via empirical attenuation laws and our studies investigate a potential to include topographic input into them. Numerical studies analyse several signals with varying shape and changing central frequency values. All tests demonstrate that the spectral amplification patterns directly affect the amplification of the Ia values. These results let to link the 2D distribution of the topographically amplified Ia values with the parameter called as smoothed curvature. The amplification values for the low-frequency signals are better correlated with the curvature smoothed over larger spatial extent, while those values for the high-frequency signals are more linked to the curvature with smaller smoothing extent. The best predictions are provided by the curvature smoothed over the extent calculated according to Geli's law. The sample equations predicting the Ia amplification based on the smoothed curvature are presented for the sinusoid-shape input signals. These laws cannot be directly implemented in the regional Newmark method, as 3D amplification of the Ia values addresses more problem complexities which are not studied here. Nevertheless, our 2D results prepare the theoretical framework which can potentially be applied to the 3D domain and, therefore, represent a robust basis for these future research targets.
NASA Astrophysics Data System (ADS)
Torgoev, Almaz; Havenith, Hans-Balder
2016-07-01
A 2D elasto-dynamic modelling of the pure topographic seismic response is performed for six models with a total length of around 23.0 km. These models are reconstructed from the real topographic settings of the landslide-prone slopes situated in the Mailuu-Suu River Valley, Southern Kyrgyzstan. The main studied parameter is the Arias Intensity (Ia, m/sec), which is applied in the GIS-based Newmark method to regionally map the seismically-induced landslide susceptibility. This method maps the Ia values via empirical attenuation laws and our studies investigate a potential to include topographic input into them. Numerical studies analyse several signals with varying shape and changing central frequency values. All tests demonstrate that the spectral amplification patterns directly affect the amplification of the Ia values. These results let to link the 2D distribution of the topographically amplified Ia values with the parameter called as smoothed curvature. The amplification values for the low-frequency signals are better correlated with the curvature smoothed over larger spatial extent, while those values for the high-frequency signals are more linked to the curvature with smaller smoothing extent. The best predictions are provided by the curvature smoothed over the extent calculated according to Geli's law. The sample equations predicting the Ia amplification based on the smoothed curvature are presented for the sinusoid-shape input signals. These laws cannot be directly implemented in the regional Newmark method, as 3D amplification of the Ia values addresses more problem complexities which are not studied here. Nevertheless, our 2D results prepare the theoretical framework which can potentially be applied to the 3D domain and, therefore, represent a robust basis for these future research targets.
Quantum damped oscillator II: Bateman's Hamiltonian vs. 2D parabolic potential barrier
Chruscinski, Dariusz . E-mail: darch@phys.uni.torun.pl
2006-04-15
We show that quantum Bateman's system which arises in the quantization of a damped harmonic oscillator is equivalent to a quantum problem with 2D parabolic potential barrier known also as 2D inverted isotropic oscillator. It turns out that this system displays the family of complex eigenvalues corresponding to the poles of analytical continuation of the resolvent operator to the complex energy plane. It is shown that this representation is more suitable than the hyperbolic one used recently by Blasone and Jizba.
Metrology for graphene and 2D materials
NASA Astrophysics Data System (ADS)
Pollard, Andrew J.
2016-09-01
The application of graphene, a one atom-thick honeycomb lattice of carbon atoms with superlative properties, such as electrical conductivity, thermal conductivity and strength, has already shown that it can be used to benefit metrology itself as a new quantum standard for resistance. However, there are many application areas where graphene and other 2D materials, such as molybdenum disulphide (MoS2) and hexagonal boron nitride (h-BN), may be disruptive, areas such as flexible electronics, nanocomposites, sensing and energy storage. Applying metrology to the area of graphene is now critical to enable the new, emerging global graphene commercial world and bridge the gap between academia and industry. Measurement capabilities and expertise in a wide range of scientific areas are required to address this challenge. The combined and complementary approach of varied characterisation methods for structural, chemical, electrical and other properties, will allow the real-world issues of commercialising graphene and other 2D materials to be addressed. Here, examples of metrology challenges that have been overcome through a multi-technique or new approach are discussed. Firstly, the structural characterisation of defects in both graphene and MoS2 via Raman spectroscopy is described, and how nanoscale mapping of vacancy defects in graphene is also possible using tip-enhanced Raman spectroscopy (TERS). Furthermore, the chemical characterisation and removal of polymer residue on chemical vapour deposition (CVD) grown graphene via secondary ion mass spectrometry (SIMS) is detailed, as well as the chemical characterisation of iron films used to grow large domain single-layer h-BN through CVD growth, revealing how contamination of the substrate itself plays a role in the resulting h-BN layer. In addition, the role of international standardisation in this area is described, outlining the current work ongoing in both the International Organization of Standardization (ISO) and the
IP4DI: A software for time-lapse 2D/3D DC-resistivity and induced polarization tomography
NASA Astrophysics Data System (ADS)
Karaoulis, M.; Revil, A.; Tsourlos, P.; Werkema, D. D.; Minsley, B. J.
2013-04-01
We propose a 2D/3D forward modelling and inversion package to invert direct current (DC)-resistivity, time-domain induced polarization (TDIP), and frequency-domain induced polarization (FDIP) data. Each cell used for the discretization of the 2D/3D problems is characterized by a DC-resistivity value and a chargeability or complex conductivity for TDIP/FDIP problems, respectively. The governing elliptic partial differential equations are solved with the finite element method, which can be applied for both real and complex numbers. The inversion can be performed either for a single snapshot of data or for a sequence of snapshots in order to monitor a dynamic process such as a salt tracer test. For the time-lapse inversion, we have developed an active time constrained (ATC) approach that is very efficient in filtering out noise in the data that is not correlated over time. The forward algorithm is benchmarked with simple analytical solutions. The inversion package IP4DI is benchmarked with three tests, two including simple geometries. The last one corresponds to a time-lapse resistivity problem for cross-well tomography during enhanced oil recovery. The algorithms are based on MATLAB® code package and a graphical user interface (GUI).
Hirobe, Tomohisa; Ito, Shosuke; Wakamatsu, Kazumasa
2013-09-01
The novel mutation named ru2(d) /Hps5(ru2-d) , characterized by light-colored coats and ruby-eyes, prohibits differentiation of melanocytes by inhibiting tyrosinase (Tyr) activity, expression of Tyr, Tyr-related protein 1 (Tyrp1), Tyrp2, and Kit. However, it is not known whether the ru2(d) allele affects pheomelanin synthesis in recessive yellow (e/Mc1r(e) ) or in pheomelanic stage in agouti (A) mice. In this study, effects of the ru2(d) allele on pheomelanin synthesis were investigated by chemical analysis of melanin present in dorsal hairs of 5-week-old mice from F2 generation between C57BL/10JHir (B10)-co-isogenic ruby-eye 2(d) and B10-congenic recessive yellow or agouti. Eumelanin content was decreased in ruby-eye 2(d) and ruby-eye 2(d) agouti mice, whereas pheomelanin content in ruby-eye 2(d) recessive yellow and ruby-eye 2(d) agouti mice did not differ from the corresponding Ru2(d) /- mice, suggesting that the ru2(d) allele inhibits eumelanin but not pheomelanin synthesis. PMID:23672590
Segmentation of 2D gel electrophoresis spots using a Markov random field
NASA Astrophysics Data System (ADS)
Hoeflich, Christopher S.; Corso, Jason J.
2009-02-01
We propose a statistical model-based approach for the segmentation of fragments of DNA as a first step in the automation of the primarily manual process of comparing two or more images resulting from the Restriction Landmark Genomic Scanning (RLGS) method. These 2D gel electrophoresis images are the product of the separation of DNA into fragments that appear as spots on X-ray films. The goal is to find instances where a spot appears in one image and not in another since a missing spot can be correlated with a region of DNA that has been affected by a disease such as cancer. The entire comparison process is typically done manually, which is tedious and very error prone. We pose the problem as the labeling of each image pixel as either a spot or non-spot and use a Markov Random Field (MRF) model and simulated annealing for inference. Neighboring spot labels are then connected to form spot regions. The MRF based model was tested on actual 2D gel electrophoresis images.
Pilot study risk assessment for selected problems at the Nevada Test Site (NTS)
Daniels, J.I.
1993-06-01
The Nevada Test Site (NTS) is located in southwestern Nevada, about 105 km (65 mi) northwest of the city of Las Vegas. A series of tests was conducted in the late 1950s and early 1960s at or near the NTS to study issues involving plutonium-bearing devices. These tests resulted in the dispersal of about 5 TBq of [sup 239,24O]Pu on the surficial soils at the test locations. Additionally, underground tests of nuclear weapons devices have been conducted at the NTS since late 1962; ground water beneath the NTS has been contaminated with radionuclides produced by these tests. These two important problems have been selected for assessment. Regarding the plutonium contamination, because the residual [sup 239]Pu decays slowly (half-life of 24,110 y), these sites could represent a long-term hazard if they are not remediated and if institutional controls are lost. To investigate the magnitude of the potential health risks for this no-remediation case, three basic exposure scenarios were defined that could bring individuals in contact with [sup 239,24O]Pu at the sites: (1) a resident living in a subdivision, (2) a resident farmer, and (3) a worker at a commercial facility -- all located at a test site. The predicted cancer risks for the resident farmer were more than a factor of three times higher than the suburban resident at the median risk level, and about a factor of ten greater than the reference worker at a commercial facility. At 100 y from the present, the 5, 50, and 95th percentile risks for the resident farmer at the most contaminated site were 4 x 10[sup [minus]6], 6 x 10[sup [minus]5], and 5 x 10[sup [minus]4], respectively. For the assessment of Pu in surface soil, the principal sources of uncertainty in the estimated risks were population mobility, the relationship between indoor and outdoor contaminant levels, and the dose and risk factors for bone, liver, and lung.
Pilot study risk assessment for selected problems at the Nevada Test Site (NTS)
Daniels, J.I.; Anspaugh, L.R.; Bogen, K.T.; Daniels, J.I.; Layton, D.W.; Straume, T.; Andricevic, R.; Jacobson, R.L.; Meinhold, A.F.; Holtzman, S.; Morris, S.C.; Hamilton, L.D.
1993-06-01
The Nevada Test Site (NTS) is located in southwestern Nevada, about 105 km (65 mi) northwest of the city of Las Vegas. A series of tests was conducted in the late 1950s and early 1960s at or near the NTS to study issues involving plutonium-bearing devices. These tests resulted in the dispersal of about 5 TBq of {sup 239,24O}Pu on the surficial soils at the test locations. Additionally, underground tests of nuclear weapons devices have been conducted at the NTS since late 1962; ground water beneath the NTS has been contaminated with radionuclides produced by these tests. These two important problems have been selected for assessment. Regarding the plutonium contamination, because the residual {sup 239}Pu decays slowly (half-life of 24,110 y), these sites could represent a long-term hazard if they are not remediated and if institutional controls are lost. To investigate the magnitude of the potential health risks for this no-remediation case, three basic exposure scenarios were defined that could bring individuals in contact with {sup 239,24O}Pu at the sites: (1) a resident living in a subdivision, (2) a resident farmer, and (3) a worker at a commercial facility -- all located at a test site. The predicted cancer risks for the resident farmer were more than a factor of three times higher than the suburban resident at the median risk level, and about a factor of ten greater than the reference worker at a commercial facility. At 100 y from the present, the 5, 50, and 95th percentile risks for the resident farmer at the most contaminated site were 4 x 10{sup {minus}6}, 6 x 10{sup {minus}5}, and 5 x 10{sup {minus}4}, respectively. For the assessment of Pu in surface soil, the principal sources of uncertainty in the estimated risks were population mobility, the relationship between indoor and outdoor contaminant levels, and the dose and risk factors for bone, liver, and lung.
Allison, Scott A; Sweet, Clifford F; Beall, Douglas P; Lewis, Thomas E; Monroe, Thomas
2005-09-01
The PACS implementation process is complicated requiring a tremendous amount of time, resources, and planning. The Department of Defense (DOD) has significant experience in developing and refining PACS acceptance testing (AT) protocols that assure contract compliance, clinical safety, and functionality. The DOD's AT experience under the initial Medical Diagnostic Imaging Support System contract led to the current Digital Imaging Network-Picture Archiving and Communications Systems (DIN-PACS) contract AT protocol. To identify the most common system and component deficiencies under the current DIN-PACS AT protocol, 14 tri-service sites were evaluated during 1998-2000. Sixteen system deficiency citations with 154 separate types of limitations were noted with problems involving the workstation, interfaces, and the Radiology Information System comprising more than 50% of the citations. Larger PACS deployments were associated with a higher number of deficiencies. The most commonly cited systems deficiencies were among the most expensive components of the PACS. PMID:15924273
Goldstien, J H; Harmon, J; McGhee, P E; Karasik, R
1975-01-01
Heart rate and skin conductance were monitored continuously while subjects heard and responded to seven riddles and seven problems which were structurally similar. Hypotheses based on cognitive problem-solving models of humor were tested by comparing changes occurring during problem solving with those that occurred during "riddle solving". While heart rate tended to accelerate once a riddle or problem was presented and decelerate once the answer or punch line was given, there were significant differences in skin conductance between riddle- and problem-solving responses, suggesting that a purely problem-solving model of humor may be untenable. A test of Berlyne's hypothesis of arousal change and humor appreciation was also made. As predicted, humor appreciation was greatest for those who showed a moderate amount of change. PMID:1113088
Non-Iterative Rigid 2D/3D Point-Set Registration Using Semidefinite Programming
NASA Astrophysics Data System (ADS)
Khoo, Yuehaw; Kapoor, Ankur
2016-07-01
We describe a convex programming framework for pose estimation in 2D/3D point-set registration with unknown point correspondences. We give two mixed-integer nonlinear program (MINP) formulations of the 2D/3D registration problem when there are multiple 2D images, and propose convex relaxations for both of the MINPs to semidefinite programs (SDP) that can be solved efficiently by interior point methods. Our approach to the 2D/3D registration problem is non-iterative in nature as we jointly solve for pose and correspondence. Furthermore, these convex programs can readily incorporate feature descriptors of points to enhance registration results. We prove that the convex programs exactly recover the solution to the original nonconvex 2D/3D registration problem under noiseless condition. We apply these formulations to the registration of 3D models of coronary vessels to their 2D projections obtained from multiple intra-operative fluoroscopic images. For this application, we experimentally corroborate the exact recovery property in the absence of noise and further demonstrate robustness of the convex programs in the presence of noise.
2D to 3D conversion implemented in different hardware
NASA Astrophysics Data System (ADS)
Ramos-Diaz, Eduardo; Gonzalez-Huitron, Victor; Ponomaryov, Volodymyr I.; Hernandez-Fragoso, Araceli
2015-02-01
Conversion of available 2D data for release in 3D content is a hot topic for providers and for success of the 3D applications, in general. It naturally completely relies on virtual view synthesis of a second view given by original 2D video. Disparity map (DM) estimation is a central task in 3D generation but still follows a very difficult problem for rendering novel images precisely. There exist different approaches in DM reconstruction, among them manually and semiautomatic methods that can produce high quality DMs but they demonstrate hard time consuming and are computationally expensive. In this paper, several hardware implementations of designed frameworks for an automatic 3D color video generation based on 2D real video sequence are proposed. The novel framework includes simultaneous processing of stereo pairs using the following blocks: CIE L*a*b* color space conversions, stereo matching via pyramidal scheme, color segmentation by k-means on an a*b* color plane, and adaptive post-filtering, DM estimation using stereo matching between left and right images (or neighboring frames in a video), adaptive post-filtering, and finally, the anaglyph 3D scene generation. Novel technique has been implemented on DSP TMS320DM648, Matlab's Simulink module over a PC with Windows 7, and using graphic card (NVIDIA Quadro K2000) demonstrating that the proposed approach can be applied in real-time processing mode. The time values needed, mean Similarity Structural Index Measure (SSIM) and Bad Matching Pixels (B) values for different hardware implementations (GPU, Single CPU, and DSP) are exposed in this paper.
3D track initiation in clutter using 2D measurements
NASA Astrophysics Data System (ADS)
Lin, Lin; Kirubarajan, Thiagalingam; Bar-Shalom, Yaakov
2001-11-01
In this paper we present an algorithm for initiating 3-D tracks using range and azimuth (bearing) measurements from a 2-D radar on a moving platform. The work is motivated by the need to track possibly low-flying targets, e.g., cruise missiles, using reports from an aircraft-based surveillance radar. Previous work on this problem considered simple linear motion in a flat earth coordinate frame. Our research extends this to a more realistic scenario where the earth"s curvature is also considered. The target is assumed to be moving along a great circle at a constant altitude. After the necessary coordinate transformations, the measurements are nonlinear functions of the target state and the observability of target altitude is severely limited. The observability, quantified by the Cramer-Rao Lower Bound (CRLB), is very sensitive to the sensor-to-target geometry. The paper presents a Maximum Likelihood (ML) estimator for estimating the target motion parameters in the Earth Centered Earth Fixed coordinate frame from 2-D range and angle measurements. In order to handle the possibility of false measurements and missed detections, which was not considered in, we use the Probabilistic Data Association (PDA) algorithm to weight the detections in a frame. The PDA-based modified global likelihood is optimized using a numerical search. The accuracies obtained by the resulting ML-PDA estimator are quantified using the CRLB for different sensor-target configurations. It is shown that the proposed estimator is efficient, that is, it meets the CRLB. Of particular interest is the achievable accuracy for estimating the target altitude, which is not observed directly by the 2-D radar, but can be only inferred from the range and bearing observations.
Code of Federal Regulations, 2010 CFR
2010-10-01
... 49 Transportation 1 2010-10-01 2010-10-01 false What procedural problems do not result in the... Problems in Drug Tests § 40.209 What procedural problems do not result in the cancellation of a test and do... aware, even if they are not considered problems that will cause a test to be cancelled as listed in...
Code of Federal Regulations, 2011 CFR
2011-10-01
... 49 Transportation 1 2011-10-01 2011-10-01 false What procedural problems do not result in the... Problems in Drug Tests § 40.209 What procedural problems do not result in the cancellation of a test and do... aware, even if they are not considered problems that will cause a test to be cancelled as listed in...
ERIC Educational Resources Information Center
Erford, Bradley T.; Alsamadi, Silvana C.
2012-01-01
Score reliability and validity of parent responses concerning their 10- to 17-year-old students were analyzed using the Screening Test for Emotional Problems-Parent Report (STEP-P), which assesses a variety of emotional problems classified under the Individuals with Disabilities Education Improvement Act. Score reliability, convergent, and…
NASA Astrophysics Data System (ADS)
Martynenko, O. N.; Serdakov, A. S.
1980-06-01
A structural program is developed for the solving of the problems of technical testing of radioelectronic equipment. The following problems are considered: construction of a full matrix of single defects, transformation of the full matrix into a minimized one, and decoding the minimized matrix in order to localize the number of a failed element.
Tay-Sontheimer, Jessica; Shireman, Laura M; Beyer, Richard P; Senn, Taurence; Witten, Daniela; Pearce, Robin E; Gaedigk, Andrea; Fomban, Cletus L Gana; Lutz, Justin D; Isoherranen, Nina; Thummel, Kenneth E; Fiehn, Oliver; Leeder, J Steven; Lin, Yvonne S
2015-01-01
Aim We sought to discover endogenous urinary biomarkers of human CYP2D6 activity. Patients & methods Healthy pediatric subjects (n = 189) were phenotyped using dextromethorphan and randomized for candidate biomarker selection and validation. Global urinary metabolomics was performed using liquid chromatography quadrupole time-of-flight mass spectrometry. Candidate biomarkers were tested in adults receiving fluoxetine, a CYP2D6 inhibitor. Results A biomarker, M1 (m/z 444.3102) was correlated with CYP2D6 activity in both the pediatric training and validation sets. Poor metabolizers had undetectable levels of M1, whereas it was present in subjects with other phenotypes. In adult subjects, a 9.56-fold decrease in M1 abundance was observed during CYP2D6 inhibition. Conclusion Identification and validation of M1 may provide a noninvasive means of CYP2D6 phenotyping. PMID:25521354
Computing 2D constrained delaunay triangulation using the GPU.
Qi, Meng; Cao, Thanh-Tung; Tan, Tiow-Seng
2013-05-01
We propose the first graphics processing unit (GPU) solution to compute the 2D constrained Delaunay triangulation (CDT) of a planar straight line graph (PSLG) consisting of points and edges. There are many existing CPU algorithms to solve the CDT problem in computational geometry, yet there has been no prior approach to solve this problem efficiently using the parallel computing power of the GPU. For the special case of the CDT problem where the PSLG consists of just points, which is simply the normal Delaunay triangulation (DT) problem, a hybrid approach using the GPU together with the CPU to partially speed up the computation has already been presented in the literature. Our work, on the other hand, accelerates the entire computation on the GPU. Our implementation using the CUDA programming model on NVIDIA GPUs is numerically robust, and runs up to an order of magnitude faster than the best sequential implementations on the CPU. This result is reflected in our experiment with both randomly generated PSLGs and real-world GIS data having millions of points and edges. PMID:23492377
Mantovani, A; Stazi, A V; Macrì, C; Maranghi, F; Ricciardi, C
1999-10-01
Endocrine disrupting chemicals (EDCs) may affect mammalian development either indirectly (by impairing implantation, placental development, lactation, etc.) or directly, altering the maturation of target tissues. Current regulatory tests for reproductive/developmental toxicity should be carefully evaluated with regard to risk assessment of EDCs, considering hazard identification (are relevant endpoints being assessed?) and dose-response assessment (are sensitive NOEL/dose-response curves being provided?). Many in vitro and in vivo assays for sex steroid disruption are available; provided that the metabolic capacities of the assays are defined, they could be integrated in a sensitive battery for early detection of steroid-disrupting potentials. The screening battery should address further regulatory in vivo tests (e.g. what specific parameters have to be investigated). As regards dose-response, qualitative differences may be observed between lower and higher exposures, showing primary hormone-related effects and frank embryotoxicity, respectively. Other problems concern (a) the identification of critical developmental windows, according to hormone concentrations and/or receptor levels in the developing target tissues; (b) the potential for interactions between chemicals with common mechanism/target (e.g. xenoestrogens); (c) most important, besides sex steroids more attention should be given to other mechanisms of endocrine disruption, e.g., thyroid effects, which can be highly relevant to prenatal and postnatal development. PMID:10467724
Is There a Space-Based Technology Solution to Problems with Preclinical Drug Toxicity Testing?
Hammond, Timothy; Allen, Patricia; Birdsall, Holly
2016-07-01
Even the finest state-of-the art preclinical drug testing, usually in primary hepatocytes, remains an imperfect science. Drugs continue to be withdrawn from the market due to unforeseen toxicity, side effects, and drug interactions. The space program may be able to provide a lifeline. Best known for rockets, space shuttles, astronauts and engineering, the space program has also delivered some serious medical science. Optimized suspension culture in NASA's specialized suspension culture devices, known as rotating wall vessels, uniquely maintains Phase I and Phase II drug metabolizing pathways in hepatocytes for weeks in cell culture. Previously prohibitively expensive, new materials and 3D printing techniques have the potential to make the NASA rotating wall vessel available inexpensively on an industrial scale. Here we address the tradeoffs inherent in the rotating wall vessel, limitations of alternative approaches for drug metabolism studies, and the market to be addressed. Better pre-clinical drug testing has the potential to significantly reduce the morbidity and mortality of one of the most common problems in modern medicine: adverse events related to pharmaceuticals. PMID:27183841
Reliability of screening tests for health-related problems among low-income elderly.
Lino, Valéria Teresa Saraiva; Portela, Margareth Crisóstomo; Camacho, Luiz Antônio Bastos; Rodrigues, Nádia Cristina Pinheiro
2014-12-01
Screening tests for health problems can identify elderly people who should undergo the Comprehensive Geriatric Assessment, enabling the planning of actions to prevent disability. The aim of this study was to analyze the inter-rater reliability (IRR) of self-assessment questions (SAQ) and performance tests (PT) recommended in Brazil, in a sample of low-income elderly people, through an exploratory study performed with 165 elderly assessed by two professionals on different days. IRR was evaluated using the intraclass correlation coefficient (ICC) for continuous variables and the kappa statistic for categorical ones. The IRR for the PT (muscle strength, mobility body mass index, vision) was excellent and presented ICC values greater than 0.75. By contrast, the IRR for SAQ (urinary incontinence, self-perceived health and hearing impairment) was intermediate. Only the fall-related item presented a good IRR. In this study single SAQ had poor reliability when compared to PT, suggesting the necessity of revision of subjective self-assessment items with low reproducibility before implementation. PMID:26247997
Modelling RF sources using 2-D PIC codes
Eppley, K.R.
1993-03-01
In recent years, many types of RF sources have been successfully modelled using 2-D PIC codes. Both cross field devices (magnetrons, cross field amplifiers, etc.) and pencil beam devices (klystrons, gyrotrons, TWT`S, lasertrons, etc.) have been simulated. All these devices involve the interaction of an electron beam with an RF circuit. For many applications, the RF structure may be approximated by an equivalent circuit, which appears in the simulation as a boundary condition on the electric field (``port approximation``). The drive term for the circuit is calculated from the energy transfer between beam and field in the drift space. For some applications it may be necessary to model the actual geometry of the structure, although this is more expensive. One problem not entirely solved is how to accurately model in 2-D the coupling to an external waveguide. Frequently this is approximated by a radial transmission line, but this sometimes yields incorrect results. We also discuss issues in modelling the cathode and injecting the beam into the PIC simulation.
2D Gridded Surface Data Value-Added Product
Tang, Q; Xie, S
2015-08-30
This report describes the Atmospheric Radiation Measurement (ARM) Best Estimate (ARMBE) 2-dimensional (2D) gridded surface data (ARMBE2DGRID) value-added product. Spatial variability is critically important to many scientific studies, especially those that involve processes of great spatial variations at high temporal frequency (e.g., precipitation, clouds, radiation, etc.). High-density ARM sites deployed at the Southern Great Plains (SGP) allow us to observe the spatial patterns of variables of scientific interests. The upcoming megasite at SGP with its enhanced spatial density will facilitate the studies at even finer scales. Currently, however, data are reported only at individual site locations at different time resolutions for different datastreams. It is difficult for users to locate all the data they need and requires extra effort to synchronize the data. To address these problems, the ARMBE2DGRID value-added product merges key surface measurements at the ARM SGP sites and interpolates the data to a regular 2D grid to facilitate the data application.
Modelling RF sources using 2-D PIC codes
Eppley, K.R.
1993-03-01
In recent years, many types of RF sources have been successfully modelled using 2-D PIC codes. Both cross field devices (magnetrons, cross field amplifiers, etc.) and pencil beam devices (klystrons, gyrotrons, TWT'S, lasertrons, etc.) have been simulated. All these devices involve the interaction of an electron beam with an RF circuit. For many applications, the RF structure may be approximated by an equivalent circuit, which appears in the simulation as a boundary condition on the electric field ( port approximation''). The drive term for the circuit is calculated from the energy transfer between beam and field in the drift space. For some applications it may be necessary to model the actual geometry of the structure, although this is more expensive. One problem not entirely solved is how to accurately model in 2-D the coupling to an external waveguide. Frequently this is approximated by a radial transmission line, but this sometimes yields incorrect results. We also discuss issues in modelling the cathode and injecting the beam into the PIC simulation.
TOPAZ2D heat transfer code users manual and thermal property data base
Shapiro, A.B.; Edwards, A.L.
1990-05-01
TOPAZ2D is a two dimensional implicit finite element computer code for heat transfer analysis. This user's manual provides information on the structure of a TOPAZ2D input file. Also included is a material thermal property data base. This manual is supplemented with The TOPAZ2D Theoretical Manual and the TOPAZ2D Verification Manual. TOPAZ2D has been implemented on the CRAY, SUN, and VAX computers. TOPAZ2D can be used to solve for the steady state or transient temperature field on two dimensional planar or axisymmetric geometries. Material properties may be temperature dependent and either isotropic or orthotropic. A variety of time and temperature dependent boundary conditions can be specified including temperature, flux, convection, and radiation. Time or temperature dependent internal heat generation can be defined locally be element or globally by material. TOPAZ2D can solve problems of diffuse and specular band radiation in an enclosure coupled with conduction in material surrounding the enclosure. Additional features include thermally controlled reactive chemical mixtures, thermal contact resistance across an interface, bulk fluid flow, phase change, and energy balances. Thermal stresses can be calculated using the solid mechanics code NIKE2D which reads the temperature state data calculated by TOPAZ2D. A three dimensional version of the code, TOPAZ3D is available. The material thermal property data base, Chapter 4, included in this manual was originally published in 1969 by Art Edwards for use with his TRUMP finite difference heat transfer code. The format of the data has been altered to be compatible with TOPAZ2D. Bob Bailey is responsible for adding the high explosive thermal property data.
3D-2D registration of cerebral angiograms: a method and evaluation on clinical images.
Mitrovic, Uroš; Špiclin, Žiga; Likar, Boštjan; Pernuš, Franjo
2013-08-01
Endovascular image-guided interventions (EIGI) involve navigation of a catheter through the vasculature followed by application of treatment at the site of anomaly using live 2D projection images for guidance. 3D images acquired prior to EIGI are used to quantify the vascular anomaly and plan the intervention. If fused with the information of live 2D images they can also facilitate navigation and treatment. For this purpose 3D-2D image registration is required. Although several 3D-2D registration methods for EIGI achieve registration accuracy below 1 mm, their clinical application is still limited by insufficient robustness or reliability. In this paper, we propose a 3D-2D registration method based on matching a 3D vasculature model to intensity gradients of live 2D images. To objectively validate 3D-2D registration methods, we acquired a clinical image database of 10 patients undergoing cerebral EIGI and established "gold standard" registrations by aligning fiducial markers in 3D and 2D images. The proposed method had mean registration accuracy below 0.65 mm, which was comparable to tested state-of-the-art methods, and execution time below 1 s. With the highest rate of successful registrations and the highest capture range the proposed method was the most robust and thus a good candidate for application in EIGI. PMID:23649179
NASA Astrophysics Data System (ADS)
Cheng, Chingyun; Kangara, Jayampathi; Arakelyan, Ilya; Thomas, John
2016-05-01
We tune the dimensionality of a strongly interacting degenerate 6 Li Fermi gas from 2D to quasi-2D, by adjusting the radial confinement of pancake-shaped clouds to control the radial chemical potential. In the 2D regime with weak radial confinement, the measured pair binding energies are in agreement with 2D-BCS mean field theory, which predicts dimer pairing energies in the many-body regime. In the qausi-2D regime obtained with increased radial confinement, the measured pairing energy deviates significantly from 2D-BCS theory. In contrast to the pairing energy, the measured radii of the cloud profiles are not fit by 2D-BCS theory in either the 2D or quasi-2D regimes, but are fit in both regimes by a beyond mean field polaron-model of the free energy. Supported by DOE, ARO, NSF, and AFOSR.
Repression of multiple CYP2D genes in mouse primary hepatocytes with a single siRNA construct.
Elraghy, Omaima; Baldwin, William S
2015-01-01
The Cyp2d subfamily is the second most abun-dant subfamily of hepatic drug-metabolizing CYPs. In mice, there are nine Cyp2d members that are believed to have redundant catalytic activity. We are testing and optimizing the ability of one short interfering RNA (siRNA) construct to knockdown the expression of multiple mouse Cyp2ds in primary hepatocytes. Expression of Cyp2d10, Cyp2d11, Cyp2d22, and Cyp2d26 was observed in the primary male mouse hepatocytes. Cyp2d9, which is male-specific and growth hormone-dependent, was not expressed in male primary hepatocytes, potentially because of its dependence on pulsatile growth hormone release from the anterior pituitary. Several different siRNAs at different concentrations and with different reagents were used to knockdown Cyp2d expression. siRNA constructs designed to repress only one construct often mildly repressed several Cyp2d isoforms. A construct designed to knockdown every Cyp2d isoform provided the best results, especially when incubated with transfection reagents designed specifically for primary cell culture. Interestingly, a construct designed to knockdown all Cyp2d isoforms, except Cyp2d10, caused a 2.5× increase in Cyp2d10 expression, presumably because of a compensatory response. However, while RNA expression is repressed 24 h after siRNA treatment, associated changes in Cyp2d-mediated metabolism are tenuous. Overall, this study provides data on the expression of murine Cyp2ds in primary cell lines, valuable information on designing siRNAs for silencing multiple murine CYPs, and potential pros and cons of using siRNA as a tool for repressing Cyp2d and estimating Cyp2d's role in murine xenobiotic metabolism. PMID:25124873
Competing coexisting phases in 2D water
Zanotti, Jean-Marc; Judeinstein, Patrick; Dalla-Bernardina, Simona; Creff, Gaëlle; Brubach, Jean-Blaise; Roy, Pascale; Bonetti, Marco; Ollivier, Jacques; Sakellariou, Dimitrios; Bellissent-Funel, Marie-Claire
2016-01-01
The properties of bulk water come from a delicate balance of interactions on length scales encompassing several orders of magnitudes: i) the Hydrogen Bond (HBond) at the molecular scale and ii) the extension of this HBond network up to the macroscopic level. Here, we address the physics of water when the three dimensional extension of the HBond network is frustrated, so that the water molecules are forced to organize in only two dimensions. We account for the large scale fluctuating HBond network by an analytical mean-field percolation model. This approach provides a coherent interpretation of the different events experimentally (calorimetry, neutron, NMR, near and far infra-red spectroscopies) detected in interfacial water at 160, 220 and 250 K. Starting from an amorphous state of water at low temperature, these transitions are respectively interpreted as the onset of creation of transient low density patches of 4-HBonded molecules at 160 K, the percolation of these domains at 220 K and finally the total invasion of the surface by them at 250 K. The source of this surprising behaviour in 2D is the frustration of the natural bulk tetrahedral local geometry and the underlying very significant increase in entropy of the interfacial water molecules. PMID:27185018
2D Radiative Processes Near Cloud Edges
NASA Technical Reports Server (NTRS)
Varnai, T.
2012-01-01
Because of the importance and complexity of dynamical, microphysical, and radiative processes taking place near cloud edges, the transition zone between clouds and cloud free air has been the subject of intense research both in the ASR program and in the wider community. One challenge in this research is that the one-dimensional (1D) radiative models widely used in both remote sensing and dynamical simulations become less accurate near cloud edges: The large horizontal gradients in particle concentrations imply that accurate radiative calculations need to consider multi-dimensional radiative interactions among areas that have widely different optical properties. This study examines the way the importance of multidimensional shortwave radiative interactions changes as we approach cloud edges. For this, the study relies on radiative simulations performed for a multiyear dataset of clouds observed over the NSA, SGP, and TWP sites. This dataset is based on Microbase cloud profiles as well as wind measurements and ARM cloud classification products. The study analyzes the way the difference between 1D and 2D simulation results increases near cloud edges. It considers both monochromatic radiances and broadband radiative heating, and it also examines the influence of factors such as cloud type and height, and solar elevation. The results provide insights into the workings of radiative processes and may help better interpret radiance measurements and better estimate the radiative impacts of this critical region.
Simulation of Yeast Cooperation in 2D.
Wang, M; Huang, Y; Wu, Z
2016-03-01
Evolution of cooperation has been an active research area in evolutionary biology in decades. An important type of cooperation is developed from group selection, when individuals form spatial groups to prevent them from foreign invasions. In this paper, we study the evolution of cooperation in a mixed population of cooperating and cheating yeast strains in 2D with the interactions among the yeast cells restricted to their small neighborhoods. We conduct a computer simulation based on a game theoretic model and show that cooperation is increased when the interactions are spatially restricted, whether the game is of a prisoner's dilemma, snow drifting, or mutual benefit type. We study the evolution of homogeneous groups of cooperators or cheaters and describe the conditions for them to sustain or expand in an opponent population. We show that under certain spatial restrictions, cooperator groups are able to sustain and expand as group sizes become large, while cheater groups fail to expand and keep them from collapse. PMID:26988702
Phase Engineering of 2D Tin Sulfides.
Mutlu, Zafer; Wu, Ryan J; Wickramaratne, Darshana; Shahrezaei, Sina; Liu, Chueh; Temiz, Selcuk; Patalano, Andrew; Ozkan, Mihrimah; Lake, Roger K; Mkhoyan, K A; Ozkan, Cengiz S
2016-06-01
Tin sulfides can exist in a variety of phases and polytypes due to the different oxidation states of Sn. A subset of these phases and polytypes take the form of layered 2D structures that give rise to a wide host of electronic and optical properties. Hence, achieving control over the phase, polytype, and thickness of tin sulfides is necessary to utilize this wide range of properties exhibited by the compound. This study reports on phase-selective growth of both hexagonal tin (IV) sulfide SnS2 and orthorhombic tin (II) sulfide SnS crystals with diameters of over tens of microns on SiO2 substrates through atmospheric pressure vapor-phase method in a conventional horizontal quartz tube furnace with SnO2 and S powders as the source materials. Detailed characterization of each phase of tin sulfide crystals is performed using various microscopy and spectroscopy methods, and the results are corroborated by ab initio density functional theory calculations. PMID:27099950
Ion Transport in 2-D Graphene Nanochannels
NASA Astrophysics Data System (ADS)
Xie, Quan; Foo, Elbert; Duan, Chuanhua
2015-11-01
Graphene membranes have recently attracted wide attention due to its great potential in water desalination and selective molecular sieving. Further developments of these membranes, including enhancing their mass transport rate and/or molecular selectivity, rely on the understanding of fundamental transport mechanisms through graphene membranes, which has not been studied experimentally before due to fabrication and measurement difficulties. Herein we report the fabrication of the basic constituent of graphene membranes, i.e. 2-D single graphene nanochannels (GNCs) and the study of ion transport in these channels. A modified bonding technique was developed to form GNCs with well-defined geometry and uniform channel height. Ion transport in such GNCs was studied using DC conductance measurement. Our preliminary results showed that the ion transport in GNCs is still governed by surface charge at low concentrations (10-6M to 10-4M). However, GNCs exhibits much higher ionic conductances than silica nanochannels with the same geometries in the surface-charge-governed regime. This conductance enhancement can be attributed to the pre-accumulation of charges on graphene surfaces. The work is supported by the Faculty Startup Fund (Boston University, USA).
Parallel map analysis on 2-D grids
Berry, M.; Comiskey, J.; Minser, K.
1993-12-31
In landscape ecology, computer modeling is used to assess habitat fragmentation and its ecological iMPLications. Specifically, maps (2-D grids) of habitat clusters must be analyzed to determine number, sizes and geometry of clusters. Models prior to this study relied upon sequential Fortran-77 programs which limited the sizes of maps and densities of clusters which could be analyzed. In this paper, we present more efficient computer models which can exploit recursion or parallelism. Significant improvements over the original Fortran-77 programs have been achieved using both recursive and nonrecursive C implementations on a variety of workstations such as the Sun Sparc 2, IBM RS/6000-350, and HP 9000-750. Parallel implementations on a 4096-processor MasPar MP-1 and a 32-processor CM-5 are also studied. Preliminary experiments suggest that speed improvements for the parallel model on the MasPar MP-1 (written in MPL) and on the CM-5 (written in C using CMMD) can be as much as 39 and 34 times faster, respectively, than the most efficient sequential C program on a Sun Sparc 2 for a 512 map. An important goal in this research effort is to produce a scalable map analysis algorithm for the identification and characterization of clusters for relatively large maps on massively-parallel computers.
2D Turbulence with Complicated Boundaries
NASA Astrophysics Data System (ADS)
Roullet, G.; McWilliams, J. C.
2014-12-01
We examine the consequences of lateral viscous boundary layers on the 2D turbulence that arises in domains with complicated boundaries (headlands, bays etc). The study is carried out numerically with LES. The numerics are carefully designed to ensure all global conservation laws, proper boundary conditions and a minimal range of dissipation scales. The turbulence dramatically differs from the classical bi-periodic case. Boundary layer separations lead to creation of many small vortices and act as a continuing energy source exciting the inverse cascade of energy throughout the domain. The detachments are very intermittent in time. In free decay, the final state depends on the effective numerical resolution: laminar with a single dominant vortex for low Re and turbulent with many vortices for large enough Re. After very long time, the turbulent end-state exhibits a striking tendency for the emergence of shielded vortices which then interact almost elastically. In the forced case, the boundary layers allow the turbulence to reach a statistical steady state without any artificial hypo-viscosity or other large-scale dissipation. Implications are discussed for the oceanic mesoscale and submesoscale turbulence.
Competing coexisting phases in 2D water
NASA Astrophysics Data System (ADS)
Zanotti, Jean-Marc; Judeinstein, Patrick; Dalla-Bernardina, Simona; Creff, Gaëlle; Brubach, Jean-Blaise; Roy, Pascale; Bonetti, Marco; Ollivier, Jacques; Sakellariou, Dimitrios; Bellissent-Funel, Marie-Claire
2016-05-01
The properties of bulk water come from a delicate balance of interactions on length scales encompassing several orders of magnitudes: i) the Hydrogen Bond (HBond) at the molecular scale and ii) the extension of this HBond network up to the macroscopic level. Here, we address the physics of water when the three dimensional extension of the HBond network is frustrated, so that the water molecules are forced to organize in only two dimensions. We account for the large scale fluctuating HBond network by an analytical mean-field percolation model. This approach provides a coherent interpretation of the different events experimentally (calorimetry, neutron, NMR, near and far infra-red spectroscopies) detected in interfacial water at 160, 220 and 250 K. Starting from an amorphous state of water at low temperature, these transitions are respectively interpreted as the onset of creation of transient low density patches of 4-HBonded molecules at 160 K, the percolation of these domains at 220 K and finally the total invasion of the surface by them at 250 K. The source of this surprising behaviour in 2D is the frustration of the natural bulk tetrahedral local geometry and the underlying very significant increase in entropy of the interfacial water molecules.
Competing coexisting phases in 2D water.
Zanotti, Jean-Marc; Judeinstein, Patrick; Dalla-Bernardina, Simona; Creff, Gaëlle; Brubach, Jean-Blaise; Roy, Pascale; Bonetti, Marco; Ollivier, Jacques; Sakellariou, Dimitrios; Bellissent-Funel, Marie-Claire
2016-01-01
The properties of bulk water come from a delicate balance of interactions on length scales encompassing several orders of magnitudes: i) the Hydrogen Bond (HBond) at the molecular scale and ii) the extension of this HBond network up to the macroscopic level. Here, we address the physics of water when the three dimensional extension of the HBond network is frustrated, so that the water molecules are forced to organize in only two dimensions. We account for the large scale fluctuating HBond network by an analytical mean-field percolation model. This approach provides a coherent interpretation of the different events experimentally (calorimetry, neutron, NMR, near and far infra-red spectroscopies) detected in interfacial water at 160, 220 and 250 K. Starting from an amorphous state of water at low temperature, these transitions are respectively interpreted as the onset of creation of transient low density patches of 4-HBonded molecules at 160 K, the percolation of these domains at 220 K and finally the total invasion of the surface by them at 250 K. The source of this surprising behaviour in 2D is the frustration of the natural bulk tetrahedral local geometry and the underlying very significant increase in entropy of the interfacial water molecules. PMID:27185018
2-D wavelet with position controlled resolution
NASA Astrophysics Data System (ADS)
Walczak, Andrzej; Puzio, Leszek
2005-09-01
Wavelet transformation localizes all irregularities in the scene. It is most effective in the case when intensities in the scene have no sharp details. It is the case often present in a medical imaging. To identify the shape one has to extract it from the scene as typical irregularity. When the scene does not contain sharp changes then common differential filters are not efficient tool for a shape extraction. The new 2-D wavelet for such task has been proposed. Described wavelet transform is axially symmetric and has varied scale in dependence on the distance from the centre of the wavelet symmetry. The analytical form of the wavelet has been presented as well as its application for details extraction in the scene. Most important feature of the wavelet transform is that it gives a multi-scale transformation, and if zoom is on the wavelet selectivity varies proportionally to the zoom step. As a result, the extracted shape does not change during zoom operation. What is more the wavelet selectivity can be fit to the local intensity gradient properly to obtain best extraction of the irregularities.
Siblings Are Special: Initial test of a New Approach for Preventing Youth Behavior Problems
Feinberg, Mark E.; Solmeyer, Anna R.; Hostetler, Michelle L.; Sakuma, Kari-Lyn; Jones, Damon; McHale, Susan M.
2012-01-01
Purpose A growing body of research documents the significance of siblings and sibling relationships for development, mental health, and behavioral risk across childhood and adolescence. Nonetheless, few well-designed efforts have been undertaken to promote positive and reduce negative youth outcomes by enhancing sibling relationships. Methods Based on a theoretical model of sibling influences, we conducted a randomized trial of Siblings Are Special, a group-format afterschool program for 5th graders with a younger sibling in 2nd through 4th grade, which entailed 12 weekly afterschool sessions and 3 Family Nights. We tested program efficacy with a pre-posttest design with 174 families randomly assigned to condition. In home visits at both time points we collected data via parent questionnaires, child interviews, and observer-rated videotaped interactions and teachers rated children’s behavior at school. Results The program enhanced positive sibling relationships, appropriate strategies for parenting siblings, and child self-control, social competence, and academic performance; program exposure was also associated with reduced maternal depression and child internalizing problems. Results were robust across the sample, not qualified by sibling gender, age, family demographics, or baseline risk. No effects were found for sibling conflict, collusion or child externalizing problems; we will examine follow-up data to determine if short-term impacts lead to reduced negative behaviors over time. Conclusions The breadth of the SAS program’s impact is consistent with research suggesting that siblings are an important influence on development and adjustment and supports our argument that a sibling focus should be incorporated into youth and family-oriented prevention programs. PMID:23298985
Multiscale simulation of 2D elastic wave propagation
NASA Astrophysics Data System (ADS)
Zhang, Wensheng; Zheng, Hui
2016-06-01
In this paper, we develop the multiscale method for simulation of elastic wave propagation. Based on the first-order velocity-stress hyperbolic form of 2D elastic wave equation, the particle velocities are solved first ona coarse grid by the finite volume method. Then the stress tensor is solved by using the multiscale basis functions which can represent the fine-scale variation of the wavefield on the coarse grid. The basis functions are computed by solving a local problem with the finite element method. The theoretical formulae and description of the multiscale method for elastic wave equation are given in more detail. The numerical computations for an inhomogeneous model with random scatter are completed. The results show the effectiveness of the multiscale method.
Advecting Procedural Textures for 2D Flow Animation
NASA Technical Reports Server (NTRS)
Kao, David; Pang, Alex; Moran, Pat (Technical Monitor)
2001-01-01
This paper proposes the use of specially generated 3D procedural textures for visualizing steady state 2D flow fields. We use the flow field to advect and animate the texture over time. However, using standard texture advection techniques and arbitrary textures will introduce some undesirable effects such as: (a) expanding texture from a critical source point, (b) streaking pattern from the boundary of the flowfield, (c) crowding of advected textures near an attracting spiral or sink, and (d) absent or lack of textures in some regions of the flow. This paper proposes a number of strategies to solve these problems. We demonstrate how the technique works using both synthetic data and computational fluid dynamics data.
The Anatomy of High-Performance 2D Similarity Calculations
Haque, Imran S.; Pande, Vijay S.
2011-01-01
Similarity measures based on the comparison of dense bit-vectors of two-dimensional chemical features are a dominant method in chemical informatics. For large-scale problems, including compound selection and machine learning, computing the intersection between two dense bit-vectors is the overwhelming bottleneck. We describe efficient implementations of this primitive, as well as example applications, using features of modern CPUs that allow 20-40x performance increases relative to typical code. Specifically, we describe fast methods for population count on modern x86 processors and cache-efficient matrix traversal and leader clustering algorithms that alleviate memory bandwidth bottlenecks in similarity matrix construction and clustering. The speed of our 2D comparison primitives is within a small factor of that obtained on GPUs, and does not require specialized hardware. PMID:21854053
Integrated Coupling of Surface and Subsurface Flow with HYDRUS-2D
NASA Astrophysics Data System (ADS)
Hartmann, Anne; Šimůnek, Jirka; Wöhling, Thomas; Schütze, Niels
2016-04-01
Describing interactions between surface and subsurface flow processes is important to adequately define water flow in natural systems. Since overland flow generation is highly influenced by rainfall and infiltration, both highly spatially heterogeneous processes, overland flow is unsteady and varies spatially. The prediction of overland flow needs to include an appropriate description of the interactions between the surface and subsurface flow. Coupling surface and subsurface water flow is a challenging task. Different approaches have been developed during the last few years, each having its own advantages and disadvantages. A new approach by Weill et al. (2009) to couple overland flow and subsurface flow based on a generalized Richards equation was implemented into the well-known subsurface flow model HYDRUS-2D (Šimůnek et al., 2011). This approach utilizes the one-dimensional diffusion wave equation to model overland flow. The diffusion wave model is integrated in HYDRUS-2D by replacing the terms of the Richards equation in a pre-defined runoff layer by terms defining the diffusion wave equation. Using this approach, pressure and flux continuity along the interface between both flow domains is provided. This direct coupling approach provides a strong coupling of both systems based on the definition of a single global system matrix to numerically solve the coupled flow problem. The advantage of the direct coupling approach, compared to the loosely coupled approach, is supposed to be a higher robustness, when many convergence problems can be avoided (Takizawa et al., 2014). The HYDRUS-2D implementation was verified using a) different test cases, including a direct comparison with the results of Weill et al. (2009), b) an analytical solution of the kinematic wave equation, and c) the results of a benchmark test of Maxwell et al. (2014), that included several known coupled surface subsurface flow models. Additionally, a sensitivity analysis evaluating the effects
2-D Animation's Not Just for Mickey Mouse.
ERIC Educational Resources Information Center
Weinman, Lynda
1995-01-01
Discusses characteristics of two-dimensional (2-D) animation; highlights include character animation, painting issues, and motion graphics. Sidebars present Silicon Graphics animations tools and 2-D animation programs for the desktop computer. (DGM)
NASA Technical Reports Server (NTRS)
Thakur, Siddarth; Wright, Jeffrey
2006-01-01
The traditional design and analysis practice for advanced propulsion systems, particularly chemical rocket engines, relies heavily on expensive full-scale prototype development and testing. Over the past decade, use of high-fidelity analysis and design tools such as CFD early in the product development cycle has been identified as one way to alleviate testing costs and to develop these devices better, faster and cheaper. Increased emphasis is being placed on developing and applying CFD models to simulate the flow field environments and performance of advanced propulsion systems. This necessitates the development of next generation computational tools which can be used effectively and reliably in a design environment by non-CFD specialists. A computational tool, called Loci-STREAM is being developed for this purpose. It is a pressure-based, Reynolds-averaged Navier-Stokes (RANS) solver for generalized unstructured grids, which is designed to handle all-speed flows (incompressible to hypersonic) and is particularly suitable for solving multi-species flow in fixed-frame combustion devices. Loci-STREAM integrates proven numerical methods for generalized grids and state-of-the-art physical models in a novel rule-based programming framework called Loci which allows: (a) seamless integration of multidisciplinary physics in a unified manner, and (b) automatic handling of massively parallel computing. The objective of the ongoing work is to develop a robust simulation capability for combustion problems in rocket engines. As an initial step towards validating this capability, a model problem is investigated in the present study which involves a gaseous oxygen/gaseous hydrogen (GO2/GH2) shear coaxial single element injector, for which experimental data are available. The sensitivity of the computed solutions to grid density, grid distribution, different turbulence models, and different near-wall treatments is investigated. A refined grid, which is clustered in the vicinity of
Generation and Radiation of Acoustic Waves from a 2D Shear Layer
NASA Technical Reports Server (NTRS)
Dahl, Milo D.
2000-01-01
A thin free shear layer containing an inflection point in the mean velocity profile is inherently unstable. Disturbances in the flow field can excite the unstable behavior of a shear layer, if the appropriate combination of frequencies and shear layer thicknesses exists, causing instability waves to grow. For other combinations of frequencies and thicknesses, these instability waves remain neutral in amplitude or decay in the downstream direction. A growing instability wave radiates noise when its phase velocity becomes supersonic relative to the ambient speed of sound. This occurs primarily when the mean jet flow velocity is supersonic. Thus, the small disturbances in the flow, which themselves may generate noise, have generated an additional noise source. It is the purpose of this problem to test the ability of CAA to compute this additional source of noise. The problem is idealized such that the exciting disturbance is a fixed known acoustic source pulsating at a single frequency. The source is placed inside of a 2D jet with parallel flow; hence, the shear layer thickness is constant. With the source amplitude small enough, the problem is governed by the following set of linear equations given in dimensional form.
ERIC Educational Resources Information Center
Erford, Bradley T.; Butler, Caitlin; Peacock, Elizabeth
2015-01-01
The Screening Test for Emotional Problems-Teacher Version (STEP-T) was designed to identify students aged 7-17 years with wide-ranging emotional disturbances. Coefficients alpha and test-retest reliability were adequate for all subscales except Anxiety. The hypothesized five-factor model fit the data very well and external aspects of validity were…
ERIC Educational Resources Information Center
Educational Testing Service, Princeton, NJ.
Four topics were emphasized during this conference on testing problems: (1) the selection of appropriate score scales for tests; (2) the experimental approach to the measurement of human motivation; (3) trends in public opinion polling since 1948 and their probable effects on predictions of the 1952 election; and (4) techniques for developing…
On 2D graphical representation of DNA sequence of nondegeneracy
NASA Astrophysics Data System (ADS)
Zhang, Yusen; Liao, Bo; Ding, Kequan
2005-08-01
Some two-dimensional (2D) graphical representations of DNA sequences have been given by Gates, Nandy, Leong and Mogenthaler, Randić, and Liao et al., which give visual characterizations of DNA sequences. In this Letter, we introduce a nondegeneracy 2D graphical representation of DNA sequence, which is different from Randić's novel 2D representation and Liao's 2D representation. We also present the nondegeneracy forms corresponding to the representations of Gates, Nandy, Leong and Mogenthaler.
Efficiency of Pareto joint inversion of 2D geophysical data using global optimization methods
NASA Astrophysics Data System (ADS)
Miernik, Katarzyna; Bogacz, Adrian; Kozubal, Adam; Danek, Tomasz; Wojdyła, Marek
2016-04-01
Pareto joint inversion of two or more sets of data is a promising new tool of modern geophysical exploration. In the first stage of our investigation we created software enabling execution of forward solvers of two geophysical methods (2D magnetotelluric and gravity) as well as inversion with possibility of constraining solution with seismic data. In the algorithm solving MT forward solver Helmholtz's equations, finite element method and Dirichlet's boundary conditions were applied. Gravity forward solver was based on Talwani's algorithm. To limit dimensionality of solution space we decided to describe model as sets of polygons, using Sharp Boundary Interface (SBI) approach. The main inversion engine was created using Particle Swarm Optimization (PSO) algorithm adapted to handle two or more target functions and to prevent acceptance of solutions which are non - realistic or incompatible with Pareto scheme. Each inversion run generates single Pareto solution, which can be added to Pareto Front. The PSO inversion engine was parallelized using OpenMP standard, what enabled execution code for practically unlimited amount of threads at once. Thereby computing time of inversion process was significantly decreased. Furthermore, computing efficiency increases with number of PSO iterations. In this contribution we analyze the efficiency of created software solution taking under consideration details of chosen global optimization engine used as a main joint minimization engine. Additionally we study the scale of possible decrease of computational time caused by different methods of parallelization applied for both forward solvers and inversion algorithm. All tests were done for 2D magnetotelluric and gravity data based on real geological media. Obtained results show that even for relatively simple mid end computational infrastructure proposed solution of inversion problem can be applied in practice and used for real life problems of geophysical inversion and interpretation.
Testing a Comprehensive Community Problem-Solving Framework for Community Coalitions
ERIC Educational Resources Information Center
Yang, Evelyn; Foster-Fishman, Pennie; Collins, Charles; Ahn, Soyeon
2012-01-01
Community problem solving is believed to help coalitions achieve community changes and subsequent population-level reductions in targeted community health problems. This study empirically examined a community problem solving model used by CADCA, a national coalition training organization, to determine if the model explains how coalitions become…
NASA Technical Reports Server (NTRS)
Kapoor, Kamlesh; Anderson, Bernhard H.; Shaw, Robert J.
1994-01-01
A two-dimensional computational code, PRLUS2D, which was developed for the reactive propulsive flows of ramjets and scramjets, was validated for two-dimensional shock-wave/turbulent-boundary-layer interactions. The problem of compression corners at supersonic speeds was solved using the RPLUS2D code. To validate the RPLUS2D code for hypersonic speeds, it was applied to a realistic hypersonic inlet geometry. Both the Baldwin-Lomax and the Chien two-equation turbulence models were used. Computational results showed that the RPLUS2D code compared very well with experimentally obtained data for supersonic compression corner flows, except in the case of large separated flows resulting from the interactions between the shock wave and turbulent boundary layer. The computational results compared well with the experiment results in a hypersonic NASA P8 inlet case, with the Chien two-equation turbulence model performing better than the Baldwin-Lomax model.
Homogenization models for 2-D grid structures
NASA Technical Reports Server (NTRS)
Banks, H. T.; Cioranescu, D.; Rebnord, D. A.
1992-01-01
In the past several years, we have pursued efforts related to the development of accurate models for the dynamics of flexible structures made of composite materials. Rather than viewing periodicity and sparseness as obstacles to be overcome, we exploit them to our advantage. We consider a variational problem on a domain that has large, periodically distributed holes. Using homogenization techniques we show that the solution to this problem is in some topology 'close' to the solution of a similar problem that holds on a much simpler domain. We study the behavior of the solution of the variational problem as the holes increase in number, but decrease in size in such a way that the total amount of material remains constant. The result is an equation that is in general more complex, but with a domain that is simply connected rather than perforated. We study the limit of the solution as the amount of material goes to zero. This second limit will, in most cases, retrieve much of the simplicity that was lost in the first limit without sacrificing the simplicity of the domain. Finally, we show that these results can be applied to the case of a vibrating Love-Kirchhoff plate with Kelvin-Voigt damping. We rely heavily on earlier results of (Du), (CS) for the static, undamped Love-Kirchhoff equation. Our efforts here result in a modification of those results to include both time dependence and Kelvin-Voigt damping.
Testing Wind as an Explanation for the Spin Problem in the Continuum-fitting Method
NASA Astrophysics Data System (ADS)
You, Bei; Straub, Odele; Czerny, Bożena; Sobolewska, Małgosia; Różańska, Agata; Bursa, Michal; Dovčiak, Michal
2016-04-01
The continuum-fitting method is one of the two most advanced methods of determining the black hole spin in accreting X-ray binary systems. There are, however, still some unresolved issues with the underlying disk models. One of these issues manifests as an apparent decrease in spin for increasing source luminosity. Here, we perform a few simple tests to establish whether outflows from the disk close to the inner radius can address this problem. We employ four different parametric models to describe the wind and compare these to the apparent decrease in spin with luminosity measured in the sources LMC X-3 and GRS 1915+105. Wind models in which parameters do not explicitly depend on the accretion rate cannot reproduce the spin measurements. Models with mass accretion rate dependent outflows, however, have spectra that emulate the observed ones. The assumption of a wind thus effectively removes the artifact of spin decrease. This solution is not unique; the same conclusion can be obtained using a truncated inner disk model. To distinguish among the valid models, we will need high-resolution X-ray data and a realistic description of the Comptonization in the wind.
A Planar Quantum Transistor Based on 2D-2D Tunneling in Double Quantum Well Heterostructures
Baca, W.E.; Blount, M.A.; Hafich, M.J.; Lyo, S.K.; Moon, J.S.; Reno, J.L.; Simmons, J.A.; Wendt, J.R.
1998-12-14
We report on our work on the double electron layer tunneling transistor (DELTT), based on the gate-control of two-dimensional -- two-dimensional (2D-2D) tunneling in a double quantum well heterostructure. While previous quantum transistors have typically required tiny laterally-defined features, by contrast the DELTT is entirely planar and can be reliably fabricated in large numbers. We use a novel epoxy-bond-and-stop-etch (EBASE) flip-chip process, whereby submicron gating on opposite sides of semiconductor epitaxial layers as thin as 0.24 microns can be achieved. Because both electron layers in the DELTT are 2D, the resonant tunneling features are unusually sharp, and can be easily modulated with one or more surface gates. We demonstrate DELTTs with peak-to-valley ratios in the source-drain I-V curve of order 20:1 below 1 K. Both the height and position of the resonant current peak can be controlled by gate voltage over a wide range. DELTTs with larger subband energy offsets ({approximately} 21 meV) exhibit characteristics that are nearly as good at 77 K, in good agreement with our theoretical calculations. Using these devices, we also demonstrate bistable memories operating at 77 K. Finally, we briefly discuss the prospects for room temperature operation, increases in gain, and high-speed.
New Approach for 2D Readout of GEM Detectors
Hasell, Douglas K
2011-10-29
Detectors based on Gas Electron Multiplication (GEM) technology are becoming more and more widely used in nuclear and high energy physics and are being applied in astronomy, medical physics, industry, and homeland security. GEM detectors are thin, low mass, insensitive to magnetic fields, and can currently provide position resolutions down to {approx}50 microns. However, the designs for reconstructing the position, in two dimensions (2D), of the charged particles striking a GEM detector are often complicated to fabricate and expensive. The objective of this proposal is to investigate a simpler procedure for producing the two dimensional readout layer of GEM detectors using readily available printed circuit board technology which can be tailored to the detector requirements. We will use the established GEM laboratory and facilities at M.I.T. currently employed in developing GEM detectors for the STAR forward tracking upgrade to simplify the testing and evaluation of the new 2D readout designs. If this new design proves successful it will benefit future nuclear and high energy physics experiments already being planned and will similarly extend and simplify the application of GEM technology to other branches of science, medicine, and industry. These benefits would be not only in lower costs for fabrication but also it increased flexibility for design and application.
The effects of aging on haptic 2D shape recognition.
Overvliet, Krista E; Wagemans, J; Krampe, Ralf T
2013-12-01
We use the image-mediation model (Klatzky & Lederman, 1987) as a framework to investigate potential sources of adult age differences in the haptic recognition of two-dimensional (2D) shapes. This model states that the low-resolution, temporally sequential, haptic input is translated into a visual image, which is then reperceived through the visual processors, before it is matched against a long-term memory representation and named. In three experiments we tested groups of 12 older (mean age 73.11) and three groups of 12 young adults (mean age 22.80) each. In Experiment 1 we confirm age-related differences in haptic 2D shape recognition, and we show the typical age × complexity interaction. In Experiment 2 we show that if we facilitate the visual translation process, age differences become smaller, but only with simple shapes and not with the more complex everyday objects. In Experiment 3 we target the last step in the model (matching and naming) for complex stimuli. We found that age differences in exploration time were considerably reduced when this component process was facilitated by providing a category name. We conclude that the image-mediation model can explain adult-age differences in haptic recognition, particularly if the role of working memory in forming the transient visual image is considered. Our findings suggest that sensorimotor skills thought to rely on peripheral processes for the most part are critically constrained by age-related changes in central processing capacity in later adulthood. PMID:23978010
Facial biometrics based on 2D vector geometry
NASA Astrophysics Data System (ADS)
Malek, Obaidul; Venetsanopoulos, Anastasios; Androutsos, Dimitrios
2014-05-01
The main challenge of facial biometrics is its robustness and ability to adapt to changes in position orientation, facial expression, and illumination effects. This research addresses the predominant deficiencies in this regard and systematically investigates a facial authentication system in the Euclidean domain. In the proposed method, Euclidean geometry in 2D vector space is being constructed for features extraction and the authentication method. In particular, each assigned point of the candidates' biometric features is considered to be a 2D geometrical coordinate in the Euclidean vector space. Algebraic shapes of the extracted candidate features are also computed and compared. The proposed authentication method is being tested on images from the public "Put Face Database". The performance of the proposed method is evaluated based on Correct Recognition (CRR), False Acceptance (FAR), and False Rejection (FRR) rates. The theoretical foundation of the proposed method along with the experimental results are also presented in this paper. The experimental results demonstrate the effectiveness of the proposed method.
ELRIS2D: A MATLAB Package for the 2D Inversion of DC Resistivity/IP Data
NASA Astrophysics Data System (ADS)
Akca, Irfan
2016-04-01
ELRIS2D is an open source code written in MATLAB for the two-dimensional inversion of direct current resistivity (DCR) and time domain induced polarization (IP) data. The user interface of the program is designed for functionality and ease of use. All available settings of the program can be reached from the main window. The subsurface is discretized using a hybrid mesh generated by the combination of structured and unstructured meshes, which reduces the computational cost of the whole inversion procedure. The inversion routine is based on the smoothness constrained least squares method. In order to verify the program, responses of two test models and field data sets were inverted. The models inverted from the synthetic data sets are consistent with the original test models in both DC resistivity and IP cases. A field data set acquired in an archaeological site is also used for the verification of outcomes of the program in comparison with the excavation results.
Mobley, Michael; Chun, Heejung
2013-04-01
Based on Jessor's problem behavior theory (PBT; R. Jessor, 1987, Problem-behavior theory, psychosocial development, and adolescent problem drinking, British Journal of Addiction, Vol. 82, pp. 331-342), the comparability of a second-order problem behavior model (SPBM) was investigated employing structural equation modeling (SEM) and latent mean differences in problem behavior engagement were examined among racial/ethnic adolescents. Within a span of nearly 25 years, this study represents the first nationally representative sample of Latino and African American adolescents utilized in testing Jessor's PBT and problem behavior syndrome (PBS). Using a sample of 5,831 Latino, African American, and European American adolescents drawn from the National Longitudinal Study of Adolescent Health, a series of invariance tests evidenced support for Jessor's PBT and PBS. Latent mean difference test results evidenced significant differences in problem behaviors (e.g., academic failure [AF], aggression [AG], substance use [SU], and risky sexual activity[RSA]) across racial/ethnic adolescent groups, which could be explained partially by PBS. A discussion of findings, limitations, and recommendations for future research is presented. PMID:23647329
Generation and Radiation of Acoustic Waves from a 2-D Shear Layer
NASA Technical Reports Server (NTRS)
Agarwal, Anurag; Morris, Philip J.
2000-01-01
A parallel numerical simulation of the radiation of sound from an acoustic source inside a 2-D jet is presented in this paper. This basic benchmark problem is used as a test case for scattering problems that are presently being solved by using the Impedance Mismatch Method (IMM). In this technique, a solid body in the domain is represented by setting the acoustic impedance of each medium, encountered by a wave, to a different value. This impedance discrepancy results in reflected and scattered waves with appropriate amplitudes. The great advantage of the use of this method is that no modifications to a simple Cartesian grid need to be made for complicated geometry bodies. Thus, high order finite difference schemes may be applied simply to all parts of the domain. In the IMM, the total perturbation field is split into incident and scattered fields. The incident pressure is assumed to be known and the equivalent sources for the scattered field are associated with the presence of the scattering body (through the impedance mismatch) and the propagation of the incident field through a non-uniform flow. An earlier version of the technique could only handle uniform flow in the vicinity of the source and at the outflow boundary. Scattering problems in non-uniform mean flow are of great practical importance (for example, scattering from a high lift device in a non-uniform mean flow or the effects of a fuselage boundary layer). The solution to this benchmark problem, which has an acoustic wave propagating through a non-uniform mean flow, serves as a test case for the extensions of the IMM technique.
Correlated Electron Phenomena in 2D Materials
NASA Astrophysics Data System (ADS)
Lambert, Joseph G.
In this thesis, I present experimental results on coherent electron phenomena in layered two-dimensional materials: single layer graphene and van der Waals coupled 2D TiSe2. Graphene is a two-dimensional single-atom thick sheet of carbon atoms first derived from bulk graphite by the mechanical exfoliation technique in 2004. Low-energy charge carriers in graphene behave like massless Dirac fermions, and their density can be easily tuned between electron-rich and hole-rich quasiparticles with electrostatic gating techniques. The sharp interfaces between regions of different carrier densities form barriers with selective transmission, making them behave as partially reflecting mirrors. When two of these interfaces are set at a separation distance within the phase coherence length of the carriers, they form an electronic version of a Fabry-Perot cavity. I present measurements and analysis of multiple Fabry-Perot modes in graphene with parallel electrodes spaced a few hundred nanometers apart. Transition metal dichalcogenide (TMD) TiSe2 is part of the family of materials that coined the term "materials beyond graphene". It contains van der Waals coupled trilayer stacks of Se-Ti-Se. Many TMD materials exhibit a host of interesting correlated electronic phases. In particular, TiSe2 exhibits chiral charge density waves (CDW) below TCDW ˜ 200 K. Upon doping with copper, the CDW state gets suppressed with Cu concentration, and CuxTiSe2 becomes superconducting with critical temperature of T c = 4.15 K. There is still much debate over the mechanisms governing the coexistence of the two correlated electronic phases---CDW and superconductivity. I will present some of the first conductance spectroscopy measurements of proximity coupled superconductor-CDW systems. Measurements reveal a proximity-induced critical current at the Nb-TiSe2 interfaces, suggesting pair correlations in the pure TiSe2. The results indicate that superconducting order is present concurrently with CDW in
Wang, Yuxian; Xie, Yongbing; Sun, Hongqi; Xiao, Jiadong; Cao, Hongbin; Wang, Shaobin
2016-01-15
Two-dimensional reduced graphene oxide (2D rGO) was employed as both a shape-directing medium and support to fabricate 2D γ-MnO2/2D rGO nano-hybrids (MnO2/rGO) via a facile hydrothermal route. For the first time, the 2D/2D hybrid materials were used for catalytic ozonation of 4-nitrophenol. The catalytic efficiency of MnO2/rGO was much higher than either MnO2 or rGO only, and rGO was suggested to play the role for promoting electron transfers. Quenching tests using tert-butanol, p-benzoquinone, and sodium azide suggested that the major radicals responsible for 4-nitrophenol degradation and mineralization are O2(-) and (1)O2, but not ·OH. Reusability tests demonstrated a high stability of the materials in catalytic ozonation with minor Mn leaching below 0.5 ppm. Degradation mechanism, reaction kinetics, reusability and a synergistic effect between catalytic ozonation and coupling peroxymonosulfate (PMS) activation were also discussed. PMID:26342576
NASA Astrophysics Data System (ADS)
Bernauer, F.; Hürkamp, K.; Rühm, W.; Tschiersch, J.
2015-08-01
Detailed characterization and classification of precipitation is an important task in atmospheric research. Line scanning 2-D video disdrometer devices are well established for rain observations. The two orthogonal views taken of each hydrometeor passing the sensitive area of the instrument qualify these devices especially for detailed characterization of nonsymmetric solid hydrometeors. However, in case of solid precipitation, problems related to the matching algorithm have to be considered and the user must be aware of the limited spatial resolution when size and shape descriptors are analyzed. Clarifying the potential of 2-D video disdrometers in deriving size, velocity and shape parameters from single recorded pictures is the aim of this work. The need of implementing a matching algorithm suitable for mixed- and solid-phase precipitation is highlighted as an essential step in data evaluation. For this purpose simple reproducible experiments with solid steel spheres and irregularly shaped Styrofoam particles are conducted. Self-consistency of shape parameter measurements is tested in 38 cases of real snowfall. As a result, it was found that reliable size and shape characterization with a relative standard deviation of less than 5 % is only possible for particles larger than 1 mm. For particles between 0.5 and 1.0 mm the relative standard deviation can grow up to 22 % for the volume, 17 % for size parameters and 14 % for shape descriptors. Testing the adapted matching algorithm with a reproducible experiment with Styrofoam particles, a mismatch probability of less than 3 % was found. For shape parameter measurements in case of real solid-phase precipitation, the 2-DVD shows self-consistent behavior.
Establishment of CYP2D6 reference samples by multiple validated genotyping platforms.
Fang, H; Liu, X; Ramírez, J; Choudhury, N; Kubo, M; Im, H K; Konkashbaev, A; Cox, N J; Ratain, M J; Nakamura, Y; O'Donnell, P H
2014-12-01
Cytochrome P450 2D6 (cytochrome P450, family 2, subfamily D, polypeptide 6 (CYP2D6)), a highly polymorphic drug-metabolizing enzyme, is involved in the metabolism of one-quarter of the most commonly prescribed medications. Here we have applied multiple genotyping methods and Sanger sequencing to assign precise and reproducible CYP2D6 genotypes, including copy numbers, for 48 HapMap samples. Furthermore, by analyzing a set of 50 human liver microsomes using endoxifen formation from N-desmethyl-tamoxifen as the phenotype of interest, we observed a significant positive correlation between CYP2D6 genotype-assigned activity score and endoxifen formation rate (rs = 0.68 by rank correlation test, P = 5.3 × 10(-8)), which corroborated the genotype-phenotype prediction derived from our genotyping methodologies. In the future, these 48 publicly available HapMap samples characterized by multiple substantiated CYP2D6 genotyping platforms could serve as a reference resource for assay development, validation, quality control and proficiency testing for other CYP2D6 genotyping projects and for programs pursuing clinical pharmacogenomic testing implementation. PMID:24980783
Establishment of CYP2D6 Reference Samples by Multiple Validated Genotyping Platforms
Fang, Hua; Liu, Xiao; Ramírez, Jacqueline; Choudhury, Noura; Kubo, Michiaki; Im, Hae Kyung; Konkashbaev, Anuar; Cox, Nancy J.; Ratain, Mark J.; Nakamura, Yusuke; O’Donnell, Peter H.
2014-01-01
Cytochrome P450 2D6 (cytochrome P450, family 2, subfamily D, polypeptide 6, or CYP2D6), a highly polymorphic drug metabolizing enzyme, is involved in the metabolism of one quarter of the most commonly prescribed medications. Here, we have applied multiple genotyping methods and Sanger sequencing to assign precise and reproducible CYP2D6 genotypes, including copy numbers, for 48 HapMap samples. Furthermore, by analyzing a set of 50 human liver microsomes using endoxifen formation from N-desmethyl-tamoxifen as the phenotype of interest, we observed a significant positive correlation between CYP2D6 genotype-assigned activity score and endoxifen formation rate (rs = 0.68 by Rank correlation test, P = 5.3 ×10−8), which corroborated the genotype-phenotype prediction derived from our genotyping methodologies. In the future, these 48 publicly available HapMap samples characterized by multiple substantiated CYP2D6 genotyping platforms could serve as a reference resource for assay development, validation, quality control, and proficiency testing for other CYP2D6 genotyping projects, and for programs pursuing clinical pharmacogenomic testing implementation. PMID:24980783
Concordance between CYP2D6 genotypes obtained from tumor-derived and germline DNA.
Rae, James M; Regan, Meredith M; Thibert, Jacklyn N; Gersch, Christina; Thomas, Dafydd; Leyland-Jones, Brian; Viale, Giuseppe; Pusztai, Lajos; Hayes, Daniel F; Skaar, Todd; Van Poznak, Catherine
2013-09-01
Formalin-fixed, paraffin-embedded tumors (FFPETs) are a valuable source of DNA for genotype association studies and are often the only germline DNA resource from cancer clinical trials. The anti-estrogen tamoxifen is metabolized into endoxifen by CYP2D6, leading to the hypothesis that patients with certain CYP2D6 genotypes may not receive benefit because of their inability to activate the drug. Studies testing this hypothesis using FFPETs have provided conflicting results. It has been postulated that CYP2D6 genotype determined using FFPET may not be accurate because of somatic tumor alterations. In this study, we determined the concordance between CYP2D6 genotypes generated using 3 tissue sources (FFPETs; formalin-fixed, paraffin-embedded unaffected lymph nodes [FFPELNs]; and whole blood cells [WBCs]) from 122 breast cancer patients. Compared with WBCs, FFPET and FFPELN genotypes were highly concordant (>94%), as were the predicted CYP2D6 metabolic phenotypes (>97%). We conclude that CYP2D6 genotypes obtained from FFPETs accurately represent the patient's CYP2D6 metabolic phenotype. PMID:23958736
Impact of CYP2D*6 in the adjuvant treatment of breast cancer patients with tamoxifen.
Markopoulos, Christos; Kykalos, Stylianos; Mantas, Dimitrios
2014-08-10
Biotransformation of tamoxifen to the potent antiestrogen endoxifen is performed by cytochrome P450 (CYP) enzymes, in particular the CYP2D6 isoform. CYP2D6*4 is one of the most frequent alleles associated with loss of enzymatic activity. The incidence of CYP2D6*4 among Caucasians is estimated up to 27%, while it is present in up to 90% of all poor metabolizers within the Caucasian population. The hypothesis under question is whether the presence of one or two non-functioning (null) alleles predicts an inferior outcome in postmenopausal women with breast cancer receiving adjuvant treatment with tamoxifen. The numerous existing studies investigating the association of CYP2D6 with treatment failure in breast cancer are inconsistent and give rather conflicting results. Currently, routine CYP2D6 testing among women with breast cancer is not recommended and the significance of CYP2D6 phenotype in decision making regarding the administration of tamoxifen is unclear. The present study summarizes current literature regarding clinical studies on CYP2D6*4, particularly in terms of response to tamoxifen therapy and breast cancer outcome. PMID:25114852
Impact of CYP2D*6 in the adjuvant treatment of breast cancer patients with tamoxifen
Markopoulos, Christos; Kykalos, Stylianos; Mantas, Dimitrios
2014-01-01
Biotransformation of tamoxifen to the potent antiestrogen endoxifen is performed by cytochrome P450 (CYP) enzymes, in particular the CYP2D6 isoform. CYP2D6*4 is one of the most frequent alleles associated with loss of enzymatic activity. The incidence of CYP2D6*4 among Caucasians is estimated up to 27%, while it is present in up to 90% of all poor metabolizers within the Caucasian population. The hypothesis under question is whether the presence of one or two non-functioning (null) alleles predicts an inferior outcome in postmenopausal women with breast cancer receiving adjuvant treatment with tamoxifen. The numerous existing studies investigating the association of CYP2D6 with treatment failure in breast cancer are inconsistent and give rather conflicting results. Currently, routine CYP2D6 testing among women with breast cancer is not recommended and the significance of CYP2D6 phenotype in decision making regarding the administration of tamoxifen is unclear. The present study summarizes current literature regarding clinical studies on CYP2D6*4, particularly in terms of response to tamoxifen therapy and breast cancer outcome. PMID:25114852
Measurements of Schottky barrier heights formed from metals and 2D transition metal dichalcogedides
NASA Astrophysics Data System (ADS)
Kim, Changsik; Moon, Inyong; Nam, Seunggeol; Cho, Yeonchoo; Shin, Hyeon-Jin; Park, Seongjun; Yoo, Won Jong
Schottky barrier height (SBH) is an important parameter that needs to be considered for designing electronic devices. However, for two dimensional (2D) materials based devices, SBH control is limited by 2D structure induced quantum confinement and 2D surface induced Fermi level pinning. In this work, we explore differences in measuring SBH between 2D and 3D materials. Recently, low temperature I-V measurement has been reported to extract SBH based on thermionic emission equation for Schottky diode. However, 2D devices are not real Schottky diode in that both source and drain metal electrodes make Schottky contact. According to our experimental results, SBH extracted from linear slope of ln (I/T3/2) against 1/T show widely diverse values, dependent on applied voltage bias and tested temperature which affect carrier transport including tunneling or thermionic emission across the metal-2D material interface. In this work, we wish to demonstrate the method to determine SBH and Fermi level pinning which are attributed to 2D transition metal dichalcogedides, differently from conventional 3D materials. .
Finger length ratios (2D:4D) in anthropoids implicate reduced prenatal androgens in social bonding.
Nelson, Emma; Shultz, Susanne
2010-03-01
The second-to-fourth digit ratio (2D:4D) has been proposed as a biomarker reflecting prenatal androgen effects (PAE), such that individuals with lower ratios have experienced higher PAE than those with higher ratios. 2D:4D has been correlated with a number of sex-linked traits in humans such as aggression, promiscuity, and competitiveness. In addition, polygynous societies reportedly have lower 2D:4D (higher PAE) than more monogamous populations. This evidence suggests that PAE may be implicated in the development of sexually selected behaviors in humans. To place 2D:4D research into a broader context, we test the relationship between digit ratios and behavior across nonhuman anthropoids; polygynous species, with higher levels of intrasexual competition, should have more pronounced markers of PAE (lower 2D:4D) than pair-bonded species. Our results accord with those found in humans: 2D:4D is lower in polygynous species and higher (lower PAE) in pair-bonded species. Old World monkeys have low, and relatively invariant 2D:4D (high PAE), which is coupled with high levels of intrasexual competition. This contrasts with higher and more variable ratios in both great apes and New World monkeys. In addition, both male and female ratios decrease with increasing levels of intrasexual competition. Human ratios are intermediate between pair-bonded and more promiscuous hominoids. We propose that PAE may be involved in promoting species characteristic social behavior in anthropoids. PMID:19862809
Corbin, William R.; Iwamoto, Derek K.; Fromme, Kim
2011-01-01
Objective: According to the acquired preparedness model (APM), personality traits related to disinhibition (i.e., impulsivity and sensation seeking) may influence the learning process, contributing to individual differences in cognitions (e.g., expectations about outcomes) that may contribute to engagement in and consequences of risk behaviors, including alcohol use. Although there is strong support for the APM, longitudinal studies have involved short-term follow-ups, and the relevance of the APM for alcohol-related consequences has not been clearly established. Method: Participants were 2,245 (59.9% female) incoming freshmen who completed the first of eight web-based surveys during the summer before college matriculation. Structural equation modeling was used to test a comprehensive longitudinal APM for both alcohol use and related consequences. Multigroup models were used to examine measurement and structural invariance by gender. Results: Positive (but not negative) alcohol expectancies during freshman year of college partially mediated the relation between senior year of high school disinhibition and both alcohol use and related problems during the fourth year of college, and multigroup models suggested that the relationships proposed in the APM operated similarly for women and men. Conclusions: This study demonstrates the temporal relations proposed in the APM across a longer period (4 years) than in previous studies among a large sample of ethnically diverse students. Further, the results are the first to validate the APM with respect to drinking consequences while controlling for levels of alcohol use. The results lend support for brief interventions targeting positive alcohol expectancies, particularly for individuals high in trait disinhibition. PMID:21683042
2-D Finite Element Heat Conduction
1989-10-30
AYER is a finite element program which implicitly solves the general two-dimensional equation of thermal conduction for plane or axisymmetric bodies. AYER takes into account the effects of time (transient problems), in-plane anisotropic thermal conductivity, a three-dimensional velocity distribution, and interface thermal contact resistance. Geometry and material distributions are arbitrary, and input is via subroutines provided by the user. As a result, boundary conditions, material properties, velocity distributions, and internal power generation may be mademore » functions of, e.g., time, temperature, location, and heat flux.« less
Grazing incidence neutron diffraction from large scale 2D structures
Toperverg, B. P.; Felcher, G. P.; Metlushko, V. V.; Leiner, V.; Siebrecht, R.; Nikonov, O.
2000-01-13
The distorted wave Born approximation (DWBA) is applied to evaluate the diffraction pattern of neutrons (or X-rays) from a 2D array of dots deposited onto a dissimilar substrate. With the radiation impinging on the surface at a grazing incidence angle {alpha}, the intensities diffracted both in and out the plane of specular reflection are calculated as a function of the periodicity of the array, height and diameter of the dots. The results are presented in the form of diffracted intensity contours in a plane with coordinates {alpha} and {alpha}{prime}, the latter being the glancing angle of scattering. The optimization of the experimental conditions for polarized neutron experiments on submicron dots is discussed. The feasibility of such measurements is confirmed by a test experiment.
Constriant inversion of 2D magnetotelluric data with anisotropic conductivities
NASA Astrophysics Data System (ADS)
Chen, X.; Weckmann, U.
2011-12-01
Within the framework of the German - South African geo-scientific research initiative Inkaba yeAfrica a series of magnetotelluric (MT) field experiments were conducted along the Agulhas-Karoo Transect in South Africa. This transect crosses several continental collision zones between the Cape Fold Belt, the Namaqua Natal Mobile Belt and the Kaapvaal Craton. Along the Cape Fold Belt (CFB) profile we can identify areas (>10 km) where MT sites exhibit phases over 90°. This phenomenon usually occurs in presence of electrical anisotropy. Due to the dense site spacing we are able to observe this behaviour consistently at several sites. The anisotropy of electrical conductivity is essentially a scale effect: Even if the conductivity is isotropic on the micro scale, it will become anisotropic on a larger scale if, in the averaging volume, preferred orientation (e.g., layering or lamination) exist. Therefore, it is necessary to understand the electrical anisotropy in more details and furthermore electrical anisotropy offers new degrees of freedom, which should allow a better interpretation of data. In 2D MT case with considering of electrical anisotropy, computing of impedance tensor requires two independent electric field solutions computed for two different source polarisations. Based on the forward problem formulation and its numerical approximation we derive partial differential equations for the sensitivities of the magnetotelluric fields with respect to the elements of the conductivity tensor within the medium. For illustration a sensitivity study for a simple synthetic model is shown. We present an algorithm for the inversion of 2D magnetotelluric data with anisotropic conductivities which is a extension of the well-known NLCG minimization algorithm to anisotropic model. To constrain the structure complexity, a penalty function consists of datamisfit, standard model roughness and quadratic variation of the conductivity tensor elements is minimized. To demonstrate the
Quantum Simulation with 2D Arrays of Trapped Ions
NASA Astrophysics Data System (ADS)
Richerme, Philip
2016-05-01
The computational difficulty of solving fully quantum many-body spin problems is a significant obstacle to understanding the behavior of strongly correlated quantum matter. This work proposes the design and construction of a 2D quantum spin simulator to investigate the physics of frustrated materials, highly entangled states, mechanisms potentially underpinning high-temperature superconductivity, and other topics inaccessible to current 1D systems. The effective quantum spins will be encoded within the well-isolated electronic levels of trapped ions, confined in a two-dimensional planar geometry, and made to interact using phonon-mediated optical dipole forces. The system will be scalable to 100+ quantum particles, far beyond the realm of classical intractability, while maintaining individual-ion control, long quantum coherence times, and site-resolved projective spin measurements. Once constructed, the two-dimensional quantum simulator will implement a broad range of spin models on a variety of reconfigurable lattices and characterize their behavior through measurements of spin-spin correlations and entanglement. This versatile tool will serve as an important experimental resource for exploring difficult quantum many-body problems in a regime where classical methods fail.
NASA Technical Reports Server (NTRS)
Banks, H. T.; Kojima, Fumio
1988-01-01
The identification of the geometrical structure of the system boundary for a two-dimensional diffusion system is reported. The domain identification problem treated here is converted into an optimization problem based on a fit-to-data criterion and theoretical convergence results for approximate identification techniques are discussed. Results of numerical experiments to demonstrate the efficacy of the theoretical ideas are reported.
Mechanical properties of 2D and 3D braided textile composites
NASA Technical Reports Server (NTRS)
Norman, Timothy L.
1991-01-01
The purpose of this research was to determine the mechanical properties of 2D and 3D braided textile composite materials. Specifically, those designed for tension or shear loading were tested under static loading to failure to investigate the effects of braiding. The overall goal of the work was to provide a structural designer with an idea of how textile composites perform under typical loading conditions. From test results for unnotched tension, it was determined that the 2D is stronger, stiffer, and has higher elongation to failure than the 3D. It was also found that the polyetherether ketone (PEEK) resin system was stronger, stiffer, and had higher elongation at failure than the resin transfer molding (RTM) epoxy. Open hole tension tests showed that PEEK resin is more notch sensitive than RTM epoxy. Of greater significance, it was found that the 3D is less notch sensitive than the 2D. Unnotched compression tests indicated, as did the tension tests, that the 2D is stronger, stiffer, and has higher elongation at failure than the RTM epoxy. The most encouraging results were from compression after impact. The 3D braided composite showed a compression after impact failure stress equal to 92 percent of the unimpacted specimen. The 2D braided composite failed at about 67 percent of the unimpacted specimen. Higher damage tolerance is observed in textiles over conventional composite materials. This is observed in the results, especially in the 3D braided materials.
Silicene: silicon conquers the 2D world
NASA Astrophysics Data System (ADS)
Le Lay, Guy; Salomon, Eric; Angot, Thierry
2016-01-01
We live in the digital age based on the silicon chip and driven by Moore's law. Last July, IBM created a surprise by announcing the fabrication of a 7 nm test chip with functional transistors using, instead of just silicon, a silicon-germanium alloy. Will silicon be dethroned?
ERIC Educational Resources Information Center
Leahy, Wayne; Hanham, José; Sweller, John
2015-01-01
The testing effect occurs when learners who are tested rather than relearning material perform better on a final test than those who relearn. Based on cognitive load theory, it was predicted that the testing effect may not be obtained when the material being learned is high in element interactivity. Three experiments investigated conditions of the…
Solutions for Some Technical Problems in Domain-Referenced Mastery Testing. Final Report.
ERIC Educational Resources Information Center
Huynh, Huynh; Saunders, Joseph C.
A basic technical framework is provided for the design and use of mastery tests. The Mastery Testing Project (MTP) prepared this framework using advanced mathematics supplemented with computer simulation based on real test data collected by the South Carolina Statewide Testing Program. The MTP focused on basic technical issues encountered in using…
Distribution of CYP2D6 alleles and phenotypes in the Brazilian population.
Friedrich, Deise C; Genro, Júlia P; Sortica, Vinicius A; Suarez-Kurtz, Guilherme; de Moraes, Maria Elizabete; Pena, Sergio D J; dos Santos, Andrea K Ribeiro; Romano-Silva, Marco A; Hutz, Mara H
2014-01-01
The CYP2D6 enzyme is one of the most important members of the cytochrome P450 superfamily. This enzyme metabolizes approximately 25% of currently prescribed medications. The CYP2D6 gene presents a high allele heterogeneity that determines great inter-individual variation. The aim of this study was to evaluate the variability of CYP2D6 alleles, genotypes and predicted phenotypes in Brazilians. Eleven single nucleotide polymorphisms and CYP2D6 duplications/multiplications were genotyped by TaqMan assays in 1020 individuals from North, Northeast, South, and Southeast Brazil. Eighteen CYP2D6 alleles were identified in the Brazilian population. The CYP2D6*1 and CYP2D6*2 alleles were the most frequent and widely distributed in different geographical regions of Brazil. The highest number of CYPD6 alleles observed was six and the frequency of individuals with more than two copies ranged from 6.3% (in Southern Brazil) to 10.2% (Northern Brazil). The analysis of molecular variance showed that CYP2D6 is homogeneously distributed across different Brazilian regions and most of the differences can be attributed to inter-individual differences. The most frequent predicted metabolic status was EM (83.5%). Overall 2.5% and 3.7% of Brazilians were PMs and UMs respectively. Genomic ancestry proportions differ only in the prevalence of intermediate metabolizers. The IM predicted phenotype is associated with a higher proportion of African ancestry and a lower proportion of European ancestry in Brazilians. PM and UM classes did not vary among regions and/or ancestry proportions therefore unique CYP2D6 testing guidelines for Brazilians are possible and could potentially avoid ineffective or adverse events outcomes due to drug prescriptions. PMID:25329392
Distribution of CYP2D6 Alleles and Phenotypes in the Brazilian Population
Sortica, Vinicius A.; Suarez-Kurtz, Guilherme; de Moraes, Maria Elizabete; Pena, Sergio D. J.; dos Santos, Ândrea K. Ribeiro; Romano-Silva, Marco A.; Hutz, Mara H.
2014-01-01
Abstract The CYP2D6 enzyme is one of the most important members of the cytochrome P450 superfamily. This enzyme metabolizes approximately 25% of currently prescribed medications. The CYP2D6 gene presents a high allele heterogeneity that determines great inter-individual variation. The aim of this study was to evaluate the variability of CYP2D6 alleles, genotypes and predicted phenotypes in Brazilians. Eleven single nucleotide polymorphisms and CYP2D6 duplications/multiplications were genotyped by TaqMan assays in 1020 individuals from North, Northeast, South, and Southeast Brazil. Eighteen CYP2D6 alleles were identified in the Brazilian population. The CYP2D6*1 and CYP2D6*2 alleles were the most frequent and widely distributed in different geographical regions of Brazil. The highest number of CYPD6 alleles observed was six and the frequency of individuals with more than two copies ranged from 6.3% (in Southern Brazil) to 10.2% (Northern Brazil). The analysis of molecular variance showed that CYP2D6 is homogeneously distributed across different Brazilian regions and most of the differences can be attributed to inter-individual differences. The most frequent predicted metabolic status was EM (83.5%). Overall 2.5% and 3.7% of Brazilians were PMs and UMs respectively. Genomic ancestry proportions differ only in the prevalence of intermediate metabolizers. The IM predicted phenotype is associated with a higher proportion of African ancestry and a lower proportion of European ancestry in Brazilians. PM and UM classes did not vary among regions and/or ancestry proportions therefore unique CYP2D6 testing guidelines for Brazilians are possible and could potentially avoid ineffective or adverse events outcomes due to drug prescriptions. PMID:25329392
Differential CYP 2D6 Metabolism Alters Primaquine Pharmacokinetics
Potter, Brittney M. J.; Xie, Lisa H.; Vuong, Chau; Zhang, Jing; Zhang, Ping; Duan, Dehui; Luong, Thu-Lan T.; Bandara Herath, H. M. T.; Dhammika Nanayakkara, N. P.; Tekwani, Babu L.; Walker, Larry A.; Nolan, Christina K.; Sciotti, Richard J.; Zottig, Victor E.; Smith, Philip L.; Paris, Robert M.; Read, Lisa T.; Li, Qigui; Pybus, Brandon S.; Sousa, Jason C.; Reichard, Gregory A.
2015-01-01
Primaquine (PQ) metabolism by the cytochrome P450 (CYP) 2D family of enzymes is required for antimalarial activity in both humans (2D6) and mice (2D). Human CYP 2D6 is highly polymorphic, and decreased CYP 2D6 enzyme activity has been linked to decreased PQ antimalarial activity. Despite the importance of CYP 2D metabolism in PQ efficacy, the exact role that these enzymes play in PQ metabolism and pharmacokinetics has not been extensively studied in vivo. In this study, a series of PQ pharmacokinetic experiments were conducted in mice with differential CYP 2D metabolism characteristics, including wild-type (WT), CYP 2D knockout (KO), and humanized CYP 2D6 (KO/knock-in [KO/KI]) mice. Plasma and liver pharmacokinetic profiles from a single PQ dose (20 mg/kg of body weight) differed significantly among the strains for PQ and carboxy-PQ. Additionally, due to the suspected role of phenolic metabolites in PQ efficacy, these were probed using reference standards. Levels of phenolic metabolites were highest in mice capable of metabolizing CYP 2D6 substrates (WT and KO/KI 2D6 mice). PQ phenolic metabolites were present in different quantities in the two strains, illustrating species-specific differences in PQ metabolism between the human and mouse enzymes. Taking the data together, this report furthers understanding of PQ pharmacokinetics in the context of differential CYP 2D metabolism and has important implications for PQ administration in humans with different levels of CYP 2D6 enzyme activity. PMID:25645856
2D barcodes: a novel and simple method for denture identification.
Sudheendra, Udyavara S; Sowmya, Kasetty; Vidhi, Mathur; Shreenivas, Kallianpur; Prathamesh, Joshi
2013-01-01
Several methods of denture marking have been described in the literature. However, most of them are expensive, time-consuming, and do not permit the incorporation of large amounts of information. We propose a novel and simple method incorporating 2D codes which has several advantages over the existing methods. A 2D code was generated in the dental office and inserted into a maxillary denture. The code was then read using software downloaded into a mobile phone giving access to the website containing details about the patient. The denture was also subjected to durability tests, which did not hamper the efficacy of the 2D code. 2D coding for dentures is a simple, less expensive method with the potential of storing a large amount of information that can be accessed on-site by the forensic investigator, thus allowing quick identification of the denture wearer. PMID:22971078
Lagrangian statistics in laboratory 2D turbulence
NASA Astrophysics Data System (ADS)
Xia, Hua; Francois, Nicolas; Punzmann, Horst; Shats, Michael
2014-05-01
Turbulent mixing in liquids and gases is ubiquitous in nature and industrial flows. Understanding statistical properties of Lagrangian trajectories in turbulence is crucial for a range of problems such as spreading of plankton in the ocean, transport of pollutants, etc. Oceanic data on trajectories of the free-drifting instruments, indicate that the trajectory statistics can often be described by a Lagrangian integral scale. Turbulence however is a state of a flow dominated by a hierarchy of scales, and it is not clear which of these scales mostly affect particle dispersion. Moreover, coherent structures often coexist with turbulence in laboratory experiments [1]. The effect of coherent structures on particle dispersion in turbulent flows is not well understood. Recent progress in scientific imaging and computational power made it possible to tackle this problem experimentally. In this talk, we report the analysis of the higher order Lagrangian statistics in laboratory two-dimensional turbulence. Our results show that fluid particle dispersion is diffusive and it is determined by a single measurable Lagrangian scale related to the forcing scale [2]. Higher order moments of the particle dispersion show strong self-similarity in fully developed turbulence [3]. Here we introduce a new dispersion law that describes single particle dispersion during the turbulence development [4]. These results offer a new way of predicting dispersion in turbulent flows in which one of the low energy scales are persistent. It may help better understanding of drifter Lagrangian statistics in the regions of the ocean where small scale coherent eddies are present [5]. Reference: 1. H. Xia, H. Punzmann, G. Falkovich and M. Shats, Physical Review Letters, 101, 194504 (2008) 2. H. Xia, N. Francois, H. Punzmann, and M. Shats, Nature Communications, 4, 2013 (2013) 3. R. Ferrari, A.J. Manfroi , W.R. Young, Physica D 154 111 (2001) 4. H. Xia, N. Francois, H. Punzmann and M. Shats, submitted (2014
2D to 3D to 2D Dimensionality Crossovers in Thin BSCCO Films
NASA Astrophysics Data System (ADS)
Williams, Gary A.
2003-03-01
With increasing temperature the superfluid fraction in very thin BSCCO films undergoes a series of dimensionality crossovers. At low temperatures the strong anisotropy causes the thermal excitations to be 2D pancake-antipancake pairs in uncoupled layers. At higher temperatures where the c-axis correlation length becomes larger than a layer there is a crossover to 3D vortex loops. These are initially elliptical, but as the 3D Tc is approached they become more circular as the anisotropy scales away, as modeled by Shenoy and Chattopadhyay [1]. Close to Tc when the correlation length becomes comparable to the film thickness there is a further crossover to a 2D Kosterlitz-Thouless transition, with a drop of the superfluid fraction to zero at T_KT which can be of the order of 1 K below T_c. Good agreement with this model is found for experiments on thin BSCCO 2212 films [2]. 1. S. R. Shenoy and B. Chattopadhyay, Phys. Rev. B 51, 9129 (1995). 2. K. Osborn et al., cond-mat/0204417.
Mechanical characterization of 2D, 2D stitched, and 3D braided/RTM materials
NASA Technical Reports Server (NTRS)
Deaton, Jerry W.; Kullerd, Susan M.; Portanova, Marc A.
1993-01-01
Braided composite materials have potential for application in aircraft structures. Fuselage frames, floor beams, wing spars, and stiffeners are examples where braided composites could find application if cost effective processing and damage tolerance requirements are met. Another important consideration for braided composites relates to their mechanical properties and how they compare to the properties of composites produced by other textile composite processes being proposed for these applications. Unfortunately, mechanical property data for braided composites do not appear extensively in the literature. Data are presented in this paper on the mechanical characterization of 2D triaxial braid, 2D triaxial braid plus stitching, and 3D (through-the-thickness) braid composite materials. The braided preforms all had the same graphite tow size and the same nominal braid architectures, (+/- 30 deg/0 deg), and were resin transfer molded (RTM) using the same mold for each of two different resin systems. Static data are presented for notched and unnotched tension, notched and unnotched compression, and compression after impact strengths at room temperature. In addition, some static results, after environmental conditioning, are included. Baseline tension and compression fatigue results are also presented, but only for the 3D braided composite material with one of the resin systems.
GetDDM: An open framework for testing optimized Schwarz methods for time-harmonic wave problems
NASA Astrophysics Data System (ADS)
Thierry, B.; Vion, A.; Tournier, S.; El Bouajaji, M.; Colignon, D.; Marsic, N.; Antoine, X.; Geuzaine, C.
2016-06-01
We present an open finite element framework, called GetDDM, for testing optimized Schwarz domain decomposition techniques for time-harmonic wave problems. After a review of Schwarz domain decomposition methods and associated transmission conditions, we discuss the implementation, based on the open source software GetDP and Gmsh. The solver, along with ready-to-use examples for Helmholtz and Maxwell's equations, is freely available online for further testing.
FLAG Simulations of the Elasticity Test Problem of Gavrilyuk et al.
Kamm, James R.; Runnels, Scott R.; Canfield, Thomas R.; Carney, Theodore C.
2014-04-23
This report contains a description of the impact problem used to compare hypoelastic and hyperelastic material models, as described by Gavrilyuk, Favrie & Saurel. That description is used to set up hypoelastic simulations in the FLAG hydrocode.
Learning from graphically integrated 2D and 3D representations improves retention of neuroanatomy
NASA Astrophysics Data System (ADS)
Naaz, Farah
Visualizations in the form of computer-based learning environments are highly encouraged in science education, especially for teaching spatial material. Some spatial material, such as sectional neuroanatomy, is very challenging to learn. It involves learning the two dimensional (2D) representations that are sampled from the three dimensional (3D) object. In this study, a computer-based learning environment was used to explore the hypothesis that learning sectional neuroanatomy from a graphically integrated 2D and 3D representation will lead to better learning outcomes than learning from a sequential presentation. The integrated representation explicitly demonstrates the 2D-3D transformation and should lead to effective learning. This study was conducted using a computer graphical model of the human brain. There were two learning groups:
Differential Cytochrome P450 2D Metabolism Alters Tafenoquine Pharmacokinetics
Vuong, Chau; Xie, Lisa H.; Potter, Brittney M. J.; Zhang, Jing; Zhang, Ping; Duan, Dehui; Nolan, Christina K.; Sciotti, Richard J.; Zottig, Victor E.; Nanayakkara, N. P. Dhammika; Tekwani, Babu L.; Walker, Larry A.; Smith, Philip L.; Paris, Robert M.; Read, Lisa T.; Li, Qigui; Pybus, Brandon S.; Sousa, Jason C.; Reichard, Gregory A.; Smith, Bryan
2015-01-01
Cytochrome P450 (CYP) 2D metabolism is required for the liver-stage antimalarial efficacy of the 8-aminoquinoline molecule tafenoquine in mice. This could be problematic for Plasmodium vivax radical cure, as the human CYP 2D ortholog (2D6) is highly polymorphic. Diminished CYP 2D6 enzyme activity, as in the poor-metabolizer phenotype, could compromise radical curative efficacy in humans. Despite the importance of CYP 2D metabolism for tafenoquine liver-stage efficacy, the exact role that CYP 2D metabolism plays in the metabolism and pharmacokinetics of tafenoquine and other 8-aminoquinoline molecules has not been extensively studied. In this study, a series of tafenoquine pharmacokinetic experiments were conducted in mice with different CYP 2D metabolism statuses, including wild-type (WT) (reflecting extensive metabolizers for CYP 2D6 substrates) and CYPmouse 2D knockout (KO) (reflecting poor metabolizers for CYP 2D6 substrates) mice. Plasma and liver pharmacokinetic profiles from a single 20-mg/kg of body weight dose of tafenoquine differed between the strains; however, the differences were less striking than previous results obtained for primaquine in the same model. Additionally, the presence of a 5,6-ortho-quinone tafenoquine metabolite was examined in both mouse strains. The 5,6-ortho-quinone species of tafenoquine was observed, and concentrations of the metabolite were highest in the WT extensive-metabolizer phenotype. Altogether, this study indicates that CYP 2D metabolism in mice affects tafenoquine pharmacokinetics and could have implications for human tafenoquine pharmacokinetics in polymorphic CYP 2D6 human populations. PMID:25870069
Spot identification on 2D electrophoresis gel images
NASA Astrophysics Data System (ADS)
Wang, Weixing
2006-09-01
2-D electrophoresis gel images can be used for identifying and characterizing many forms of a particular protein encoded by a single gene. Conventional approaches to gel analysis require the three steps: (1) Spot detection on each gel; (2) Spot matching between gels; and (3) Spot quantification and comparison. Many researchers and developers attempt to automate all steps as much as possible, but errors in the detection and matching stages are common. In order to carry out gel image analysis, one first needs to accurately detect and measure the protein spots in a gel image. This paper presents the algorithms for automatically delineating gel spots. The fusion of two types of segmentation algorithms was implemented. One is edge (discontinuity) based type, and the other is region based type. The primary integration of the two types of image segmentation algorithms have been tested too, the test results clearly show that the integrated algorithm can automatically delineate gel spots not only on a simple image and also on a complex image, and it is much better that either only edge based algorithm or only region based algorithm. Based on the testing and analysis results, the fusion of edge information and region information for gel image segmentation is good for this kind of images.
ERIC Educational Resources Information Center
Liu, Lisa L.; Lau, Anna S.; Chen, Angela Chia-Chen; Dinh, Khanh T.; Kim, Su Yeong
2009-01-01
Associations among neighborhood disadvantage, maternal acculturation, parenting and conduct problems were investigated in a sample of 444 Chinese American adolescents. Adolescents (54% female, 46% male) ranged from 12 to 15 years of age (mean age = 13.0 years). Multilevel modeling was employed to test the hypothesis that the association between…
Test Problems for Reactive Flow HE Model in the ALE3D Code and Limited Sensitivity Study
Gerassimenko, M.
2000-03-01
We document quick running test problems for a reactive flow model of HE initiation incorporated into ALE3D. A quarter percent change in projectile velocity changes the outcome from detonation to HE burn that dies down. We study the sensitivity of calculated HE behavior to several parameters of practical interest where modeling HE initiation with ALE3D.
ERIC Educational Resources Information Center
Masson, J. D.; Dagnan, D.; Evans, J.
2010-01-01
Background: There is a need for validated, standardised tools for the assessment of executive functions in adults with intellectual disabilities (ID). This study examines the validity of a test of planning and problem solving (Tower of London) with adults with ID. Method: Participants completed an adapted version of the Tower of London (ToL) while…
ERIC Educational Resources Information Center
Educational Testing Service, Los Angeles, CA.
At the eighth Western Regional Conference on Testing Problems, the following speeches were given: (1) "Background for Recent Curriculum Developments" by Ralph W. Tyler; (2) "Recent Curriculum Developments in Mathematics" by Robert E. K. Rourke; (3) "Recent Curriculum Developments in Science" by Kenneth E. Davis; (4) "Changes in Measurement as a…
ERIC Educational Resources Information Center
Hambrick, David Z.; Libarkin, Julie C.; Petcovic, Heather L.; Baker, Kathleen M.; Elkins, Joe; Callahan, Caitlin N.; Turner, Sheldon P.; Rench, Tara A.; LaDue, Nicole D.
2012-01-01
Sources of individual differences in scientific problem solving were investigated. Participants representing a wide range of experience in geology completed tests of visuospatial ability and geological knowledge, and performed a geological bedrock mapping task, in which they attempted to infer the geological structure of an area in the Tobacco…
A Dutch Comprehension Test for Identifying Reading Problems in L1 and L2 Students.
ERIC Educational Resources Information Center
Hacquebord, Hilde
1999-01-01
Evaluates the Dutch Text Comprehension Test, developed to help improve the education of secondary school minority students in the Netherlands. Finds the test proved reliable for both L1 students and L2 students. Suggests, although classification results were not optimal, the test can be used for diagnostic purposes provided a relatively large…
Independence Problems for Certain Tests Based on the Shine-Bower Error Term
ERIC Educational Resources Information Center
Shine, Lester C.
1975-01-01
For the Shine-Bower single-subject ANOVA the numerator and demoninator of all F tests based on the Shine-Bower error term are independent of each other. The same property holds for all such tests in the Shine Combined ANOVA except for the test for the trial by subject interaction. (Author/RC)
Haloperidol plasma concentration in Japanese psychiatric subjects with gene duplication of CYP2D6
Ohnuma, Tohru; Shibata, Nobuto; Matsubara, Yoichiro; Arai, Heii
2003-01-01
Aims The cytochrome P-450 2D6 (CYP2D6) gene duplication/multiduplication producing an increase in enzyme activity, and the common Japanese mutation, CYP2D6*10A producing a decrease of enzyme activity were screened in a large number of Japanese psychiatric subjects (n = 111) in order to investigate whether these mutated alleles affected the plasma concentration of haloperidol. Methods Polymerase chain reaction-restriction fragment length polymorphism (PCR-RFLP) method was performed to identify the CYP2D6*10A and CYP2D6*2 genotypes in subjects who had been taking haloperidol. For the screening of duplicated active CYP2D6 gene, allele-specific long PCR was performed. Plasma concentration of haloperidol was measured by the enzyme immunoassay, and expressed as ‘plasma concentration dose ratio’ to normalize individual differences. Results The plasma concentration–dose ratio showed large interindividual differences of approximately 18-fold. PCR-RFLP methods revealed that 29 (26.1%), 10 (9.0%), 39 (35.1%), 0 (0%), seven (6.3%) and 26 (23.4%) cases possessed the CYP2D6 genotypes *1/*1, *1/*2, *1/*10A, *2/*2, *2/*10A and *10 A/*10A, respectively. Six cases (5.4%) had duplicated CYP2D6 genes. There were no significant differences of plasma concentration–dose ratio between the groups classified by CYP2D6*10A and *2 genotypes (Kruskal–Wallis test; P = 0.37), even in those cases whose daily doses were lower than 20 mg (n = 90, P = 0.91). Subjects having duplicated genes (n = 6) did not show significant differences of plasma concentration–dose ratio by comparison with subjects who had no duplicated genes (Mann–Whitney U-test; P = 0.80). Conclusions Gene duplication, and the common Japanese mutation CYP2D6*10A on CYP2D6 gene are not likely to be the main modulatory factors of plasma concentration of haloperidol in Japanese psychiatric subjects. PMID:12919180
A Geometric Boolean Library for 2D Objects
2006-01-05
The 2D Boolean Library is a collection of C++ classes -- which primarily represent 2D geometric data and relationships, and routines -- which contain algorithms for 2D geometric Boolean operations and utility functions. Classes are provided for 2D points, lines, arcs, edgeuses, loops, surfaces and mask sets. Routines are provided that incorporate the Boolean operations Union(OR), XOR, Intersection and Difference. Various analytical geometry routines and routines for importing and exporting the data in various filemore » formats, are also provided in the library.« less
A Geometric Boolean Library for 2D Objects
McBride, Corey L.; Yarberry, Victor; Jorgensen, Craig
2006-01-05
The 2D Boolean Library is a collection of C++ classes -- which primarily represent 2D geometric data and relationships, and routines -- which contain algorithms for 2D geometric Boolean operations and utility functions. Classes are provided for 2D points, lines, arcs, edgeuses, loops, surfaces and mask sets. Routines are provided that incorporate the Boolean operations Union(OR), XOR, Intersection and Difference. Various analytical geometry routines and routines for importing and exporting the data in various file formats, are also provided in the library.
Estimating 2-D vector velocities using multidimensional spectrum analysis.
Oddershede, Niels; Løvstakken, Lasse; Torp, Hans; Jensen, Jørgen Arendt
2008-08-01
Wilson (1991) presented an ultrasonic wideband estimator for axial blood flow velocity estimation through the use of the 2-D Fourier transform. It was shown how a single velocity component was concentrated along a line in the 2-D Fourier space, where the slope was given by the axial velocity. Later, it was shown that this approach could also be used for finding the lateral velocity component by also including a lateral sampling. A single velocity component would then be concentrated along a plane in the 3-D Fourier space, tilted according to the 2 velocity components. This paper presents 2 new velocity estimators for finding both the axial and lateral velocity components. The estimators essentially search for the plane in the 3- D Fourier space, where the integrated power spectrum is largest. The first uses the 3-D Fourier transform to find the power spectrum, while the second uses a minimum variance approach. Based on this plane, the axial and lateral velocity components are estimated. Several phantom measurements, for flow-to-depth angles of 60, 75, and 90 degrees, were performed. Multiple parallel lines were beamformed simultaneously, and 2 different receive apodization schemes were tried. The 2 estimators were then applied to the data. The axial velocity component was estimated with an average standard deviation below 2.8% of the peak velocity, while the average standard deviation of the lateral velocity estimates was between 2.0% and 16.4%. The 2 estimators were also tested on in vivo data from a transverse scan of the common carotid artery, showing the potential of the vector velocity estimation method under in vivo conditions. PMID:18986918
AnisWave2D: User's Guide to the 2d Anisotropic Finite-DifferenceCode
Toomey, Aoife
2005-01-06
This document describes a parallel finite-difference code for modeling wave propagation in 2D, fully anisotropic materials. The code utilizes a mesh refinement scheme to improve computational efficiency. Mesh refinement allows the grid spacing to be tailored to the velocity model, so that fine grid spacing can be used in low velocity zones where the seismic wavelength is short, and coarse grid spacing can be used in zones with higher material velocities. Over-sampling of the seismic wavefield in high velocity zones is therefore avoided. The code has been implemented to run in parallel over multiple processors and allows large-scale models and models with large velocity contrasts to be simulated with ease.
Klassifikation von Standardebenen in der 2D-Echokardiographie mittels 2D-3D-Bildregistrierung
NASA Astrophysics Data System (ADS)
Bergmeir, Christoph; Subramanian, Navneeth
Zum Zweck der Entwicklung eines Systems, das einen unerfahrenen Anwender von Ultraschall (US) zur Aufnahme relevanter anatomischer Strukturen leitet, untersuchen wir die Machbarkeit von 2D-US zu 3D-CT Registrierung. Wir verwenden US-Aufnahmen von Standardebenen des Herzens, welche zu einem 3D-CT-Modell registriert werden. Unser Algorithmus unterzieht sowohl die US-Bilder als auch den CT-Datensatz Vorverarbeitungsschritten, welche die Daten durch Segmentierung auf wesentliche Informationen in Form von Labein für Muskel und Blut reduzieren. Anschließend werden diese Label zur Registrierung mittels der Match-Cardinality-Metrik genutzt. Durch mehrmaliges Registrieren mit verschiedenen Initialisierungen ermitteln wir die im US-Bild sichtbare Standardebene. Wir evaluierten die Methode auf sieben US-Bildern von Standardebenen. Fünf davon wurden korrekt zugeordnet.
Topological evolutionary computing in the optimal design of 2D and 3D structures
NASA Astrophysics Data System (ADS)
Burczynski, T.; Poteralski, A.; Szczepanik, M.
2007-10-01
An application of evolutionary algorithms and the finite-element method to the topology optimization of 2D structures (plane stress, bending plates, and shells) and 3D structures is described. The basis of the topological evolutionary optimization is the direct control of the density material distribution (or thickness for 2D structures) by the evolutionary algorithm. The structures are optimized for stress, mass, and compliance criteria. The numerical examples demonstrate that this method is an effective technique for solving problems in computer-aided optimal design.
No relation between 2D : 4D fetal testosterone marker and dyslexia.
Boets, Bart; De Smedt, Bert; Wouters, Jan; Lemay, Katrien; Ghesquière, Pol
2007-09-17
It has been suggested that high levels of prenatal testosterone exposure are implied in the aetiology of dyslexia and its frequently co-occurring sensory problems. This study examined 2D : 4D digit ratio (a marker of fetal testosterone exposure) in dyslexic and normal reading children. No group differences in 2D : 4D were observed. Digit ratio did not show the postulated relation with reading, spelling, phonological ability, speech perception, auditory processing and visual processing. These findings challenge the validity of theories that allocate a prominent role to fetal testosterone exposure in the aetiology of dyslexia and its sensory impairments. PMID:17712280
Determination of the Ground Vibrational State Parameters of the C2D4 Molecule
NASA Astrophysics Data System (ADS)
Fomchenko, A. L.; Zhang, F.; Gromova, O. V.; Buttersack, T.
2016-07-01
The object of the study is the C2D4 molecule, as it is important to know its properties to address numerous problems of molecular physics. The analysis of high-resolution spectra of the deuterated ethylene molecule was made in the range of 600-1200 cm-1, specifically bands ν7 and ν12. The results obtained were used to determine high-accurate values of the vibrational-rotational levels of the ground vibrational state of the C2D4 molecule.
Global well-posedness of the 2D Boussinesq equations with fractional Laplacian dissipation
NASA Astrophysics Data System (ADS)
Ye, Zhuan; Xu, Xiaojing
2016-04-01
As a continuation of the previous work [48], in this paper we focus on the Cauchy problem of the two-dimensional (2D) incompressible Boussinesq equations with fractional Laplacian dissipation. We give an elementary proof of the global regularity of the smooth solutions of the 2D Boussinesq equations with a new range of fractional powers of the Laplacian. The argument is based on the nonlinear lower bounds for the fractional Laplacian established in [13]. Consequently, this result significantly improves the recent works [13,45,48].
Functional characterization of CYP2D6 enhancer polymorphisms
Wang, Danxin; Papp, Audrey C.; Sun, Xiaochun
2015-01-01
CYP2D6 metabolizes nearly 25% of clinically used drugs. Genetic polymorphisms cause large inter-individual variability in CYP2D6 enzyme activity and are currently used as biomarker to predict CYP2D6 metabolizer phenotype. Previously, we had identified a region 115 kb downstream of CYP2D6 as enhancer for CYP2D6, containing two completely linked single nucleotide polymorphisms (SNPs), rs133333 and rs5758550, associated with enhanced transcription. However, the enhancer effect on CYP2D6 expression, and the causative variant, remained to be ascertained. To characterize the CYP2D6 enhancer element, we applied chromatin conformation capture combined with the next-generation sequencing (4C assays) and chromatin immunoprecipitation with P300 antibody, in HepG2 and human primary culture hepatocytes. The results confirmed the role of the previously identified enhancer region in CYP2D6 expression, expanding the number of candidate variants to three highly linked SNPs (rs133333, rs5758550 and rs4822082). Among these, only rs5758550 demonstrated regulating enhancer activity in a reporter gene assay. Use of clustered regularly interspaced short palindromic repeats mediated genome editing in HepG2 cells targeting suspected enhancer regions decreased CYP2D6 mRNA expression by 70%, only upon deletion of the rs5758550 region. These results demonstrate robust effects of both the enhancer element and SNP rs5758550 on CYP2D6 expression, supporting consideration of rs5758550 for CYP2D6 genotyping panels to yield more accurate phenotype prediction. PMID:25381333
Code of Federal Regulations, 2011 CFR
2011-10-01
... 49 Transportation 1 2011-10-01 2011-10-01 false What is the effect of procedural problems that are... of Transportation PROCEDURES FOR TRANSPORTATION WORKPLACE DRUG AND ALCOHOL TESTING PROGRAMS Problems in Alcohol Testing § 40.275 What is the effect of procedural problems that are not sufficient...
Melton-Celsa, Angela R; O'Brien, Alison D; Feng, Peter C H
2015-11-01
Shiga toxin (Stx)-producing Escherichia coli (STEC) strains are food- and waterborne pathogens that are often transmitted via beef products or fresh produce. STEC strains cause both sporadic infections and outbreaks, which may result in hemorrhagic colitis and hemolytic uremic syndrome. STEC strains may elaborate Stx1, Stx2, and/or subtypes of those toxins. Epidemiological evidence indicates that STEC that produce subtypes Stx2a, Stx2c, and/or Stx2d are more often associated with serious illness. The Stx2d subtype becomes more toxic to Vero cells after incubation with intestinal mucus or elastase, a process named "activation." Stx2d is not generally found in the E. coli serotypes most commonly connected to STEC outbreaks. However, STEC strains that are stx2d positive can be isolated from foods, an occurrence that gives rise to the question of whether those food isolates are potential human pathogens. In this study, we examined 14 STEC strains from fresh produce that were stx2d positive and found that they all produced the mucus-activatable Stx2d and that a subset of the strains tested were virulent in streptomycin-treated mice. PMID:26555533
The use of 2D and 3D information in a perceptual-cognitive judgement task.
Put, Koen; Wagemans, Johan; Spitz, Jochim; Gallardo, Manuel Armenteros; Williams, A Mark; Helsen, Werner F
2014-01-01
We examined whether the use of three-dimensional (3D) simulations in an off-field offside decision-making task is beneficial compared to the more widely available two-dimensional (2D) simulations. Thirty-three assistant referees, who were all involved in professional football, participated in the experiment. They assessed 40 offside situations in both 2D and 3D formats using a counterbalanced design. A distinction was made between offside situations near (i.e., 15 m) and far (i.e., 30 m) from the touchline. Subsequently, a frame recognition task was performed in which assistant referees were asked to indicate which of the five pictures represented the previous video scene. A higher response accuracy score was observed under 3D (80.0%) compared to 2D (75.0%) conditions, in particular for the situations near the touchline (3D: 81.8%; 2D: 72.7%). No differences were reported between 2D and 3D in the frame recognition task. Findings suggest that in highly dynamic and complex situations, the visual system can benefit from the availability of 3D information, especially for relatively fine, metric position judgements. In the memory task, in which a mental abstraction had to be made from a dynamic situation to a static snapshot, 3D stereo disparities do not add anything over and beyond 2D simulations. The specific task demands should be taken into account when considering the most appropriate format for testing and training. PMID:24857384
NASA Astrophysics Data System (ADS)
Jeromin, A.; Schaffarczyk, A. P.; Puczylowski, J.; Peinke, J.; Hölling, M.
2014-12-01
For the investigation of atmospheric turbulent flows on small scales a new anemometer was developed, the so-called 2d-Atmospheric Laser Cantilever Anemometer (2d-ALCA). It performs highly resolved measurements with a spatial resolution in millimeter range and temporal resolution in kHz range, thus detecting very small turbulent structures. The anemometer is a redesign of the successfully operating 2d-LCA for laboratory application. The new device was designed to withstand hostile operating environments (rain and saline, humid air). In February 2012, the 2d-ALCA was used for the first time in a test field. The device was mounted in about 53 m above ground level on a lattice tower near the German North Sea coast. Wind speed was measured by the 2d-ALCA at 10 kHz sampling rate and by cup anemometers at 1 Hz. The instantaneous wind speed ranged from 8 m/s to 19 m/s at an average turbulence level of about 7 %. Wind field characteristics were analyzed based on cup anemometer as well as 2d-ALCA. The combination of both devices allowed the study of atmospheric turbulence over several magnitudes in turbulent scales.
Effect of CYP2D6 genetic polymorphism on the metabolism of citalopram in vitro.
Hu, Xiao-Xia; Yuan, Ling-Jing; Fang, Ping; Mao, Yong-Hui; Zhan, Yun-Yun; Li, Xiang-Yu; Dai, Da-Peng; Cai, Jian-Ping; Hu, Guo-Xin
2016-04-01
Genetic polymorphisms of CYP2D6 significantly influence the efficacy and safety of some drugs, which might cause adverse effects and therapeutic failure. We aimed at investigating the role of CYP2D6 in the metabolism of citalopram and identifying the effect of 24 CYP2D6 allelic variants we found in Chinese Han population on the metabolism of citalopram in vitro. These CYP2D6 variants expressed by insect cells system were incubated with 10-1000 μM citalopram for 30 min at 37 °C and the reaction was terminated by cooling to -80 °C immediately. Citalopram and its metabolites were analyzed by high-performance liquid chromatography (HPLC). The intrinsic clearance (Vmax/Km) values of the variants toward citalopram metabolites were significantly altered, 38-129% for demethylcitalopram and 13-138% for citalopram N-oxide when compared with CYP2D6*1. Most of the tested rare alleles exhibited significantly decreased values due to increased Km and/or decreased Vmax values. We conclude that recombinant system could be used to investigate the enzymes involved in drug metabolism and these findings suggest that more attention should be paid to subjects carrying these CYP2D6 alleles when administering citalopram in the clinic. PMID:27016952
NASA Astrophysics Data System (ADS)
Laibe, Guillaume; Price, Daniel J.
2011-12-01
In this paper we present the analytic solutions for two test problems involving two-fluid mixtures of dust and gas in an astrophysical context. The solutions provide a means of benchmarking numerical codes designed to simulate the non-linear dynamics of dusty gas. The first problem, DUSTYBOX, consists of two interpenetrating homogeneous fluids moving with relative velocity difference. We provide exact solutions to the full non-linear problem for a range of drag formulations appropriate to astrophysical fluids (i.e. various prescriptions for Epstein and Stokes drag in different regimes). The second problem, DUSTYWAVE, consists of the propagation of linear acoustic waves in a two-fluid gas-dust mixture. We provide the analytic solution for the case when the two fluids are interacting via a linear drag term. Both test problems are simple to set up in any numerical code and can be run with periodic boundary conditions. The solutions we derive are completely general with respect to both the dust-to-gas ratio and the amplitude of the drag coefficient. A stability analysis of waves in a gas-dust system is also presented, showing that sound waves in an astrophysical dust-gas mixture are linearly stable.
Hyun, Eugin; Jin, Young-Seok; Lee, Jong-Hun
2016-01-01
For an automotive pedestrian detection radar system, fast-ramp based 2D range-Doppler Frequency Modulated Continuous Wave (FMCW) radar is effective for distinguishing between moving targets and unwanted clutter. However, when a weak moving target such as a pedestrian exists together with strong clutter, the pedestrian may be masked by the side-lobe of the clutter even though they are notably separated in the Doppler dimension. To prevent this problem, one popular solution is the use of a windowing scheme with a weighting function. However, this method leads to a spread spectrum, so the pedestrian with weak signal power and slow Doppler may also be masked by the main-lobe of clutter. With a fast-ramp based FMCW radar, if the target is moving, the complex spectrum of the range- Fast Fourier Transform (FFT) is changed with a constant phase difference over ramps. In contrast, the clutter exhibits constant phase irrespective of the ramps. Based on this fact, in this paper we propose a pedestrian detection for highly cluttered environments using a coherent phase difference method. By detecting the coherent phase difference from the complex spectrum of the range-FFT, we first extract the range profile of the moving pedestrians. Then, through the Doppler FFT, we obtain the 2D range-Doppler map for only the pedestrian. To test the proposed detection scheme, we have developed a real-time data logging system with a 24 GHz FMCW transceiver. In laboratory tests, we verified that the signal processing results from the proposed method were much better than those expected from the conventional 2D FFT-based detection method. PMID:26805835
Hyun, Eugin; Jin, Young-Seok; Lee, Jong-Hun
2016-01-01
For an automotive pedestrian detection radar system, fast-ramp based 2D range-Doppler Frequency Modulated Continuous Wave (FMCW) radar is effective for distinguishing between moving targets and unwanted clutter. However, when a weak moving target such as a pedestrian exists together with strong clutter, the pedestrian may be masked by the side-lobe of the clutter even though they are notably separated in the Doppler dimension. To prevent this problem, one popular solution is the use of a windowing scheme with a weighting function. However, this method leads to a spread spectrum, so the pedestrian with weak signal power and slow Doppler may also be masked by the main-lobe of clutter. With a fast-ramp based FMCW radar, if the target is moving, the complex spectrum of the range- Fast Fourier Transform (FFT) is changed with a constant phase difference over ramps. In contrast, the clutter exhibits constant phase irrespective of the ramps. Based on this fact, in this paper we propose a pedestrian detection for highly cluttered environments using a coherent phase difference method. By detecting the coherent phase difference from the complex spectrum of the range-FFT, we first extract the range profile of the moving pedestrians. Then, through the Doppler FFT, we obtain the 2D range-Doppler map for only the pedestrian. To test the proposed detection scheme, we have developed a real-time data logging system with a 24 GHz FMCW transceiver. In laboratory tests, we verified that the signal processing results from the proposed method were much better than those expected from the conventional 2D FFT-based detection method. PMID:26805835
NASA Astrophysics Data System (ADS)
Chae, Dongho; Constantin, Peter; Wu, Jiahong
2014-09-01
We give an example of a well posed, finite energy, 2D incompressible active scalar equation with the same scaling as the surface quasi-geostrophic equation and prove that it can produce finite time singularities. In spite of its simplicity, this seems to be the first such example. Further, we construct explicit solutions of the 2D Boussinesq equations whose gradients grow exponentially in time for all time. In addition, we introduce a variant of the 2D Boussinesq equations which is perhaps a more faithful companion of the 3D axisymmetric Euler equations than the usual 2D Boussinesq equations.