The Cauchy problem for space-time monopole equations in Sobolev spaces
NASA Astrophysics Data System (ADS)
Huh, Hyungjin; Yim, Jihyun
2018-04-01
We consider the initial value problem of space-time monopole equations in one space dimension with initial data in Sobolev space Hs. Observing null structures of the system, we prove local well-posedness in almost critical space. Unconditional uniqueness and global existence are proved for s ≥ 0. Moreover, we show that the H1 Sobolev norm grows at a rate of at most c exp(ct2).
Robust fuel- and time-optimal control of uncertain flexible space structures
NASA Technical Reports Server (NTRS)
Wie, Bong; Sinha, Ravi; Sunkel, John; Cox, Ken
1993-01-01
The problem of computing open-loop, fuel- and time-optimal control inputs for flexible space structures in the face of modeling uncertainty is investigated. Robustified, fuel- and time-optimal pulse sequences are obtained by solving a constrained optimization problem subject to robustness constraints. It is shown that 'bang-off-bang' pulse sequences with a finite number of switchings provide a practical tradeoff among the maneuvering time, fuel consumption, and performance robustness of uncertain flexible space structures.
A note on the electromagnetic irradiation in a holed spatial region: A space-time approach
NASA Astrophysics Data System (ADS)
Botelho, Luiz C. L.
2017-02-01
We study the role of the homological topological property of a space-time with holes (a multiple connected manifold) on the formal solution of the electromagnetic irradiation problem taking place on these “holed” space-times. In this paper, in addition to the main focus of study, we present as well important studies on this irradiation problem on other mathematical frameworks.
NASA Astrophysics Data System (ADS)
Hoover, Wm. G.; Hoover, Carol G.
2012-02-01
We compare the Gram-Schmidt and covariant phase-space-basis-vector descriptions for three time-reversible harmonic oscillator problems, in two, three, and four phase-space dimensions respectively. The two-dimensional problem can be solved analytically. The three-dimensional and four-dimensional problems studied here are simultaneously chaotic, time-reversible, and dissipative. Our treatment is intended to be pedagogical, for use in an updated version of our book on Time Reversibility, Computer Simulation, and Chaos. Comments are very welcome.
Collaborative Problem Solving in Shared Space
ERIC Educational Resources Information Center
Lin, Lin; Mills, Leila A.; Ifenthaler, Dirk
2015-01-01
The purpose of this study was to examine collaborative problem solving in a shared virtual space. The main question asked was: How will the performance and processes differ between collaborative problem solvers and independent problem solvers over time? A total of 104 university students (63 female and 41 male) participated in an experimental…
Space-Time Error Representation and Estimation in Navier-Stokes Calculations
NASA Technical Reports Server (NTRS)
Barth, Timothy J.
2006-01-01
The mathematical framework for a-posteriori error estimation of functionals elucidated by Eriksson et al. [7] and Becker and Rannacher [3] is revisited in a space-time context. Using these theories, a hierarchy of exact and approximate error representation formulas are presented for use in error estimation and mesh adaptivity. Numerical space-time results for simple model problems as well as compressible Navier-Stokes flow at Re = 300 over a 2D circular cylinder are then presented to demonstrate elements of the error representation theory for time-dependent problems.
Time-Accurate, Unstructured-Mesh Navier-Stokes Computations with the Space-Time CESE Method
NASA Technical Reports Server (NTRS)
Chang, Chau-Lyan
2006-01-01
Application of the newly emerged space-time conservation element solution element (CESE) method to compressible Navier-Stokes equations is studied. In contrast to Euler equations solvers, several issues such as boundary conditions, numerical dissipation, and grid stiffness warrant systematic investigations and validations. Non-reflecting boundary conditions applied at the truncated boundary are also investigated from the stand point of acoustic wave propagation. Validations of the numerical solutions are performed by comparing with exact solutions for steady-state as well as time-accurate viscous flow problems. The test cases cover a broad speed regime for problems ranging from acoustic wave propagation to 3D hypersonic configurations. Model problems pertinent to hypersonic configurations demonstrate the effectiveness of the CESE method in treating flows with shocks, unsteady waves, and separations. Good agreement with exact solutions suggests that the space-time CESE method provides a viable alternative for time-accurate Navier-Stokes calculations of a broad range of problems.
Optimal perturbations for nonlinear systems using graph-based optimal transport
NASA Astrophysics Data System (ADS)
Grover, Piyush; Elamvazhuthi, Karthik
2018-06-01
We formulate and solve a class of finite-time transport and mixing problems in the set-oriented framework. The aim is to obtain optimal discrete-time perturbations in nonlinear dynamical systems to transport a specified initial measure on the phase space to a final measure in finite time. The measure is propagated under system dynamics in between the perturbations via the associated transfer operator. Each perturbation is described by a deterministic map in the measure space that implements a version of Monge-Kantorovich optimal transport with quadratic cost. Hence, the optimal solution minimizes a sum of quadratic costs on phase space transport due to the perturbations applied at specified times. The action of the transport map is approximated by a continuous pseudo-time flow on a graph, resulting in a tractable convex optimization problem. This problem is solved via state-of-the-art solvers to global optimality. We apply this algorithm to a problem of transport between measures supported on two disjoint almost-invariant sets in a chaotic fluid system, and to a finite-time optimal mixing problem by choosing the final measure to be uniform. In both cases, the optimal perturbations are found to exploit the phase space structures, such as lobe dynamics, leading to efficient global transport. As the time-horizon of the problem is increased, the optimal perturbations become increasingly localized. Hence, by combining the transfer operator approach with ideas from the theory of optimal mass transportation, we obtain a discrete-time graph-based algorithm for optimal transport and mixing in nonlinear systems.
Parallel satellite orbital situational problems solver for space missions design and control
NASA Astrophysics Data System (ADS)
Atanassov, Atanas Marinov
2016-11-01
Solving different scientific problems for space applications demands implementation of observations, measurements or realization of active experiments during time intervals in which specific geometric and physical conditions are fulfilled. The solving of situational problems for determination of these time intervals when the satellite instruments work optimally is a very important part of all activities on every stage of preparation and realization of space missions. The elaboration of universal, flexible and robust approach for situation analysis, which is easily portable toward new satellite missions, is significant for reduction of missions' preparation times and costs. Every situation problem could be based on one or more situation conditions. Simultaneously solving different kinds of situation problems based on different number and types of situational conditions, each one of them satisfied on different segments of satellite orbit requires irregular calculations. Three formal approaches are presented. First one is related to situation problems description that allows achieving flexibility in situation problem assembling and presentation in computer memory. The second formal approach is connected with developing of situation problem solver organized as processor that executes specific code for every particular situational condition. The third formal approach is related to solver parallelization utilizing threads and dynamic scheduling based on "pool of threads" abstraction and ensures a good load balance. The developed situation problems solver is intended for incorporation in the frames of multi-physics multi-satellite space mission's design and simulation tools.
NASA Astrophysics Data System (ADS)
Stepanov, F. I.
2018-04-01
The mechanical properties of a material which is modeled by an exponential creep kernel characterized by a spectrum of relaxation and retardation times are studied. The research is carried out considering a contact problem for a solid indenter sliding over a viscoelastic half-space. The contact pressure, indentation depth of the indenter, and the deformation component of the friction coefficient are analyzed with respect to the case of half-space material modeled by single relaxation and retardation times.
The Spectrum of Mathematical Models.
ERIC Educational Resources Information Center
Karplus, Walter J.
1983-01-01
Mathematical modeling problems encountered in many disciplines are discussed in terms of the modeling process and applications of models. The models are classified according to three types of abstraction: continuous-space-continuous-time, discrete-space-continuous-time, and discrete-space-discrete-time. Limitations in different kinds of modeling…
NASA Astrophysics Data System (ADS)
Torabi, H.; Pariz, N.; Karimpour, A.
2016-02-01
This paper investigates fractional Kalman filters when time-delay is entered in the observation signal in the discrete-time stochastic fractional order state-space representation. After investigating the common fractional Kalman filter, we try to derive a fractional Kalman filter for time-delay fractional systems. A detailed derivation is given. Fractional Kalman filters will be used to estimate recursively the states of fractional order state-space systems based on minimizing the cost function when there is a constant time delay (d) in the observation signal. The problem will be solved by converting the filtering problem to a usual d-step prediction problem for delay-free fractional systems.
CSM solutions of rotating blade dynamics using integrating matrices
NASA Technical Reports Server (NTRS)
Lakin, William D.
1992-01-01
The dynamic behavior of flexible rotating beams continues to receive considerable research attention as it constitutes a fundamental problem in applied mechanics. Further, beams comprise parts of many rotating structures of engineering significance. A topic of particular interest at the present time involves the development of techniques for obtaining the behavior in both space and time of a rotor acted upon by a simple airload loading. Most current work on problems of this type use solution techniques based on normal modes. It is certainly true that normal modes cannot be disregarded, as knowledge of natural blade frequencies is always important. However, the present work has considered a computational structural mechanics (CSM) approach to rotor blade dynamics problems in which the physical properties of the rotor blade provide input for a direct numerical solution of the relevant boundary-and-initial-value problem. Analysis of the dynamics of a given rotor system may require solution of the governing equations over a long time interval corresponding to many revolutions of the loaded flexible blade. For this reason, most of the common techniques in computational mechanics, which treat the space-time behavior concurrently, cannot be applied to the rotor dynamics problem without a large expenditure of computational resources. By contrast, the integrating matrix technique of computational mechanics has the ability to consistently incorporate boundary conditions and 'remove' dependence on a space variable. For problems involving both space and time, this feature of the integrating matrix approach thus can generate a 'splitting' which forms the basis of an efficient CSM method for numerical solution of rotor dynamics problems.
NASA Astrophysics Data System (ADS)
Jia, Shouqing; La, Dongsheng; Ma, Xuelian
2018-04-01
The finite difference time domain (FDTD) algorithm and Green function algorithm are implemented into the numerical simulation of electromagnetic waves in Schwarzschild space-time. FDTD method in curved space-time is developed by filling the flat space-time with an equivalent medium. Green function in curved space-time is obtained by solving transport equations. Simulation results validate both the FDTD code and Green function code. The methods developed in this paper offer a tool to solve electromagnetic scattering problems.
NASA Technical Reports Server (NTRS)
Shakib, Farzin; Hughes, Thomas J. R.
1991-01-01
A Fourier stability and accuracy analysis of the space-time Galerkin/least-squares method as applied to a time-dependent advective-diffusive model problem is presented. Two time discretizations are studied: a constant-in-time approximation and a linear-in-time approximation. Corresponding space-time predictor multi-corrector algorithms are also derived and studied. The behavior of the space-time algorithms is compared to algorithms based on semidiscrete formulations.
Time as an Observable in Nonrelativistic Quantum Mechanics
NASA Technical Reports Server (NTRS)
Hahne, G. E.
2003-01-01
The argument follows from the viewpoint that quantum mechanics is taken not in the usual form involving vectors and linear operators in Hilbert spaces, but as a boundary value problem for a special class of partial differential equations-in the present work, the nonrelativistic Schrodinger equation for motion of a structureless particle in four- dimensional space-time in the presence of a potential energy distribution that can be time-as well as space-dependent. The domain of interest is taken to be one of two semi-infinite boxes, one bounded by two t=constant planes and the other by two t=constant planes. Each gives rise to a characteristic boundary value problem: one in which the initial, input values on one t=constant wall are given, with zero asymptotic wavefunction values in all spatial directions, the output being the values on the second t=constant wall; the second with certain input values given on both z=constant walls, with zero asymptotic values in all directions involving time and the other spatial coordinates, the output being the complementary values on the z=constant walls. The first problem corresponds to ordinary quantum mechanics; the second, to a fully time-dependent version of a problem normally considered only for the steady state (time-independent Schrodinger equation). The second problem is formulated in detail. A conserved indefinite metric is associated with space-like propagation, where the sign of the norm of a unidirectional state corresponds to its spatial direction of travel.
International interest in space assets under the Cape Town Convention
NASA Astrophysics Data System (ADS)
Ametova, Lutfiie
2013-12-01
Private human access to outer space is impossible without space equipment. Nowadays space equipment is increasingly being financed by private sector. Private sector financiers, naturally, seek to secure their interest in space equipment. At the same time, increasing international cooperation in space industry leads to some problems of legal character. Thus, space equipment involved in international cooperation programs crosses national borders and is subject to a certain jurisdiction in a given period of time. The problem is that when an interest is created in one jurisdiction, it may not necessarily be recognised in another one. In order to provide a unified approach to interests vested in space equipment an international legal instrument is necessary. The Cape Town Convention represents an international instrument designed to provide a unified approach to interests vested in mobile equipment, including space assets.
Techniques for shuttle trajectory optimization
NASA Technical Reports Server (NTRS)
Edge, E. R.; Shieh, C. J.; Powers, W. F.
1973-01-01
The application of recently developed function-space Davidon-type techniques to the shuttle ascent trajectory optimization problem is discussed along with an investigation of the recently developed PRAXIS algorithm for parameter optimization. At the outset of this analysis, the major deficiency of the function-space algorithms was their potential storage problems. Since most previous analyses of the methods were with relatively low-dimension problems, no storage problems were encountered. However, in shuttle trajectory optimization, storage is a problem, and this problem was handled efficiently. Topics discussed include: the shuttle ascent model and the development of the particular optimization equations; the function-space algorithms; the operation of the algorithm and typical simulations; variable final-time problem considerations; and a modification of Powell's algorithm.
NASA Astrophysics Data System (ADS)
Raine, D. J.; Heller, M.
Analyzing the development of the structure of space-time from the theory of Aristotle to the present day, the present work attempts to sketch a science of relativistic mechanics. The concept of relativity is discussed in relation to the way in which space-time splits up into space and time, and in relation to Mach's principle concerning the relativity of inertia. Particular attention is given to the following topics: Aristotelian dynamics; Copernican kinematics; Newtonian dynamics; the space-time of classical dynamics; classical space-time in the presence of gravity; the space-time of special relativity; the space-time of general relativity; solutions and problems in general relativity; Mach's principle and the dynamics of space-time; theories of inertial mass; the integral formation of general relativity; and the frontiers of relativity (e.g., unified field theories and quantum gravity).
Inverse Problems for Semilinear Wave Equations on Lorentzian Manifolds
NASA Astrophysics Data System (ADS)
Lassas, Matti; Uhlmann, Gunther; Wang, Yiran
2018-06-01
We consider inverse problems in space-time ( M, g), a 4-dimensional Lorentzian manifold. For semilinear wave equations {\\square_g u + H(x, u) = f}, where {\\square_g} denotes the usual Laplace-Beltrami operator, we prove that the source-to-solution map {L: f → u|_V}, where V is a neighborhood of a time-like geodesic {μ}, determines the topological, differentiable structure and the conformal class of the metric of the space-time in the maximal set, where waves can propagate from {μ} and return back. Moreover, on a given space-time ( M, g), the source-to-solution map determines some coefficients of the Taylor expansion of H in u.
Joint space-time geostatistical model for air quality surveillance
NASA Astrophysics Data System (ADS)
Russo, A.; Soares, A.; Pereira, M. J.
2009-04-01
Air pollution and peoples' generalized concern about air quality are, nowadays, considered to be a global problem. Although the introduction of rigid air pollution regulations has reduced pollution from industry and power stations, the growing number of cars on the road poses a new pollution problem. Considering the characteristics of the atmospheric circulation and also the residence times of certain pollutants in the atmosphere, a generalized and growing interest on air quality issues led to research intensification and publication of several articles with quite different levels of scientific depth. As most natural phenomena, air quality can be seen as a space-time process, where space-time relationships have usually quite different characteristics and levels of uncertainty. As a result, the simultaneous integration of space and time is not an easy task to perform. This problem is overcome by a variety of methodologies. The use of stochastic models and neural networks to characterize space-time dispersion of air quality is becoming a common practice. The main objective of this work is to produce an air quality model which allows forecasting critical concentration episodes of a certain pollutant by means of a hybrid approach, based on the combined use of neural network models and stochastic simulations. A stochastic simulation of the spatial component with a space-time trend model is proposed to characterize critical situations, taking into account data from the past and a space-time trend from the recent past. To identify near future critical episodes, predicted values from neural networks are used at each monitoring station. In this paper, we describe the design of a hybrid forecasting tool for ambient NO2 concentrations in Lisbon, Portugal.
A Game-Theoretic Approach to Branching Time Abstract-Check-Refine Process
NASA Technical Reports Server (NTRS)
Wang, Yi; Tamai, Tetsuo
2009-01-01
Since the complexity of software systems continues to grow, most engineers face two serious problems: the state space explosion problem and the problem of how to debug systems. In this paper, we propose a game-theoretic approach to full branching time model checking on three-valued semantics. The three-valued models and logics provide successful abstraction that overcomes the state space explosion problem. The game style model checking that generates counter-examples can guide refinement or identify validated formulas, which solves the system debugging problem. Furthermore, output of our game style method will give significant information to engineers in detecting where errors have occurred and what the causes of the errors are.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Raine, D.J.; Heller, M.
1981-01-01
Analyzing the development of the structure of space-time from the theory of Aristotle to the present day, the present work attempts to sketch a science of relativistic mechanics. The concept of relativity is discussed in relation to the way in which space-time splits up into space and time, and in relation to Mach's principle concerning the relativity of inertia. Particular attention is given to the following topics: Aristotelian dynamics Copernican kinematics Newtonian dynamics the space-time of classical dynamics classical space-time in the presence of gravity the space-time of special relativity the space-time of general relativity solutions and problems in generalmore » relativity Mach's principle and the dynamics of space-time theories of inertial mass the integral formation of general relativity and the frontiers of relativity (e.g., unified field theories and quantum gravity).« less
NASA Astrophysics Data System (ADS)
Weidemaier, P.
2005-06-01
The trace problem on the hypersurface y_n=0 is investigated for a function u=u(y,t) \\in L_q(0,T;W_{\\underline p}^{\\underline m}(\\mathbb R_+^n)) with \\partial_t u \\in L_q(0,T; L_{\\underline p}(\\mathbb R_+^n)), that is, Sobolev spaces with mixed Lebesgue norm L_{\\underline p,q}(\\mathbb R^n_+\\times(0,T))=L_q(0,T;L_{\\underline p}(\\mathbb R_+^n)) are considered; here \\underline p=(p_1,\\dots,p_n) is a vector and \\mathbb R^n_+=\\mathbb R^{n-1} \\times (0,\\infty). Such function spaces are useful in the context of parabolic equations. They allow, in particular, different exponents of summability in space and time. It is shown that the sharp regularity of the trace in the time variable is characterized by the Lizorkin-Triebel space F_{q,p_n}^{1-1/(p_nm_n)}(0,T;L_{\\widetilde{\\underline p}}(\\mathbb R^{n-1})), \\underline p=(\\widetilde{\\underline p},p_n). A similar result is established for first order spatial derivatives of u. These results allow one to determine the exact spaces for the data in the inhomogeneous Dirichlet and Neumann problems for parabolic equations of the second order if the solution is in the space L_q(0,T; W_p^2(\\Omega)) \\cap W_q^1(0,T;L_p(\\Omega)) with p \\le q.
Divergence identities in curved space-time a resolution of the stress-energy problem
NASA Astrophysics Data System (ADS)
Yilmaz, Hüseyin
1989-03-01
It is noted that the joint use of two basic differential identities in curved space-time, namely, 1) the Einstein-Hilbert identity (1915), and 2) the identity of P. Freud (1939), permits a viable alternative to general relativity and a resolution of the "field stress-energy" problem of the gravitational theory. (A tribute to Eugene P. Wigner's 1957 presidential address to the APS)
Application of Multi-Hypothesis Sequential Monte Carlo for Breakup Analysis
NASA Astrophysics Data System (ADS)
Faber, W. R.; Zaidi, W.; Hussein, I. I.; Roscoe, C. W. T.; Wilkins, M. P.; Schumacher, P. W., Jr.
As more objects are launched into space, the potential for breakup events and space object collisions is ever increasing. These events create large clouds of debris that are extremely hazardous to space operations. Providing timely, accurate, and statistically meaningful Space Situational Awareness (SSA) data is crucial in order to protect assets and operations in space. The space object tracking problem, in general, is nonlinear in both state dynamics and observations, making it ill-suited to linear filtering techniques such as the Kalman filter. Additionally, given the multi-object, multi-scenario nature of the problem, space situational awareness requires multi-hypothesis tracking and management that is combinatorially challenging in nature. In practice, it is often seen that assumptions of underlying linearity and/or Gaussianity are used to provide tractable solutions to the multiple space object tracking problem. However, these assumptions are, at times, detrimental to tracking data and provide statistically inconsistent solutions. This paper details a tractable solution to the multiple space object tracking problem applicable to space object breakup events. Within this solution, simplifying assumptions of the underlying probability density function are relaxed and heuristic methods for hypothesis management are avoided. This is done by implementing Sequential Monte Carlo (SMC) methods for both nonlinear filtering as well as hypothesis management. This goal of this paper is to detail the solution and use it as a platform to discuss computational limitations that hinder proper analysis of large breakup events.
Fortran programs for the time-dependent Gross-Pitaevskii equation in a fully anisotropic trap
NASA Astrophysics Data System (ADS)
Muruganandam, P.; Adhikari, S. K.
2009-10-01
Here we develop simple numerical algorithms for both stationary and non-stationary solutions of the time-dependent Gross-Pitaevskii (GP) equation describing the properties of Bose-Einstein condensates at ultra low temperatures. In particular, we consider algorithms involving real- and imaginary-time propagation based on a split-step Crank-Nicolson method. In a one-space-variable form of the GP equation we consider the one-dimensional, two-dimensional circularly-symmetric, and the three-dimensional spherically-symmetric harmonic-oscillator traps. In the two-space-variable form we consider the GP equation in two-dimensional anisotropic and three-dimensional axially-symmetric traps. The fully-anisotropic three-dimensional GP equation is also considered. Numerical results for the chemical potential and root-mean-square size of stationary states are reported using imaginary-time propagation programs for all the cases and compared with previously obtained results. Also presented are numerical results of non-stationary oscillation for different trap symmetries using real-time propagation programs. A set of convenient working codes developed in Fortran 77 are also provided for all these cases (twelve programs in all). In the case of two or three space variables, Fortran 90/95 versions provide some simplification over the Fortran 77 programs, and these programs are also included (six programs in all). Program summaryProgram title: (i) imagetime1d, (ii) imagetime2d, (iii) imagetime3d, (iv) imagetimecir, (v) imagetimesph, (vi) imagetimeaxial, (vii) realtime1d, (viii) realtime2d, (ix) realtime3d, (x) realtimecir, (xi) realtimesph, (xii) realtimeaxial Catalogue identifier: AEDU_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEDU_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 122 907 No. of bytes in distributed program, including test data, etc.: 609 662 Distribution format: tar.gz Programming language: FORTRAN 77 and Fortran 90/95 Computer: PC Operating system: Linux, Unix RAM: 1 GByte (i, iv, v), 2 GByte (ii, vi, vii, x, xi), 4 GByte (iii, viii, xii), 8 GByte (ix) Classification: 2.9, 4.3, 4.12 Nature of problem: These programs are designed to solve the time-dependent Gross-Pitaevskii nonlinear partial differential equation in one-, two- or three-space dimensions with a harmonic, circularly-symmetric, spherically-symmetric, axially-symmetric or anisotropic trap. The Gross-Pitaevskii equation describes the properties of a dilute trapped Bose-Einstein condensate. Solution method: The time-dependent Gross-Pitaevskii equation is solved by the split-step Crank-Nicolson method by discretizing in space and time. The discretized equation is then solved by propagation, in either imaginary or real time, over small time steps. The method yields the solution of stationary and/or non-stationary problems. Additional comments: This package consists of 12 programs, see "Program title", above. FORTRAN77 versions are provided for each of the 12 and, in addition, Fortran 90/95 versions are included for ii, iii, vi, viii, ix, xii. For the particular purpose of each program please see the below. Running time: Minutes on a medium PC (i, iv, v, vii, x, xi), a few hours on a medium PC (ii, vi, viii, xii), days on a medium PC (iii, ix). Program summary (1)Title of program: imagtime1d.F Title of electronic file: imagtime1d.tar.gz Catalogue identifier: Program summary URL: Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Distribution format: tar.gz Computers: PC/Linux, workstation/UNIX Maximum RAM memory: 1 GByte Programming language used: Fortran 77 Typical running time: Minutes on a medium PC Unusual features: None Nature of physical problem: This program is designed to solve the time-dependent Gross-Pitaevskii nonlinear partial differential equation in one-space dimension with a harmonic trap. The Gross-Pitaevskii equation describes the properties of a dilute trapped Bose-Einstein condensate. Method of solution: The time-dependent Gross-Pitaevskii equation is solved by the split-step Crank-Nicolson method by discretizing in space and time. The discretized equation is then solved by propagation in imaginary time over small time steps. The method yields the solution of stationary problems. Program summary (2)Title of program: imagtimecir.F Title of electronic file: imagtimecir.tar.gz Catalogue identifier: Program summary URL: Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Distribution format: tar.gz Computers: PC/Linux, workstation/UNIX Maximum RAM memory: 1 GByte Programming language used: Fortran 77 Typical running time: Minutes on a medium PC Unusual features: None Nature of physical problem: This program is designed to solve the time-dependent Gross-Pitaevskii nonlinear partial differential equation in two-space dimensions with a circularly-symmetric trap. The Gross-Pitaevskii equation describes the properties of a dilute trapped Bose-Einstein condensate. Method of solution: The time-dependent Gross-Pitaevskii equation is solved by the split-step Crank-Nicolson method by discretizing in space and time. The discretized equation is then solved by propagation in imaginary time over small time steps. The method yields the solution of stationary problems. Program summary (3)Title of program: imagtimesph.F Title of electronic file: imagtimesph.tar.gz Catalogue identifier: Program summary URL: Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Distribution format: tar.gz Computers: PC/Linux, workstation/UNIX Maximum RAM memory: 1 GByte Programming language used: Fortran 77 Typical running time: Minutes on a medium PC Unusual features: None Nature of physical problem: This program is designed to solve the time-dependent Gross-Pitaevskii nonlinear partial differential equation in three-space dimensions with a spherically-symmetric trap. The Gross-Pitaevskii equation describes the properties of a dilute trapped Bose-Einstein condensate. Method of solution: The time-dependent Gross-Pitaevskii equation is solved by the split-step Crank-Nicolson method by discretizing in space and time. The discretized equation is then solved by propagation in imaginary time over small time steps. The method yields the solution of stationary problems. Program summary (4)Title of program: realtime1d.F Title of electronic file: realtime1d.tar.gz Catalogue identifier: Program summary URL: Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Distribution format: tar.gz Computers: PC/Linux, workstation/UNIX Maximum RAM memory: 2 GByte Programming language used: Fortran 77 Typical running time: Minutes on a medium PC Unusual features: None Nature of physical problem: This program is designed to solve the time-dependent Gross-Pitaevskii nonlinear partial differential equation in one-space dimension with a harmonic trap. The Gross-Pitaevskii equation describes the properties of a dilute trapped Bose-Einstein condensate. Method of solution: The time-dependent Gross-Pitaevskii equation is solved by the split-step Crank-Nicolson method by discretizing in space and time. The discretized equation is then solved by propagation in real time over small time steps. The method yields the solution of stationary and non-stationary problems. Program summary (5)Title of program: realtimecir.F Title of electronic file: realtimecir.tar.gz Catalogue identifier: Program summary URL: Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Distribution format: tar.gz Computers: PC/Linux, workstation/UNIX Maximum RAM memory: 2 GByte Programming language used: Fortran 77 Typical running time: Minutes on a medium PC Unusual features: None Nature of physical problem: This program is designed to solve the time-dependent Gross-Pitaevskii nonlinear partial differential equation in two-space dimensions with a circularly-symmetric trap. The Gross-Pitaevskii equation describes the properties of a dilute trapped Bose-Einstein condensate. Method of solution: The time-dependent Gross-Pitaevskii equation is solved by the split-step Crank-Nicolson method by discretizing in space and time. The discretized equation is then solved by propagation in real time over small time steps. The method yields the solution of stationary and non-stationary problems. Program summary (6)Title of program: realtimesph.F Title of electronic file: realtimesph.tar.gz Catalogue identifier: Program summary URL: Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Distribution format: tar.gz Computers: PC/Linux, workstation/UNIX Maximum RAM memory: 2 GByte Programming language used: Fortran 77 Typical running time: Minutes on a medium PC Unusual features: None Nature of physical problem: This program is designed to solve the time-dependent Gross-Pitaevskii nonlinear partial differential equation in three-space dimensions with a spherically-symmetric trap. The Gross-Pitaevskii equation describes the properties of a dilute trapped Bose-Einstein condensate. Method of solution: The time-dependent Gross-Pitaevskii equation is solved by the split-step Crank-Nicolson method by discretizing in space and time. The discretized equation is then solved by propagation in real time over small time steps. The method yields the solution of stationary and non-stationary problems. Program summary (7)Title of programs: imagtimeaxial.F and imagtimeaxial.f90 Title of electronic file: imagtimeaxial.tar.gz Catalogue identifier: Program summary URL: Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Distribution format: tar.gz Computers: PC/Linux, workstation/UNIX Maximum RAM memory: 2 GByte Programming language used: Fortran 77 and Fortran 90 Typical running time: Few hours on a medium PC Unusual features: None Nature of physical problem: This program is designed to solve the time-dependent Gross-Pitaevskii nonlinear partial differential equation in three-space dimensions with an axially-symmetric trap. The Gross-Pitaevskii equation describes the properties of a dilute trapped Bose-Einstein condensate. Method of solution: The time-dependent Gross-Pitaevskii equation is solved by the split-step Crank-Nicolson method by discretizing in space and time. The discretized equation is then solved by propagation in imaginary time over small time steps. The method yields the solution of stationary problems. Program summary (8)Title of program: imagtime2d.F and imagtime2d.f90 Title of electronic file: imagtime2d.tar.gz Catalogue identifier: Program summary URL: Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Distribution format: tar.gz Computers: PC/Linux, workstation/UNIX Maximum RAM memory: 2 GByte Programming language used: Fortran 77 and Fortran 90 Typical running time: Few hours on a medium PC Unusual features: None Nature of physical problem: This program is designed to solve the time-dependent Gross-Pitaevskii nonlinear partial differential equation in two-space dimensions with an anisotropic trap. The Gross-Pitaevskii equation describes the properties of a dilute trapped Bose-Einstein condensate. Method of solution: The time-dependent Gross-Pitaevskii equation is solved by the split-step Crank-Nicolson method by discretizing in space and time. The discretized equation is then solved by propagation in imaginary time over small time steps. The method yields the solution of stationary problems. Program summary (9)Title of program: realtimeaxial.F and realtimeaxial.f90 Title of electronic file: realtimeaxial.tar.gz Catalogue identifier: Program summary URL: Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Distribution format: tar.gz Computers: PC/Linux, workstation/UNIX Maximum RAM memory: 4 GByte Programming language used: Fortran 77 and Fortran 90 Typical running time Hours on a medium PC Unusual features: None Nature of physical problem: This program is designed to solve the time-dependent Gross-Pitaevskii nonlinear partial differential equation in three-space dimensions with an axially-symmetric trap. The Gross-Pitaevskii equation describes the properties of a dilute trapped Bose-Einstein condensate. Method of solution: The time-dependent Gross-Pitaevskii equation is solved by the split-step Crank-Nicolson method by discretizing in space and time. The discretized equation is then solved by propagation in real time over small time steps. The method yields the solution of stationary and non-stationary problems. Program summary (10)Title of program: realtime2d.F and realtime2d.f90 Title of electronic file: realtime2d.tar.gz Catalogue identifier: Program summary URL: Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Distribution format: tar.gz Computers: PC/Linux, workstation/UNIX Maximum RAM memory: 4 GByte Programming language used: Fortran 77 and Fortran 90 Typical running time: Hours on a medium PC Unusual features: None Nature of physical problem: This program is designed to solve the time-dependent Gross-Pitaevskii nonlinear partial differential equation in two-space dimensions with an anisotropic trap. The Gross-Pitaevskii equation describes the properties of a dilute trapped Bose-Einstein condensate. Method of solution: The time-dependent Gross-Pitaevskii equation is solved by the split-step Crank-Nicolson method by discretizing in space and time. The discretized equation is then solved by propagation in real time over small time steps. The method yields the solution of stationary and non-stationary problems. Program summary (11)Title of program: imagtime3d.F and imagtime3d.f90 Title of electronic file: imagtime3d.tar.gz Catalogue identifier: Program summary URL: Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Distribution format: tar.gz Computers: PC/Linux, workstation/UNIX Maximum RAM memory: 4 GByte Programming language used: Fortran 77 and Fortran 90 Typical running time: Few days on a medium PC Unusual features: None Nature of physical problem: This program is designed to solve the time-dependent Gross-Pitaevskii nonlinear partial differential equation in three-space dimensions with an anisotropic trap. The Gross-Pitaevskii equation describes the properties of a dilute trapped Bose-Einstein condensate. Method of solution: The time-dependent Gross-Pitaevskii equation is solved by the split-step Crank-Nicolson method by discretizing in space and time. The discretized equation is then solved by propagation in imaginary time over small time steps. The method yields the solution of stationary problems. Program summary (12)Title of program: realtime3d.F and realtime3d.f90 Title of electronic file: realtime3d.tar.gz Catalogue identifier: Program summary URL: Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Distribution format: tar.gz Computers: PC/Linux, workstation/UNIX Maximum Ram Memory: 8 GByte Programming language used: Fortran 77 and Fortran 90 Typical running time: Days on a medium PC Unusual features: None Nature of physical problem: This program is designed to solve the time-dependent Gross-Pitaevskii nonlinear partial differential equation in three-space dimensions with an anisotropic trap. The Gross-Pitaevskii equation describes the properties of a dilute trapped Bose-Einstein condensate. Method of solution: The time-dependent Gross-Pitaevskii equation is solved by the split-step Crank-Nicolson method by discretizing in space and time. The discretized equation is then solved by propagation in real time over small time steps. The method yields the solution of stationary and non-stationary problems.
Fast Time and Space Parallel Algorithms for Solution of Parabolic Partial Differential Equations
NASA Technical Reports Server (NTRS)
Fijany, Amir
1993-01-01
In this paper, fast time- and Space -Parallel agorithms for solution of linear parabolic PDEs are developed. It is shown that the seemingly strictly serial iterations of the time-stepping procedure for solution of the problem can be completed decoupled.
Mixed Integer Programming and Heuristic Scheduling for Space Communication
NASA Technical Reports Server (NTRS)
Lee, Charles H.; Cheung, Kar-Ming
2013-01-01
Optimal planning and scheduling for a communication network was created where the nodes within the network are communicating at the highest possible rates while meeting the mission requirements and operational constraints. The planning and scheduling problem was formulated in the framework of Mixed Integer Programming (MIP) to introduce a special penalty function to convert the MIP problem into a continuous optimization problem, and to solve the constrained optimization problem using heuristic optimization. The communication network consists of space and ground assets with the link dynamics between any two assets varying with respect to time, distance, and telecom configurations. One asset could be communicating with another at very high data rates at one time, and at other times, communication is impossible, as the asset could be inaccessible from the network due to planetary occultation. Based on the network's geometric dynamics and link capabilities, the start time, end time, and link configuration of each view period are selected to maximize the communication efficiency within the network. Mathematical formulations for the constrained mixed integer optimization problem were derived, and efficient analytical and numerical techniques were developed to find the optimal solution. By setting up the problem using MIP, the search space for the optimization problem is reduced significantly, thereby speeding up the solution process. The ratio of the dimension of the traditional method over the proposed formulation is approximately an order N (single) to 2*N (arraying), where N is the number of receiving antennas of a node. By introducing a special penalty function, the MIP problem with non-differentiable cost function and nonlinear constraints can be converted into a continuous variable problem, whose solution is possible.
Astronautics and psychology: recommendations for the psychological training of astronauts.
Haupt, G F
1991-11-01
The methods presently applied in the psychological training of astronauts are based on the principle of ensuring maximum performance of astronauts during missions. The shortcomings are obvious since those undergoing training provide nothing but the best ability to cope with Earth problem situations and add simply an experience of space problem situations as they are presently conceived. Earth attitudes and Earth behaviour remain and are simply modified. Through the utilization of interdisciplinary space knowledge a much higher degree of problem anticipation could be achieved and the astronaut be psychologically transformed into a space-being. This would at the same time stimulate interdisciplinary space research. The interdisciplinary space knowledge already available suggests that space requires not only physical and mental adjustments, but a profoundly new relationship with life.
From Discrete Space-Time to Minkowski Space: Basic Mechanisms, Methods and Perspectives
NASA Astrophysics Data System (ADS)
Finster, Felix
This survey article reviews recent results on fermion systems in discrete space-time and corresponding systems in Minkowski space. After a basic introduction to the discrete setting, we explain a mechanism of spontaneous symmetry breaking which leads to the emergence of a discrete causal structure. As methods to study the transition between discrete space-time and Minkowski space, we describe a lattice model for a static and isotropic space-time, outline the analysis of regularization tails of vacuum Dirac sea configurations, and introduce a Lorentz invariant action for the masses of the Dirac seas. We mention the method of the continuum limit, which allows to analyze interacting systems. Open problems are discussed.
NASA Astrophysics Data System (ADS)
Kel'manov, A. V.; Khandeev, V. I.
2016-02-01
The strongly NP-hard problem of partitioning a finite set of points of Euclidean space into two clusters of given sizes (cardinalities) minimizing the sum (over both clusters) of the intracluster sums of squared distances from the elements of the clusters to their centers is considered. It is assumed that the center of one of the sought clusters is specified at the desired (arbitrary) point of space (without loss of generality, at the origin), while the center of the other one is unknown and determined as the mean value over all elements of this cluster. It is shown that unless P = NP, there is no fully polynomial-time approximation scheme for this problem, and such a scheme is substantiated in the case of a fixed space dimension.
Robust on-off pulse control of flexible space vehicles
NASA Technical Reports Server (NTRS)
Wie, Bong; Sinha, Ravi
1993-01-01
The on-off reaction jet control system is often used for attitude and orbital maneuvering of various spacecraft. Future space vehicles such as the orbital transfer vehicles, orbital maneuvering vehicles, and space station will extensively use reaction jets for orbital maneuvering and attitude stabilization. The proposed robust fuel- and time-optimal control algorithm is used for a three-mass spacing model of flexible spacecraft. A fuel-efficient on-off control logic is developed for robust rest-to-rest maneuver of a flexible vehicle with minimum excitation of structural modes. The first part of this report is concerned with the problem of selecting a proper pair of jets for practical trade-offs among the maneuvering time, fuel consumption, structural mode excitation, and performance robustness. A time-optimal control problem subject to parameter robustness constraints is formulated and solved. The second part of this report deals with obtaining parameter insensitive fuel- and time- optimal control inputs by solving a constrained optimization problem subject to robustness constraints. It is shown that sensitivity to modeling errors can be significantly reduced by the proposed, robustified open-loop control approach. The final part of this report deals with sliding mode control design for uncertain flexible structures. The benchmark problem of a flexible structure is used as an example for the feedback sliding mode controller design with bounded control inputs and robustness to parameter variations is investigated.
Mesh refinement strategy for optimal control problems
NASA Astrophysics Data System (ADS)
Paiva, L. T.; Fontes, F. A. C. C.
2013-10-01
Direct methods are becoming the most used technique to solve nonlinear optimal control problems. Regular time meshes having equidistant spacing are frequently used. However, in some cases these meshes cannot cope accurately with nonlinear behavior. One way to improve the solution is to select a new mesh with a greater number of nodes. Another way, involves adaptive mesh refinement. In this case, the mesh nodes have non equidistant spacing which allow a non uniform nodes collocation. In the method presented in this paper, a time mesh refinement strategy based on the local error is developed. After computing a solution in a coarse mesh, the local error is evaluated, which gives information about the subintervals of time domain where refinement is needed. This procedure is repeated until the local error reaches a user-specified threshold. The technique is applied to solve the car-like vehicle problem aiming minimum consumption. The approach developed in this paper leads to results with greater accuracy and yet with lower overall computational time as compared to using a time meshes having equidistant spacing.
Computational complexity of ecological and evolutionary spatial dynamics
Ibsen-Jensen, Rasmus; Chatterjee, Krishnendu; Nowak, Martin A.
2015-01-01
There are deep, yet largely unexplored, connections between computer science and biology. Both disciplines examine how information proliferates in time and space. Central results in computer science describe the complexity of algorithms that solve certain classes of problems. An algorithm is deemed efficient if it can solve a problem in polynomial time, which means the running time of the algorithm is a polynomial function of the length of the input. There are classes of harder problems for which the fastest possible algorithm requires exponential time. Another criterion is the space requirement of the algorithm. There is a crucial distinction between algorithms that can find a solution, verify a solution, or list several distinct solutions in given time and space. The complexity hierarchy that is generated in this way is the foundation of theoretical computer science. Precise complexity results can be notoriously difficult. The famous question whether polynomial time equals nondeterministic polynomial time (i.e., P = NP) is one of the hardest open problems in computer science and all of mathematics. Here, we consider simple processes of ecological and evolutionary spatial dynamics. The basic question is: What is the probability that a new invader (or a new mutant) will take over a resident population? We derive precise complexity results for a variety of scenarios. We therefore show that some fundamental questions in this area cannot be answered by simple equations (assuming that P is not equal to NP). PMID:26644569
Parallel algorithms for mapping pipelined and parallel computations
NASA Technical Reports Server (NTRS)
Nicol, David M.
1988-01-01
Many computational problems in image processing, signal processing, and scientific computing are naturally structured for either pipelined or parallel computation. When mapping such problems onto a parallel architecture it is often necessary to aggregate an obvious problem decomposition. Even in this context the general mapping problem is known to be computationally intractable, but recent advances have been made in identifying classes of problems and architectures for which optimal solutions can be found in polynomial time. Among these, the mapping of pipelined or parallel computations onto linear array, shared memory, and host-satellite systems figures prominently. This paper extends that work first by showing how to improve existing serial mapping algorithms. These improvements have significantly lower time and space complexities: in one case a published O(nm sup 3) time algorithm for mapping m modules onto n processors is reduced to an O(nm log m) time complexity, and its space requirements reduced from O(nm sup 2) to O(m). Run time complexity is further reduced with parallel mapping algorithms based on these improvements, which run on the architecture for which they create the mappings.
A space-time lower-upper symmetric Gauss-Seidel scheme for the time-spectral method
NASA Astrophysics Data System (ADS)
Zhan, Lei; Xiong, Juntao; Liu, Feng
2016-05-01
The time-spectral method (TSM) offers the advantage of increased order of accuracy compared to methods using finite-difference in time for periodic unsteady flow problems. Explicit Runge-Kutta pseudo-time marching and implicit schemes have been developed to solve iteratively the space-time coupled nonlinear equations resulting from TSM. Convergence of the explicit schemes is slow because of the stringent time-step limit. Many implicit methods have been developed for TSM. Their computational efficiency is, however, still limited in practice because of delayed implicit temporal coupling, multiple iterative loops, costly matrix operations, or lack of strong diagonal dominance of the implicit operator matrix. To overcome these shortcomings, an efficient space-time lower-upper symmetric Gauss-Seidel (ST-LU-SGS) implicit scheme with multigrid acceleration is presented. In this scheme, the implicit temporal coupling term is split as one additional dimension of space in the LU-SGS sweeps. To improve numerical stability for periodic flows with high frequency, a modification to the ST-LU-SGS scheme is proposed. Numerical results show that fast convergence is achieved using large or even infinite Courant-Friedrichs-Lewy (CFL) numbers for unsteady flow problems with moderately high frequency and with the use of moderately high numbers of time intervals. The ST-LU-SGS implicit scheme is also found to work well in calculating periodic flow problems where the frequency is not known a priori and needed to be determined by using a combined Fourier analysis and gradient-based search algorithm.
Deconvolution of mixing time series on a graph
Blocker, Alexander W.; Airoldi, Edoardo M.
2013-01-01
In many applications we are interested in making inference on latent time series from indirect measurements, which are often low-dimensional projections resulting from mixing or aggregation. Positron emission tomography, super-resolution, and network traffic monitoring are some examples. Inference in such settings requires solving a sequence of ill-posed inverse problems, yt = Axt, where the projection mechanism provides information on A. We consider problems in which A specifies mixing on a graph of times series that are bursty and sparse. We develop a multilevel state-space model for mixing times series and an efficient approach to inference. A simple model is used to calibrate regularization parameters that lead to efficient inference in the multilevel state-space model. We apply this method to the problem of estimating point-to-point traffic flows on a network from aggregate measurements. Our solution outperforms existing methods for this problem, and our two-stage approach suggests an efficient inference strategy for multilevel models of multivariate time series. PMID:25309135
NASA Technical Reports Server (NTRS)
1997-01-01
Kennedy Space Center specialists aided Space, Energy, Time Saving (SETS) Systems, Inc. in working out the problems they encountered with their new electronic "tankless" water heater. The flow switch design suffered intermittent problems. Hiring several testing and engineering firms produced only graphs, printouts, and a large expense, but no solutions. Then through the Kennedy Space Center/State of Florida Technology Outreach Program, SETS was referred to Michael Brooks, a 21-year space program veteran and flowmeter expert. Run throughout Florida to provide technical service to businesses at no cost, the program applies scientific and engineering expertise originally developed for space applications to the Florida business community. Brooks discovered several key problems, resulting in a new design that turned out to be simpler, yielding a 63 percent reduction in labor and material costs over the old design.
Test Equal Bending by Gravity for Space and Time
NASA Astrophysics Data System (ADS)
Sweetser, Douglas
2009-05-01
For the simplest problem of gravity - a static, non-rotating, spherically symmetric source - the solution for spacetime bending around the Sun should be evenly split between time and space. That is true to first order in M/R, and confirmed by experiment. At second order, general relativity predicts different amounts of contribution from time and space without a physical justification. I show an exponential metric is consistent with light bending to first order, measurably different at second order. All terms to all orders show equal contributions from space and time. Beautiful minimalism is Nature's way.
A Space Affine Matching Approach to fMRI Time Series Analysis.
Chen, Liang; Zhang, Weishi; Liu, Hongbo; Feng, Shigang; Chen, C L Philip; Wang, Huili
2016-07-01
For fMRI time series analysis, an important challenge is to overcome the potential delay between hemodynamic response signal and cognitive stimuli signal, namely the same frequency but different phase (SFDP) problem. In this paper, a novel space affine matching feature is presented by introducing the time domain and frequency domain features. The time domain feature is used to discern different stimuli, while the frequency domain feature to eliminate the delay. And then we propose a space affine matching (SAM) algorithm to match fMRI time series by our affine feature, in which a normal vector is estimated using gradient descent to explore the time series matching optimally. The experimental results illustrate that the SAM algorithm is insensitive to the delay between the hemodynamic response signal and the cognitive stimuli signal. Our approach significantly outperforms GLM method while there exists the delay. The approach can help us solve the SFDP problem in fMRI time series matching and thus of great promise to reveal brain dynamics.
NASA Astrophysics Data System (ADS)
Shiangjen, Kanokwatt; Chaijaruwanich, Jeerayut; Srisujjalertwaja, Wijak; Unachak, Prakarn; Somhom, Samerkae
2018-02-01
This article presents an efficient heuristic placement algorithm, namely, a bidirectional heuristic placement, for solving the two-dimensional rectangular knapsack packing problem. The heuristic demonstrates ways to maximize space utilization by fitting the appropriate rectangle from both sides of the wall of the current residual space layer by layer. The iterative local search along with a shift strategy is developed and applied to the heuristic to balance the exploitation and exploration tasks in the solution space without the tuning of any parameters. The experimental results on many scales of packing problems show that this approach can produce high-quality solutions for most of the benchmark datasets, especially for large-scale problems, within a reasonable duration of computational time.
Heuristics in Problem Solving: The Role of Direction in Controlling Search Space
ERIC Educational Resources Information Center
Chu, Yun; Li, Zheng; Su, Yong; Pizlo, Zygmunt
2010-01-01
Isomorphs of a puzzle called m+m resulted in faster solution times and an easily reproduced solution path in a labeled version of the problem compared to a more difficult binary version. We conjecture that performance is related to a type of heuristic called direction that not only constrains search space in the labeled version, but also…
Operability engineering in the Deep Space Network
NASA Technical Reports Server (NTRS)
Wilkinson, Belinda
1993-01-01
Many operability problems exist at the three Deep Space Communications Complexes (DSCC's) of the Deep Space Network (DSN). Four years ago, the position of DSN Operability Engineer was created to provide the opportunity for someone to take a system-level approach to solving these problems. Since that time, a process has been developed for personnel and development engineers and for enforcing user interface standards in software designed for the DSCC's. Plans are for the participation of operations personnel in the product life-cycle to expand in the future.
NASA Astrophysics Data System (ADS)
Kel'manov, A. V.; Motkova, A. V.
2018-01-01
A strongly NP-hard problem of partitioning a finite set of points of Euclidean space into two clusters is considered. The solution criterion is the minimum of the sum (over both clusters) of weighted sums of squared distances from the elements of each cluster to its geometric center. The weights of the sums are equal to the cardinalities of the desired clusters. The center of one cluster is given as input, while the center of the other is unknown and is determined as the point of space equal to the mean of the cluster elements. A version of the problem is analyzed in which the cardinalities of the clusters are given as input. A polynomial-time 2-approximation algorithm for solving the problem is constructed.
Time-dependent wave splitting and source separation
NASA Astrophysics Data System (ADS)
Grote, Marcus J.; Kray, Marie; Nataf, Frédéric; Assous, Franck
2017-02-01
Starting from classical absorbing boundary conditions, we propose a method for the separation of time-dependent scattered wave fields due to multiple sources or obstacles. In contrast to previous techniques, our method is local in space and time, deterministic, and avoids a priori assumptions on the frequency spectrum of the signal. Numerical examples in two space dimensions illustrate the usefulness of wave splitting for time-dependent scattering problems.
NASA Astrophysics Data System (ADS)
Tian, X.; Zhang, Y.
2018-03-01
Herglotz variational principle, in which the functional is defined by a differential equation, generalizes the classical ones defining the functional by an integral. The principle gives a variational principle description of nonconservative systems even when the Lagrangian is independent of time. This paper focuses on studying the Noether's theorem and its inverse of a Birkhoffian system in event space based on the Herglotz variational problem. Firstly, according to the Herglotz variational principle of a Birkhoffian system, the principle of a Birkhoffian system in event space is established. Secondly, its parametric equations and two basic formulae for the variation of Pfaff-Herglotz action of a Birkhoffian system in event space are obtained. Furthermore, the definition and criteria of Noether symmetry of the Birkhoffian system in event space based on the Herglotz variational problem are given. Then, according to the relationship between the Noether symmetry and conserved quantity, the Noether's theorem is derived. Under classical conditions, Noether's theorem of a Birkhoffian system in event space based on the Herglotz variational problem reduces to the classical ones. In addition, Noether's inverse theorem of the Birkhoffian system in event space based on the Herglotz variational problem is also obtained. In the end of the paper, an example is given to illustrate the application of the results.
Exact Solutions to Time-dependent Mdps
NASA Technical Reports Server (NTRS)
Boyan, Justin A.; Littman, Michael L.
2000-01-01
We describe an extension of the Markov decision process model in which a continuous time dimension is included in the state space. This allows for the representation and exact solution of a wide range of problems in which transitions or rewards vary over time. We examine problems based on route planning with public transportation and telescope observation scheduling.
Precision time distribution within a deep space communications complex
NASA Technical Reports Server (NTRS)
Curtright, J. B.
1972-01-01
The Precision Time Distribution System (PTDS) at the Golstone Deep Space Communications Complex is a practical application of existing technology to the solution of a local problem. The problem was to synchronize four station timing systems to a master source with a relative accuracy consistently and significantly better than 10 microseconds. The solution involved combining a precision timing source, an automatic error detection assembly and a microwave distribution network into an operational system. Upon activation of the completed PTDS two years ago, synchronization accuracy at Goldstone (two station relative) was improved by an order of magnitude. It is felt that the validation of the PTDS mechanization is now completed. Other facilities which have site dispersion and synchronization accuracy requirements similar to Goldstone may find the PTDS mechanization useful in solving their problem. At present, the two station relative synchronization accuracy at Goldstone is better than one microsecond.
Dark matter: a problem in relativistic metrology?
NASA Astrophysics Data System (ADS)
Lusanna, Luca
2017-05-01
Besides the tidal degrees of freedom of Einstein general relativity (GR) (namely the two polarizations of gravitational waves after linearization of the theory) there are the inertial gauge ones connected with the freedom in the choice of the 4-coordinates of the space-time, i.e. in the choice of the notions of time and 3-space (the 3+1 splitting of space-time) and in their use to define a non-inertial frame (the inertial ones being forbidden by the equivalence principle) by means of a set of conventions for the relativistic metrology of the space-time (like the GPS ones near the Earth). The canonical York basis of canonical ADM gravity allows us to identify the Hamiltonian inertial gauge variables in globally hyperbolic asymptotically Minkowskian space-times without super-translations and to define the family of non-harmonic Schwinger time gauges. In these 3+1 splittings of space-time the freedom in the choice of time (the problem of clock synchronization) is described by the inertial gauge variable York time (the trace of the extrinsic curvature of the instantaneous 3-spaces). This inertial gauge freedom and the non-Euclidean nature of the instantaneous 3-spaces required by the equivalence principle need to be incorporated as metrical conventions in a relativistic suitable extension of the existing (essentially Galilean) ICRS celestial reference system. In this paper I make a short review of the existing possibilities to explain the presence of dark matter (or at least of part of it) as a relativistic inertial effect induced by the non- Euclidean nature of the 3-spaces. After a Hamiltonian Post-Minkowskian (HPM) linearization of canonical ADM tetrad gravity with particles, having equal inertial and gravitational masses, as matter, followed by a Post-Newtonian (PN) expansion, we find that the Newtonian equality of inertial and gravitational masses breaks down and that the inertial gauge York time produces an increment of the inertial masses explaining at least part of what is called dark matter in all its astrophysical signatures.
NASA Technical Reports Server (NTRS)
Wang, Xiao-Yen; Chang, Sin-Chung; Jorgenson, Philip C. E.
2000-01-01
The benchmark problems in Category 1 (Internal Propagation) of the third Computational Aeroacoustics (CAA) Work-shop sponsored by NASA Glenn Research Center are solved using the space-time conservation element and solution element (CE/SE) method. The first problem addresses the propagation of sound waves through a nearly choked transonic nozzle. The second one concerns shock-sound interaction in a supersonic nozzle. A quasi one-dimension CE/SE Euler solver for a nonuniform mesh is developed and employed to solve both problems. Numerical solutions are compared with the analytical solution for both problems. It is demonstrated that the CE/SE method is capable of solving aeroacoustic problems with/without shock waves in a simple way. Furthermore, the simple nonreflecting boundary condition used in the CE/SE method which is not based on the characteristic theory works very well.
From black holes to quantum gravity
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sanchez, N.
1987-01-01
Since modern physics now deals simultaneously with quantum theory, general relativity, cosmology and elementary particle physics, this volume caters to the need for a book of such a wide scope of interest. Aspects of grand unification, the thermodynamics of space-time, the loss of quantum coherence and the problem of time are expertly treated within a unified presentation. Contents: Introduction; The Global Structure of Space-time in the Classical Theory of General Relativity; Connection between the Structure of the Space-time and the Propagation of Quantum Fields; The Different Approaches to Quantization; Outlook and Conclusions.
Optimal use of human and machine resources for Space Station assembly operations
NASA Technical Reports Server (NTRS)
Parrish, Joseph C.
1988-01-01
This paper investigates the issues involved in determining the best mix of human and machine resources for assembly of the Space Station. It presents the current Station assembly sequence, along with descriptions of the available assembly resources. A number of methodologies for optimizing the human/machine tradeoff problem have been developed, but the Space Station assembly offers some unique issues that have not yet been addressed. These include a strong constraint on available EVA time for early flights and a phased deployment of assembly resources over time. A methodology for incorporating the previously developed decision methods to the special case of the Space Station is presented. This methodology emphasizes an application of multiple qualitative and quantitative techniques, including simulation and decision analysis, for producing an objective, robust solution to the tradeoff problem.
Approximating the 0-1 Multiple Knapsack Problem with Agent Decomposition and Market Negotiation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Smolinski, B.
The 0-1 multiple knapsack problem appears in many domains from financial portfolio management to cargo ship stowing. Methods for solving it range from approximate algorithms, such as greedy algorithms, to exact algorithms, such as branch and bound. Approximate algorithms have no bounds on how poorly they perform and exact algorithms can suffer from exponential time and space complexities with large data sets. This paper introduces a market model based on agent decomposition and market auctions for approximating the 0-1 multiple knapsack problem, and an algorithm that implements the model (M(x)). M(x) traverses the solution space rather than getting caught inmore » a local maximum, overcoming an inherent problem of many greedy algorithms. The use of agents ensures that infeasible solutions are not considered while traversing the solution space and that traversal of the solution space is not just random, but is also directed. M(x) is compared to a bound and bound algorithm (BB) and a simple greedy algorithm with a random shuffle (G(x)). The results suggest that M(x) is a good algorithm for approximating the 0-1 Multiple Knapsack problem. M(x) almost always found solutions that were close to optimal in a fraction of the time it took BB to run and with much less memory on large test data sets. M(x) usually performed better than G(x) on hard problems with correlated data.« less
Mobile robot motion estimation using Hough transform
NASA Astrophysics Data System (ADS)
Aldoshkin, D. N.; Yamskikh, T. N.; Tsarev, R. Yu
2018-05-01
This paper proposes an algorithm for estimation of mobile robot motion. The geometry of surrounding space is described with range scans (samples of distance measurements) taken by the mobile robot’s range sensors. A similar sample of space geometry in any arbitrary preceding moment of time or the environment map can be used as a reference. The suggested algorithm is invariant to isotropic scaling of samples or map that allows using samples measured in different units and maps made at different scales. The algorithm is based on Hough transform: it maps from measurement space to a straight-line parameters space. In the straight-line parameters, space the problems of estimating rotation, scaling and translation are solved separately breaking down a problem of estimating mobile robot localization into three smaller independent problems. The specific feature of the algorithm presented is its robustness to noise and outliers inherited from Hough transform. The prototype of the system of mobile robot orientation is described.
A PC based time domain reflectometer for space station cable fault isolation
NASA Technical Reports Server (NTRS)
Pham, Michael; McClean, Marty; Hossain, Sabbir; Vo, Peter; Kouns, Ken
1994-01-01
Significant problems are faced by astronauts on orbit in the Space Station when trying to locate electrical faults in multi-segment avionics and communication cables. These problems necessitate the development of an automated portable device that will detect and locate cable faults using the pulse-echo technique known as Time Domain Reflectometry. A breadboard time domain reflectometer (TDR) circuit board was designed and developed at the NASA-JSC. The TDR board works in conjunction with a GRiD lap-top computer to automate the fault detection and isolation process. A software program was written to automatically display the nature and location of any possible faults. The breadboard system can isolate open circuit and short circuit faults within two feet in a typical space station cable configuration. Follow-on efforts planned for 1994 will produce a compact, portable prototype Space Station TDR capable of automated switching in multi-conductor cables for high fidelity evaluation. This device has many possible commercial applications, including commercial and military aircraft avionics, cable TV, telephone, communication, information and computer network systems. This paper describes the principle of time domain reflectometry and the methodology for on-orbit avionics utility distribution system repair, utilizing the newly developed device called the Space Station Time Domain Reflectometer (SSTDR).
Adaptive density trajectory cluster based on time and space distance
NASA Astrophysics Data System (ADS)
Liu, Fagui; Zhang, Zhijie
2017-10-01
There are some hotspot problems remaining in trajectory cluster for discovering mobile behavior regularity, such as the computation of distance between sub trajectories, the setting of parameter values in cluster algorithm and the uncertainty/boundary problem of data set. As a result, based on the time and space, this paper tries to define the calculation method of distance between sub trajectories. The significance of distance calculation for sub trajectories is to clearly reveal the differences in moving trajectories and to promote the accuracy of cluster algorithm. Besides, a novel adaptive density trajectory cluster algorithm is proposed, in which cluster radius is computed through using the density of data distribution. In addition, cluster centers and number are selected by a certain strategy automatically, and uncertainty/boundary problem of data set is solved by designed weighted rough c-means. Experimental results demonstrate that the proposed algorithm can perform the fuzzy trajectory cluster effectively on the basis of the time and space distance, and obtain the optimal cluster centers and rich cluster results information adaptably for excavating the features of mobile behavior in mobile and sociology network.
Optimal stimulus scheduling for active estimation of evoked brain networks.
Kafashan, MohammadMehdi; Ching, ShiNung
2015-12-01
We consider the problem of optimal probing to learn connections in an evoked dynamic network. Such a network, in which each edge measures an input-output relationship between sites in sensor/actuator-space, is relevant to emerging applications in neural mapping and neural connectivity estimation. We show that the problem of scheduling nodes to a probe (i.e., stimulate) amounts to a problem of optimal sensor scheduling. By formulating the evoked network in state-space, we show that the solution to the greedy probing strategy has a convenient form and, under certain conditions, is optimal over a finite horizon. We adopt an expectation maximization technique to update the state-space parameters in an online fashion and demonstrate the efficacy of the overall approach in a series of detailed numerical examples. The proposed method provides a principled means to actively probe time-varying connections in neuronal networks. The overall method can be implemented in real time and is particularly well-suited to applications in stimulation-based cortical mapping in which the underlying network dynamics are changing over time.
Optimal stimulus scheduling for active estimation of evoked brain networks
NASA Astrophysics Data System (ADS)
Kafashan, MohammadMehdi; Ching, ShiNung
2015-12-01
Objective. We consider the problem of optimal probing to learn connections in an evoked dynamic network. Such a network, in which each edge measures an input-output relationship between sites in sensor/actuator-space, is relevant to emerging applications in neural mapping and neural connectivity estimation. Approach. We show that the problem of scheduling nodes to a probe (i.e., stimulate) amounts to a problem of optimal sensor scheduling. Main results. By formulating the evoked network in state-space, we show that the solution to the greedy probing strategy has a convenient form and, under certain conditions, is optimal over a finite horizon. We adopt an expectation maximization technique to update the state-space parameters in an online fashion and demonstrate the efficacy of the overall approach in a series of detailed numerical examples. Significance. The proposed method provides a principled means to actively probe time-varying connections in neuronal networks. The overall method can be implemented in real time and is particularly well-suited to applications in stimulation-based cortical mapping in which the underlying network dynamics are changing over time.
An investigation of the use of temporal decomposition in space mission scheduling
NASA Technical Reports Server (NTRS)
Bullington, Stanley E.; Narayanan, Venkat
1994-01-01
This research involves an examination of techniques for solving scheduling problems in long-duration space missions. The mission timeline is broken up into several time segments, which are then scheduled incrementally. Three methods are presented for identifying the activities that are to be attempted within these segments. The first method is a mathematical model, which is presented primarily to illustrate the structure of the temporal decomposition problem. Since the mathematical model is bound to be computationally prohibitive for realistic problems, two heuristic assignment procedures are also presented. The first heuristic method is based on dispatching rules for activity selection, and the second heuristic assigns performances of a model evenly over timeline segments. These heuristics are tested using a sample Space Station mission and a Spacelab mission. The results are compared with those obtained by scheduling the missions without any problem decomposition. The applicability of this approach to large-scale mission scheduling problems is also discussed.
NASA Technical Reports Server (NTRS)
Gupta, R. N.; Moss, J. N.; Simmonds, A. L.
1982-01-01
Two flow-field codes employing the time- and space-marching numerical techniques were evaluated. Both methods were used to analyze the flow field around a massively blown Jupiter entry probe under perfect-gas conditions. In order to obtain a direct point-by-point comparison, the computations were made by using identical grids and turbulence models. For the same degree of accuracy, the space-marching scheme takes much less time as compared to the time-marching method and would appear to provide accurate results for the problems with nonequilibrium chemistry, free from the effect of local differences in time on the final solution which is inherent in time-marching methods. With the time-marching method, however, the solutions are obtainable for the realistic entry probe shapes with massive or uniform surface blowing rates; whereas, with the space-marching technique, it is difficult to obtain converged solutions for such flow conditions. The choice of the numerical method is, therefore, problem dependent. Both methods give equally good results for the cases where results are compared with experimental data.
Autonomous space processor for orbital debris
NASA Technical Reports Server (NTRS)
Ramohalli, Kumar; Marine, Micky; Colvin, James; Crockett, Richard; Sword, Lee; Putz, Jennifer; Woelfle, Sheri
1991-01-01
The development of an Autonomous Space Processor for Orbital Debris (ASPOD) was the goal. The nature of this craft, which will process, in situ, orbital debris using resources available in low Earth orbit (LEO) is explained. The serious problem of orbital debris is briefly described and the nature of the large debris population is outlined. The focus was on the development of a versatile robotic manipulator to augment an existing robotic arm, the incorporation of remote operation of the robotic arms, and the formulation of optimal (time and energy) trajectory planning algorithms for coordinated robotic arms. The mechanical design of the new arm is described in detail. The work envelope is explained showing the flexibility of the new design. Several telemetry communication systems are described which will enable the remote operation of the robotic arms. The trajectory planning algorithms are fully developed for both the time optimal and energy optimal problems. The time optimal problem is solved using phase plane techniques while the energy optimal problem is solved using dynamic programming.
NASA Astrophysics Data System (ADS)
Andonov, Zdravko
This R&D represent innovative multidimensional 6D-N(6n)D Space-Time (S-T) Methodology, 6D-6nD Coordinate Systems, 6D Equations, new 6D strategy and technology for development of Planetary Space Sciences, S-T Data Management and S-T Computational To-mography. . . The Methodology is actual for brain new RS Microwaves' Satellites and Compu-tational Tomography Systems development, aimed to defense sustainable Earth, Moon, & Sun System evolution. Especially, extremely important are innovations for monitoring and protec-tion of strategic threelateral system H-OH-H2O Hydrogen, Hydroxyl and Water), correspond-ing to RS VHRS (Very High Resolution Systems) of 1.420-1.657-22.089GHz microwaves. . . One of the Greatest Paradox and Challenge of World Science is the "transformation" of J. L. Lagrange 4D Space-Time (S-T) System to H. Minkovski 4D S-T System (O-X,Y,Z,icT) for Einstein's "Theory of Relativity". As a global result: -In contemporary Advanced Space Sciences there is not real adequate 4D-6D Space-Time Coordinate System and 6D Advanced Cosmos Strategy & Methodology for Multidimensional and Multitemporal Space-Time Data Management and Tomography. . . That's one of the top actual S-T Problems. Simple and optimal nD S-T Methodology discovery is extremely important for all Universities' Space Sci-ences' Education Programs, for advances in space research and especially -for all young Space Scientists R&D!... The top ten 21-Century Challenges ahead of Planetary and Space Sciences, Space Data Management and Computational Space Tomography, important for successfully de-velopment of Young Scientist Generations, are following: 1. R&D of W. R. Hamilton General Idea for transformation all Space Sciences to Time Sciences, beginning with 6D Eukonal for 6D anisotropic mediums & velocities. Development of IERS Earth & Space Systems (VLBI; LLR; GPS; SLR; DORIS Etc.) for Planetary-Space Data Management & Computational Planetary & Space Tomography. 2. R&D of S. W. Hawking Paradigm for 2D Complex Time and Quan-tum Wave Cosmology Paradigm for Decision of the Main Problem of Contemporary Physics. 3. R&D of Einstein-Minkowski Geodesies' Paradigm in the 4D-Space-Time Continuum to 6D-6nD Space-Time Continuum Paradigms and 6D S-T Equations. . . 4. R&D of Erwin Schrüdinger 4D S-T Universe' Evolutional Equation; It's David Bohm 4D generalization for anisotropic mediums and innovative 6D -for instantaneously quantum measurement -Bohm-Schrüdinger 6D S-T Universe' Evolutional Equation. 5. R&D of brain new 6D Planning of S-T Experi-ments, brain new 6D Space Technicks and Space Technology Generalizations, especially for 6D RS VHRS Research, Monitoring and 6D Computational Tomography. 6. R&D of "6D Euler-Poisson Equations" and "6D Kolmogorov Turbulence Theory" for GeoDynamics and for Space Dynamics as evolution of Gauss-Riemann Paradigms. 7. R&D of N. Boneff NASA RD for Asteroid "Eros" & Space Science' Laws Evolution. 8. R&D of H. Poincare Paradigm for Nature and Cosmos as 6D Group of Transferences. 9. R&D of K. Popoff N-Body General Problem & General Thermodynamic S-T Theory as Einstein-Prigogine-Landau' Paradigms Development. ü 10. R&D of 1st GUT since 1958 by N. S. Kalitzin (Kalitzin N. S., 1958: Uber eine einheitliche Feldtheorie. ZAHeidelberg-ARI, WZHUmnR-B., 7 (2), 207-215) and "Multitemporal Theory of Relativity" -With special applications to Photon Rockets and all Space-Time R&D. GENERAL CONCLUSION: Multidimensional Space-Time Methodology is advance in space research, corresponding to the IAF-IAA-COSPAR Innovative Strategy and R&D Programs -UNEP, UNDP, GEOSS, GMES, Etc.
Sources of spurious force oscillations from an immersed boundary method for moving-body problems
NASA Astrophysics Data System (ADS)
Lee, Jongho; Kim, Jungwoo; Choi, Haecheon; Yang, Kyung-Soo
2011-04-01
When a discrete-forcing immersed boundary method is applied to moving-body problems, it produces spurious force oscillations on a solid body. In the present study, we identify two sources of these force oscillations. One source is from the spatial discontinuity in the pressure across the immersed boundary when a grid point located inside a solid body becomes that of fluid with a body motion. The addition of mass source/sink together with momentum forcing proposed by Kim et al. [J. Kim, D. Kim, H. Choi, An immersed-boundary finite volume method for simulations of flow in complex geometries, Journal of Computational Physics 171 (2001) 132-150] reduces the spurious force oscillations by alleviating this pressure discontinuity. The other source is from the temporal discontinuity in the velocity at the grid points where fluid becomes solid with a body motion. The magnitude of velocity discontinuity decreases with decreasing the grid spacing near the immersed boundary. Four moving-body problems are simulated by varying the grid spacing at a fixed computational time step and at a constant CFL number, respectively. It is found that the spurious force oscillations decrease with decreasing the grid spacing and increasing the computational time step size, but they depend more on the grid spacing than on the computational time step size.
GRADIENT: Graph Analytic Approach for Discovering Irregular Events, Nascent and Temporal
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hogan, Emilie
2015-03-31
Finding a time-ordered signature within large graphs is a computationally complex problem due to the combinatorial explosion of potential patterns. GRADIENT is designed to search and understand that problem space.
GRADIENT: Graph Analytic Approach for Discovering Irregular Events, Nascent and Temporal
Hogan, Emilie
2018-01-16
Finding a time-ordered signature within large graphs is a computationally complex problem due to the combinatorial explosion of potential patterns. GRADIENT is designed to search and understand that problem space.
Tether Impact Rate Simulation and Prediction with Orbiting Satellites
NASA Technical Reports Server (NTRS)
Harrison, Jim
2002-01-01
Space elevators and other large space structures have been studied and proposed as worthwhile by futuristic space planners for at least a couple of decades. In June 1999 the Marshall Space Flight Center sponsored a Space Elevator workshop in Huntsville, Alabama, to bring together technical experts and advanced planners to discuss the current status and to define the magnitude of the technical and programmatic problems connected with the development of these massive space systems. One obvious problem that was identified, although not for the first time, were the collision probabilities between space elevators and orbital debris. Debate and uncertainty presently exist about the extent of the threat to these large structures, one in this study as large in size as a space elevator. We have tentatively concluded that orbital debris although a major concern not sufficient justification to curtail the study and development of futuristic new millennium concepts like the space elevators.
Power optimization of wireless media systems with space-time block codes.
Yousefi'zadeh, Homayoun; Jafarkhani, Hamid; Moshfeghi, Mehran
2004-07-01
We present analytical and numerical solutions to the problem of power control in wireless media systems with multiple antennas. We formulate a set of optimization problems aimed at minimizing total power consumption of wireless media systems subject to a given level of QoS and an available bit rate. Our formulation takes into consideration the power consumption related to source coding, channel coding, and transmission of multiple-transmit antennas. In our study, we consider Gauss-Markov and video source models, Rayleigh fading channels along with the Bernoulli/Gilbert-Elliott loss models, and space-time block codes.
Regge calculus and observations. II. Further applications.
NASA Astrophysics Data System (ADS)
Williams, Ruth M.; Ellis, G. F. R.
1984-11-01
The method, developed in an earlier paper, for tracing geodesies of particles and light rays through Regge calculus space-times, is applied to a number of problems in the Schwarzschild geometry. It is possible to obtain accurate predictions of light bending by taking sufficiently small Regge blocks. Calculations of perihelion precession, Thomas precession, and the distortion of a ball of fluid moving on a geodesic can also show good agreement with the analytic solution. However difficulties arise in obtaining accurate predictions for general orbits in these space-times. Applications to other problems in general relativity are discussed briefly.
On the initial value problem for the wave equation in Friedmann-Robertson-Walker space-times.
Abbasi, Bilal; Craig, Walter
2014-09-08
The propagator W ( t 0 , t 1 )( g , h ) for the wave equation in a given space-time takes initial data ( g ( x ), h ( x )) on a Cauchy surface {( t , x ) : t = t 0 } and evaluates the solution ( u ( t 1 , x ),∂ t u ( t 1 , x )) at other times t 1 . The Friedmann-Robertson-Walker space-times are defined for t 0 , t 1 >0, whereas for t 0 →0, there is a metric singularity. There is a spherical means representation for the general solution of the wave equation with the Friedmann-Robertson-Walker background metric in the three spatial dimensional cases of curvature K =0 and K =-1 given by S. Klainerman and P. Sarnak. We derive from the expression of their representation three results about the wave propagator for the Cauchy problem in these space-times. First, we give an elementary proof of the sharp rate of time decay of solutions with compactly supported data. Second, we observe that the sharp Huygens principle is not satisfied by solutions, unlike in the case of three-dimensional Minkowski space-time (the usual Huygens principle of finite propagation speed is satisfied, of course). Third, we show that for 0< t 0 < t the limit, [Formula: see text] exists, it is independent of h ( x ), and for all reasonable initial data g ( x ), it gives rise to a well-defined solution for all t >0 emanating from the space-time singularity at t =0. Under reflection t →- t , the Friedmann-Robertson-Walker metric gives a space-time metric for t <0 with a singular future at t =0, and the same solution formulae hold. We thus have constructed solutions u ( t , x ) of the wave equation in Friedmann-Robertson-Walker space-times which exist for all [Formula: see text] and [Formula: see text], where in conformally regularized coordinates, these solutions are continuous through the singularity t =0 of space-time, taking on specified data u (0,⋅)= g (⋅) at the singular time.
Representation of solution for fully nonlocal diffusion equations with deviation time variable
NASA Astrophysics Data System (ADS)
Drin, I. I.; Drin, S. S.; Drin, Ya. M.
2018-01-01
We prove the solvability of the Cauchy problem for a nonlocal heat equation which is of fractional order both in space and time. The representation formula for classical solutions for time- and space- fractional partial differential operator Dat + a2 (-Δ) γ/2 (0 <= α <= 1, γ ɛ (0, 2]) and deviation time variable is given in terms of the Fox H-function, using the step by step method.
User's guide to four-body and three-body trajectory optimization programs
NASA Technical Reports Server (NTRS)
Pu, C. L.; Edelbaum, T. N.
1974-01-01
A collection of computer programs and subroutines written in FORTRAN to calculate 4-body (sun-earth-moon-space) and 3-body (earth-moon-space) optimal trajectories is presented. The programs incorporate a variable step integration technique and a quadrature formula to correct single step errors. The programs provide capability to solve initial value problem, two point boundary value problem of a transfer from a given initial position to a given final position in fixed time, optimal 2-impulse transfer from an earth parking orbit of given inclination to a given final position and velocity in fixed time and optimal 3-impulse transfer from a given position to a given final position and velocity in fixed time.
NASA Technical Reports Server (NTRS)
Steffen, Chris
1990-01-01
An overview of the time-delay problem and the reliability problem which arise in trying to perform robotic construction operations at a remote space location are presented. The effects of the time-delay upon the control system design will be itemized. A high level overview of a decentralized method of control which is expected to perform better than the centralized approach in solving the time-delay problem is given. The lower level, decentralized, autonomous, Troter Move-Bar algorithm is also presented (Troters are coordinated independent robots). The solution of the reliability problem is connected to adding redundancy to the system. One method of adding redundancy is given.
On the complexity and approximability of some Euclidean optimal summing problems
NASA Astrophysics Data System (ADS)
Eremeev, A. V.; Kel'manov, A. V.; Pyatkin, A. V.
2016-10-01
The complexity status of several well-known discrete optimization problems with the direction of optimization switching from maximum to minimum is analyzed. The task is to find a subset of a finite set of Euclidean points (vectors). In these problems, the objective functions depend either only on the norm of the sum of the elements from the subset or on this norm and the cardinality of the subset. It is proved that, if the dimension of the space is a part of the input, then all these problems are strongly NP-hard. Additionally, it is shown that, if the space dimension is fixed, then all the problems are NP-hard even for dimension 2 (on a plane) and there are no approximation algorithms with a guaranteed accuracy bound for them unless P = NP. It is shown that, if the coordinates of the input points are integer, then all the problems can be solved in pseudopolynomial time in the case of a fixed space dimension.
Duggal, K L
2016-01-01
A new technique is used to study a family of time-dependent null horizons, called " Evolving Null Horizons " (ENHs), of generalized Robertson-Walker (GRW) space-time [Formula: see text] such that the metric [Formula: see text] satisfies a kinematic condition. This work is different from our early papers on the same issue where we used (1 + n )-splitting space-time but only some special subcases of GRW space-time have this formalism. Also, in contrast to previous work, we have proved that each member of ENHs is totally umbilical in [Formula: see text]. Finally, we show that there exists an ENH which is always a null horizon evolving into a black hole event horizon and suggest some open problems.
Closed-Loop Optimal Control Implementations for Space Applications
2016-12-01
analyses of a series of optimal control problems, several real- time optimal control algorithms are developed that continuously adapt to feedback on the...through the analyses of a series of optimal control problems, several real- time optimal control algorithms are developed that continuously adapt to...information is estimated to average 1 hour per response, including the time for reviewing instruction, searching existing data sources, gathering
NASA Astrophysics Data System (ADS)
Alemany, Kristina
Electric propulsion has recently become a viable technology for spacecraft, enabling shorter flight times, fewer required planetary gravity assists, larger payloads, and/or smaller launch vehicles. With the maturation of this technology, however, comes a new set of challenges in the area of trajectory design. Because low-thrust trajectory optimization has historically required long run-times and significant user-manipulation, mission design has relied on expert-based knowledge for selecting departure and arrival dates, times of flight, and/or target bodies and gravitational swing-bys. These choices are generally based on known configurations that have worked well in previous analyses or simply on trial and error. At the conceptual design level, however, the ability to explore the full extent of the design space is imperative to locating the best solutions in terms of mass and/or flight times. Beginning in 2005, the Global Trajectory Optimization Competition posed a series of difficult mission design problems, all requiring low-thrust propulsion and visiting one or more asteroids. These problems all had large ranges on the continuous variables---launch date, time of flight, and asteroid stay times (when applicable)---as well as being characterized by millions or even billions of possible asteroid sequences. Even with recent advances in low-thrust trajectory optimization, full enumeration of these problems was not possible within the stringent time limits of the competition. This investigation develops a systematic methodology for determining a broad suite of good solutions to the combinatorial, low-thrust, asteroid tour problem. The target application is for conceptual design, where broad exploration of the design space is critical, with the goal being to rapidly identify a reasonable number of promising solutions for future analysis. The proposed methodology has two steps. The first step applies a three-level heuristic sequence developed from the physics of the problem, which allows for efficient pruning of the design space. The second phase applies a global optimization scheme to locate a broad suite of good solutions to the reduced problem. The global optimization scheme developed combines a novel branch-and-bound algorithm with a genetic algorithm and an industry-standard low-thrust trajectory optimization program to solve for the following design variables: asteroid sequence, launch date, times of flight, and asteroid stay times. The methodology is developed based on a small sample problem, which is enumerated and solved so that all possible discretized solutions are known. The methodology is then validated by applying it to a larger intermediate sample problem, which also has a known solution. Next, the methodology is applied to several larger combinatorial asteroid rendezvous problems, using previously identified good solutions as validation benchmarks. These problems include the 2nd and 3rd Global Trajectory Optimization Competition problems. The methodology is shown to be capable of achieving a reduction in the number of asteroid sequences of 6-7 orders of magnitude, in terms of the number of sequences that require low-thrust optimization as compared to the number of sequences in the original problem. More than 70% of the previously known good solutions are identified, along with several new solutions that were not previously reported by any of the competitors. Overall, the methodology developed in this investigation provides an organized search technique for the low-thrust mission design of asteroid rendezvous problems.
A survey of methods of feasible directions for the solution of optimal control problems
NASA Technical Reports Server (NTRS)
Polak, E.
1972-01-01
Three methods of feasible directions for optimal control are reviewed. These methods are an extension of the Frank-Wolfe method, a dual method devised by Pironneau and Polack, and a Zontendijk method. The categories of continuous optimal control problems are shown as: (1) fixed time problems with fixed initial state, free terminal state, and simple constraints on the control; (2) fixed time problems with inequality constraints on both the initial and the terminal state and no control constraints; (3) free time problems with inequality constraints on the initial and terminal states and simple constraints on the control; and (4) fixed time problems with inequality state space contraints and constraints on the control. The nonlinear programming algorithms are derived for each of the methods in its associated category.
A POD reduced order model for resolving angular direction in neutron/photon transport problems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Buchan, A.G., E-mail: andrew.buchan@imperial.ac.uk; Calloo, A.A.; Goffin, M.G.
2015-09-01
This article presents the first Reduced Order Model (ROM) that efficiently resolves the angular dimension of the time independent, mono-energetic Boltzmann Transport Equation (BTE). It is based on Proper Orthogonal Decomposition (POD) and uses the method of snapshots to form optimal basis functions for resolving the direction of particle travel in neutron/photon transport problems. A unique element of this work is that the snapshots are formed from the vector of angular coefficients relating to a high resolution expansion of the BTE's angular dimension. In addition, the individual snapshots are not recorded through time, as in standard POD, but instead theymore » are recorded through space. In essence this work swaps the roles of the dimensions space and time in standard POD methods, with angle and space respectively. It is shown here how the POD model can be formed from the POD basis functions in a highly efficient manner. The model is then applied to two radiation problems; one involving the transport of radiation through a shield and the other through an infinite array of pins. Both problems are selected for their complex angular flux solutions in order to provide an appropriate demonstration of the model's capabilities. It is shown that the POD model can resolve these fluxes efficiently and accurately. In comparison to high resolution models this POD model can reduce the size of a problem by up to two orders of magnitude without compromising accuracy. Solving times are also reduced by similar factors.« less
Space-time adaptive solution of inverse problems with the discrete adjoint method
NASA Astrophysics Data System (ADS)
Alexe, Mihai; Sandu, Adrian
2014-08-01
This paper develops a framework for the construction and analysis of discrete adjoint sensitivities in the context of time dependent, adaptive grid, adaptive step models. Discrete adjoints are attractive in practice since they can be generated with low effort using automatic differentiation. However, this approach brings several important challenges. The space-time adjoint of the forward numerical scheme may be inconsistent with the continuous adjoint equations. A reduction in accuracy of the discrete adjoint sensitivities may appear due to the inter-grid transfer operators. Moreover, the optimization algorithm may need to accommodate state and gradient vectors whose dimensions change between iterations. This work shows that several of these potential issues can be avoided through a multi-level optimization strategy using discontinuous Galerkin (DG) hp-adaptive discretizations paired with Runge-Kutta (RK) time integration. We extend the concept of dual (adjoint) consistency to space-time RK-DG discretizations, which are then shown to be well suited for the adaptive solution of time-dependent inverse problems. Furthermore, we prove that DG mesh transfer operators on general meshes are also dual consistent. This allows the simultaneous derivation of the discrete adjoint for both the numerical solver and the mesh transfer logic with an automatic code generation mechanism such as algorithmic differentiation (AD), potentially speeding up development of large-scale simulation codes. The theoretical analysis is supported by numerical results reported for a two-dimensional non-stationary inverse problem.
Three Years of Global Positioning System Experience on International Space Station
NASA Technical Reports Server (NTRS)
Gomez, Susan
2005-01-01
The International Space Station global positioning systems (GPS) receiver was activated in April 2002. Since that time, numerous software anomalies surfaced that had to be worked around. Some of the software problems required waivers, such as the time function, while others required extensive operator intervention, such as numerous power cycles. Eventually, enough anomalies surfaced that the three pieces of code included in the GPS unit have been re-written and the GPS units were upgraded. The technical aspects of the problems are discussed, as well as the underlying causes that led to the delivery of a product that has had numerous problems. The technical aspects of the problems included physical phenomena that were not well understood, such as the affect that the ionosphere would have on the GPS measurements. The underlying causes were traced to inappropriate use of legacy software, changing requirements, inadequate software processes, unrealistic schedules, incorrect contract type, and unclear ownership responsibilities.
Three Years of Global Positioning System Experience on International Space Station
NASA Technical Reports Server (NTRS)
Gomez, Susan
2006-01-01
The International Space Station global positioning system (GPS) receiver was activated in April 2002. Since that time, numerous software anomalies surfaced that had to be worked around. Some of the software problems required waivers, such as the time function, while others required extensive operator intervention, such as numerous power cycles. Eventually enough anomalies surfaced that the three pieces of code included in the GPS unit have been re-written and the GPS units upgraded. The technical aspects of the problems are discussed, as well as the underlying causes that led to the delivery of a product that has had so many problems. The technical aspects of the problems included physical phenomena that were not well understood, such as the affect that the ionosphere would have on the GPS measurements. The underlying causes were traced to inappropriate use of legacy software, changing requirements, inadequate software processes, unrealistic schedules, incorrect contract type, and unclear ownership responsibilities..
Integrated Thermal Response Modeling System For Hypersonic Entry Vehicles
NASA Technical Reports Server (NTRS)
Chen, Y.-K.; Milos, F. S.; Partridge, Harry (Technical Monitor)
2000-01-01
We describe all extension of the Markov decision process model in which a continuous time dimension is included ill the state space. This allows for the representation and exact solution of a wide range of problems in which transitions or rewards vary over time. We examine problems based on route planning with public transportation and telescope observation scheduling.
NASA Technical Reports Server (NTRS)
Wang, Xiao Yen; Chang, Sin-Chung; Jorgenson, Philip C. E.
1999-01-01
The space-time conservation element and solution element(CE/SE) method is used to study the sound-shock interaction problem. The order of accuracy of numerical schemes is investigated. The linear model problem.govemed by the 1-D scalar convection equation, sound-shock interaction problem governed by the 1-D Euler equations, and the 1-D shock-tube problem which involves moving shock waves and contact surfaces are solved to investigate the order of accuracy of numerical schemes. It is concluded that the accuracy of the CE/SE numerical scheme with designed 2nd-order accuracy becomes 1st order when a moving shock wave exists. However, the absolute error in the CE/SE solution downstream of the shock wave is on the same order as that obtained using a fourth-order accurate essentially nonoscillatory (ENO) scheme. No special techniques are used for either high-frequency low-amplitude waves or shock waves.
A k-Space Method for Moderately Nonlinear Wave Propagation
Jing, Yun; Wang, Tianren; Clement, Greg T.
2013-01-01
A k-space method for moderately nonlinear wave propagation in absorptive media is presented. The Westervelt equation is first transferred into k-space via Fourier transformation, and is solved by a modified wave-vector time-domain scheme. The present approach is not limited to forward propagation or parabolic approximation. One- and two-dimensional problems are investigated to verify the method by comparing results to analytic solutions and finite-difference time-domain (FDTD) method. It is found that to obtain accurate results in homogeneous media, the grid size can be as little as two points per wavelength, and for a moderately nonlinear problem, the Courant–Friedrichs–Lewy number can be as large as 0.4. Through comparisons with the conventional FDTD method, the k-space method for nonlinear wave propagation is shown here to be computationally more efficient and accurate. The k-space method is then employed to study three-dimensional nonlinear wave propagation through the skull, which shows that a relatively accurate focusing can be achieved in the brain at a high frequency by sending a low frequency from the transducer. Finally, implementations of the k-space method using a single graphics processing unit shows that it required about one-seventh the computation time of a single-core CPU calculation. PMID:22899114
NASA Technical Reports Server (NTRS)
Banks, H. T.; Silcox, R. J.; Keeling, S. L.; Wang, C.
1989-01-01
A unified treatment of the linear quadratic tracking (LQT) problem, in which a control system's dynamics are modeled by a linear evolution equation with a nonhomogeneous component that is linearly dependent on the control function u, is presented; the treatment proceeds from the theoretical formulation to a numerical approximation framework. Attention is given to two categories of LQT problems in an infinite time interval: the finite energy and the finite average energy. The behavior of the optimal solution for finite time-interval problems as the length of the interval tends to infinity is discussed. Also presented are the formulations and properties of LQT problems in a finite time interval.
Learning Search Control Knowledge for Deep Space Network Scheduling
NASA Technical Reports Server (NTRS)
Gratch, Jonathan; Chien, Steve; DeJong, Gerald
1993-01-01
While the general class of most scheduling problems is NP-hard in worst-case complexity, in practice, for specific distributions of problems and constraints, domain-specific solutions have been shown to perform in much better than exponential time.
Fast metabolite identification with Input Output Kernel Regression.
Brouard, Céline; Shen, Huibin; Dührkop, Kai; d'Alché-Buc, Florence; Böcker, Sebastian; Rousu, Juho
2016-06-15
An important problematic of metabolomics is to identify metabolites using tandem mass spectrometry data. Machine learning methods have been proposed recently to solve this problem by predicting molecular fingerprint vectors and matching these fingerprints against existing molecular structure databases. In this work we propose to address the metabolite identification problem using a structured output prediction approach. This type of approach is not limited to vector output space and can handle structured output space such as the molecule space. We use the Input Output Kernel Regression method to learn the mapping between tandem mass spectra and molecular structures. The principle of this method is to encode the similarities in the input (spectra) space and the similarities in the output (molecule) space using two kernel functions. This method approximates the spectra-molecule mapping in two phases. The first phase corresponds to a regression problem from the input space to the feature space associated to the output kernel. The second phase is a preimage problem, consisting in mapping back the predicted output feature vectors to the molecule space. We show that our approach achieves state-of-the-art accuracy in metabolite identification. Moreover, our method has the advantage of decreasing the running times for the training step and the test step by several orders of magnitude over the preceding methods. celine.brouard@aalto.fi Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press.
Fast metabolite identification with Input Output Kernel Regression
Brouard, Céline; Shen, Huibin; Dührkop, Kai; d'Alché-Buc, Florence; Böcker, Sebastian; Rousu, Juho
2016-01-01
Motivation: An important problematic of metabolomics is to identify metabolites using tandem mass spectrometry data. Machine learning methods have been proposed recently to solve this problem by predicting molecular fingerprint vectors and matching these fingerprints against existing molecular structure databases. In this work we propose to address the metabolite identification problem using a structured output prediction approach. This type of approach is not limited to vector output space and can handle structured output space such as the molecule space. Results: We use the Input Output Kernel Regression method to learn the mapping between tandem mass spectra and molecular structures. The principle of this method is to encode the similarities in the input (spectra) space and the similarities in the output (molecule) space using two kernel functions. This method approximates the spectra-molecule mapping in two phases. The first phase corresponds to a regression problem from the input space to the feature space associated to the output kernel. The second phase is a preimage problem, consisting in mapping back the predicted output feature vectors to the molecule space. We show that our approach achieves state-of-the-art accuracy in metabolite identification. Moreover, our method has the advantage of decreasing the running times for the training step and the test step by several orders of magnitude over the preceding methods. Availability and implementation: Contact: celine.brouard@aalto.fi Supplementary information: Supplementary data are available at Bioinformatics online. PMID:27307628
A Note on the Problem of Proper Time in Weyl Space-Time
NASA Astrophysics Data System (ADS)
Avalos, R.; Dahia, F.; Romero, C.
2018-02-01
We discuss the question of whether or not a general Weyl structure is a suitable mathematical model of space-time. This is an issue that has been in debate since Weyl formulated his unified field theory for the first time. We do not present the discussion from the point of view of a particular unification theory, but instead from a more general standpoint, in which the viability of such a structure as a model of space-time is investigated. Our starting point is the well known axiomatic approach to space-time given by Elhers, Pirani and Schild (EPS). In this framework, we carry out an exhaustive analysis of what is required for a consistent definition for proper time and show that such a definition leads to the prediction of the so-called "second clock effect". We take the view that if, based on experience, we were to reject space-time models predicting this effect, this could be incorporated as the last axiom in the EPS approach. Finally, we provide a proof that, in this case, we are led to a Weyl integrable space-time as the most general structure that would be suitable to model space-time.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tavakoli, Rouhollah, E-mail: rtavakoli@sharif.ir
An unconditionally energy stable time stepping scheme is introduced to solve Cahn–Morral-like equations in the present study. It is constructed based on the combination of David Eyre's time stepping scheme and Schur complement approach. Although the presented method is general and independent of the choice of homogeneous free energy density function term, logarithmic and polynomial energy functions are specifically considered in this paper. The method is applied to study the spinodal decomposition in multi-component systems and optimal space tiling problems. A penalization strategy is developed, in the case of later problem, to avoid trivial solutions. Extensive numerical experiments demonstrate themore » success and performance of the presented method. According to the numerical results, the method is convergent and energy stable, independent of the choice of time stepsize. Its MATLAB implementation is included in the appendix for the numerical evaluation of algorithm and reproduction of the presented results. -- Highlights: •Extension of Eyre's convex–concave splitting scheme to multiphase systems. •Efficient solution of spinodal decomposition in multi-component systems. •Efficient solution of least perimeter periodic space partitioning problem. •Developing a penalization strategy to avoid trivial solutions. •Presentation of MATLAB implementation of the introduced algorithm.« less
Modular thermal analyzer routine, volume 1
NASA Technical Reports Server (NTRS)
Oren, J. A.; Phillips, M. A.; Williams, D. R.
1972-01-01
The Modular Thermal Analyzer Routine (MOTAR) is a general thermal analysis routine with strong capabilities for performing thermal analysis of systems containing flowing fluids, fluid system controls (valves, heat exchangers, etc.), life support systems, and thermal radiation situations. Its modular organization permits the analysis of a very wide range of thermal problems for simple problems containing a few conduction nodes to those containing complicated flow and radiation analysis with each problem type being analyzed with peak computational efficiency and maximum ease of use. The organization and programming methods applied to MOTAR achieved a high degree of computer utilization efficiency in terms of computer execution time and storage space required for a given problem. The computer time required to perform a given problem on MOTAR is approximately 40 to 50 percent that required for the currently existing widely used routines. The computer storage requirement for MOTAR is approximately 25 percent more than the most commonly used routines for the most simple problems but the data storage techniques for the more complicated options should save a considerable amount of space.
2016-01-01
A new technique is used to study a family of time-dependent null horizons, called “Evolving Null Horizons” (ENHs), of generalized Robertson-Walker (GRW) space-time (M¯,g¯) such that the metric g¯ satisfies a kinematic condition. This work is different from our early papers on the same issue where we used (1 + n)-splitting space-time but only some special subcases of GRW space-time have this formalism. Also, in contrast to previous work, we have proved that each member of ENHs is totally umbilical in (M¯,g¯). Finally, we show that there exists an ENH which is always a null horizon evolving into a black hole event horizon and suggest some open problems. PMID:27722202
Classical dynamics on curved Snyder space
NASA Astrophysics Data System (ADS)
Ivetić, B.; Meljanac, S.; Mignemi, S.
2014-05-01
We study the classical dynamics of a particle in nonrelativistic Snyder-de Sitter space. We show that for spherically symmetric systems, parameterizing the solutions in terms of an auxiliary time variable, which is a function only of the physical time and of the energy and angular momentum of the particles, one can reduce the problem to the equivalent one in classical mechanics. We also discuss a relativistic extension of these results, and a generalization to the case in which the algebra is realized in flat space.
Gust Acoustics Computation with a Space-Time CE/SE Parallel 3D Solver
NASA Technical Reports Server (NTRS)
Wang, X. Y.; Himansu, A.; Chang, S. C.; Jorgenson, P. C. E.; Reddy, D. R. (Technical Monitor)
2002-01-01
The benchmark Problem 2 in Category 3 of the Third Computational Aero-Acoustics (CAA) Workshop is solved using the space-time conservation element and solution element (CE/SE) method. This problem concerns the unsteady response of an isolated finite-span swept flat-plate airfoil bounded by two parallel walls to an incident gust. The acoustic field generated by the interaction of the gust with the flat-plate airfoil is computed by solving the 3D (three-dimensional) Euler equations in the time domain using a parallel version of a 3D CE/SE solver. The effect of the gust orientation on the far-field directivity is studied. Numerical solutions are presented and compared with analytical solutions, showing a reasonable agreement.
Inversion of geophysical potential field data using the finite element method
NASA Astrophysics Data System (ADS)
Lamichhane, Bishnu P.; Gross, Lutz
2017-12-01
The inversion of geophysical potential field data can be formulated as an optimization problem with a constraint in the form of a partial differential equation (PDE). It is common practice, if possible, to provide an analytical solution for the forward problem and to reduce the problem to a finite dimensional optimization problem. In an alternative approach the optimization is applied to the problem and the resulting continuous problem which is defined by a set of coupled PDEs is subsequently solved using a standard PDE discretization method, such as the finite element method (FEM). In this paper, we show that under very mild conditions on the data misfit functional and the forward problem in the three-dimensional space, the continuous optimization problem and its FEM discretization are well-posed including the existence and uniqueness of respective solutions. We provide error estimates for the FEM solution. A main result of the paper is that the FEM spaces used for the forward problem and the Lagrange multiplier need to be identical but can be chosen independently from the FEM space used to represent the unknown physical property. We will demonstrate the convergence of the solution approximations in a numerical example. The second numerical example which investigates the selection of FEM spaces, shows that from the perspective of computational efficiency one should use 2 to 4 times finer mesh for the forward problem in comparison to the mesh of the physical property.
Problems of sampling and radiation balances: Their problematics
NASA Technical Reports Server (NTRS)
Crommelynck, D.
1980-01-01
Problems associated with the measurement of the Earth radiation balances are addressed. It is demonstrated that the knowledge of the different radiation budgets with their components is largely dependent on the space time sampling of the radiation field of the Earth atmosphere system. Whichever instrumental approach is adopted (wide angle view of high resolution) it affects the space time integration of the fluxes measured directly or calculated. In this case the necessary knowledge of the reflection pattern depends in addition on the angular sampling of the radiances. A series of questions is considered, the answers of which are a prerequisite to the the organization of a global observation system.
Accommodation of practical constraints by a linear programming jet select. [for Space Shuttle
NASA Technical Reports Server (NTRS)
Bergmann, E.; Weiler, P.
1983-01-01
An experimental spacecraft control system will be incorporated into the Space Shuttle flight software and exercised during a forthcoming mission to evaluate its performance and handling qualities. The control system incorporates a 'phase space' control law to generate rate change requests and a linear programming jet select to compute jet firings. Posed as a linear programming problem, jet selection must represent the rate change request as a linear combination of jet acceleration vectors where the coefficients are the jet firing times, while minimizing the fuel expended in satisfying that request. This problem is solved in real time using a revised Simplex algorithm. In order to implement the jet selection algorithm in the Shuttle flight control computer, it was modified to accommodate certain practical features of the Shuttle such as limited computer throughput, lengthy firing times, and a large number of control jets. To the authors' knowledge, this is the first such application of linear programming. It was made possible by careful consideration of the jet selection problem in terms of the properties of linear programming and the Simplex algorithm. These modifications to the jet select algorithm may by useful for the design of reaction controlled spacecraft.
Linear and nonlinear dynamic analysis by boundary element method. Ph.D. Thesis, 1986 Final Report
NASA Technical Reports Server (NTRS)
Ahmad, Shahid
1991-01-01
An advanced implementation of the direct boundary element method (BEM) applicable to free-vibration, periodic (steady-state) vibration and linear and nonlinear transient dynamic problems involving two and three-dimensional isotropic solids of arbitrary shape is presented. Interior, exterior, and half-space problems can all be solved by the present formulation. For the free-vibration analysis, a new real variable BEM formulation is presented which solves the free-vibration problem in the form of algebraic equations (formed from the static kernels) and needs only surface discretization. In the area of time-domain transient analysis, the BEM is well suited because it gives an implicit formulation. Although the integral formulations are elegant, because of the complexity of the formulation it has never been implemented in exact form. In the present work, linear and nonlinear time domain transient analysis for three-dimensional solids has been implemented in a general and complete manner. The formulation and implementation of the nonlinear, transient, dynamic analysis presented here is the first ever in the field of boundary element analysis. Almost all the existing formulation of BEM in dynamics use the constant variation of the variables in space and time which is very unrealistic for engineering problems and, in some cases, it leads to unacceptably inaccurate results. In the present work, linear and quadratic isoparametric boundary elements are used for discretization of geometry and functional variations in space. In addition, higher order variations in time are used. These methods of analysis are applicable to piecewise-homogeneous materials, such that not only problems of the layered media and the soil-structure interaction can be analyzed but also a large problem can be solved by the usual sub-structuring technique. The analyses have been incorporated in a versatile, general-purpose computer program. Some numerical problems are solved and, through comparisons with available analytical and numerical results, the stability and high accuracy of these dynamic analysis techniques are established.
Space and time in the quantum universe.
NASA Astrophysics Data System (ADS)
Smolin, L.
This paper is devoted to the problem of constructing a quantum theory that could describe a closed system - a quantum cosmology. The author argues that this problem is an aspect of a much older problem - that of how to eliminate from the physical theories "ideal elements", which are elements of the mathematical structure whose interpretation requires the existence of things outside the dynamical system described by the theory. This discussion is aimed at uncovering criteria that a theory of quantum cosmology must satisfy, if it is to give physically sensible predictions. The author proposes three such criteria and shows that conventional quantum cosmology can only satisfy them, if there is an intrinsic time coordinate on the phase space of the theory. It is shown that approaches based on correlations in the wave function, that do not use an inner product, cannot satisfy these criteria. As example, the author discusses the problem of quantizing a class of relational dynamical models invented by Barbour and Bertotti. The dynamical structure of these theories is closely analogous to general relativity, and the problem of their measurement theory is also similar. It is concluded that these theories can only be sensibly quantized if they contain an intrinsic time.
Nonuniform dependence on initial data for compressible gas dynamics: The periodic Cauchy problem
NASA Astrophysics Data System (ADS)
Keyfitz, B. L.; Tığlay, F.
2017-11-01
We start with the classic result that the Cauchy problem for ideal compressible gas dynamics is locally well posed in time in the sense of Hadamard; there is a unique solution that depends continuously on initial data in Sobolev space Hs for s > d / 2 + 1 where d is the space dimension. We prove that the data to solution map for periodic data in two dimensions although continuous is not uniformly continuous.
Space-dependent perfusion coefficient estimation in a 2D bioheat transfer problem
NASA Astrophysics Data System (ADS)
Bazán, Fermín S. V.; Bedin, Luciano; Borges, Leonardo S.
2017-05-01
In this work, a method for estimating the space-dependent perfusion coefficient parameter in a 2D bioheat transfer model is presented. In the method, the bioheat transfer model is transformed into a time-dependent semidiscrete system of ordinary differential equations involving perfusion coefficient values as parameters, and the estimation problem is solved through a nonlinear least squares technique. In particular, the bioheat problem is solved by the method of lines based on a highly accurate pseudospectral approach, and perfusion coefficient values are estimated by the regularized Gauss-Newton method coupled with a proper regularization parameter. The performance of the method on several test problems is illustrated numerically.
Towards conformal loop quantum gravity
NASA Astrophysics Data System (ADS)
H-T Wang, Charles
2006-03-01
A discussion is given of recent developments in canonical gravity that assimilates the conformal analysis of gravitational degrees of freedom. The work is motivated by the problem of time in quantum gravity and is carried out at the metric and the triad levels. At the metric level, it is shown that by extending the Arnowitt-Deser-Misner (ADM) phase space of general relativity (GR), a conformal form of geometrodynamics can be constructed. In addition to the Hamiltonian and Diffeomorphism constraints, an extra first class constraint is introduced to generate conformal transformations. This phase space consists of York's mean extrinsic curvature time, conformal three-metric and their momenta. At the triad level, the phase space of GR is further enlarged by incorporating spin-gauge as well as conformal symmetries. This leads to a canonical formulation of GR using a new set of real spin connection variables. The resulting gravitational constraints are first class, consisting of the Hamiltonian constraint and the canonical generators for spin-gauge and conformorphism transformations. The formulation has a remarkable feature of being parameter-free. Indeed, it is shown that a conformal parameter of the Barbero-Immirzi type can be absorbed by the conformal symmetry of the extended phase space. This gives rise to an alternative approach to loop quantum gravity that addresses both the conceptual problem of time and the technical problem of functional calculus in quantum gravity.
Zhang, Jisheng; Jia, Limin; Niu, Shuyun; Zhang, Fan; Tong, Lu; Zhou, Xuesong
2015-01-01
It is essential for transportation management centers to equip and manage a network of fixed and mobile sensors in order to quickly detect traffic incidents and further monitor the related impact areas, especially for high-impact accidents with dramatic traffic congestion propagation. As emerging small Unmanned Aerial Vehicles (UAVs) start to have a more flexible regulation environment, it is critically important to fully explore the potential for of using UAVs for monitoring recurring and non-recurring traffic conditions and special events on transportation networks. This paper presents a space-time network- based modeling framework for integrated fixed and mobile sensor networks, in order to provide a rapid and systematic road traffic monitoring mechanism. By constructing a discretized space-time network to characterize not only the speed for UAVs but also the time-sensitive impact areas of traffic congestion, we formulate the problem as a linear integer programming model to minimize the detection delay cost and operational cost, subject to feasible flying route constraints. A Lagrangian relaxation solution framework is developed to decompose the original complex problem into a series of computationally efficient time-dependent and least cost path finding sub-problems. Several examples are used to demonstrate the results of proposed models in UAVs’ route planning for small and medium-scale networks. PMID:26076404
A random utility based estimation framework for the household activity pattern problem.
DOT National Transportation Integrated Search
2016-06-01
This paper develops a random utility based estimation framework for the Household Activity : Pattern Problem (HAPP). Based on the realization that output of complex activity-travel decisions : form a continuous pattern in space-time dimension, the es...
An object correlation and maneuver detection approach for space surveillance
NASA Astrophysics Data System (ADS)
Huang, Jian; Hu, Wei-Dong; Xin, Qin; Du, Xiao-Yong
2012-10-01
Object correlation and maneuver detection are persistent problems in space surveillance and maintenance of a space object catalog. We integrate these two problems into one interrelated problem, and consider them simultaneously under a scenario where space objects only perform a single in-track orbital maneuver during the time intervals between observations. We mathematically formulate this integrated scenario as a maximum a posteriori (MAP) estimation. In this work, we propose a novel approach to solve the MAP estimation. More precisely, the corresponding posterior probability of an orbital maneuver and a joint association event can be approximated by the Joint Probabilistic Data Association (JPDA) algorithm. Subsequently, the maneuvering parameters are estimated by optimally solving the constrained non-linear least squares iterative process based on the second-order cone programming (SOCP) algorithm. The desired solution is derived according to the MAP criterions. The performance and advantages of the proposed approach have been shown by both theoretical analysis and simulation results. We hope that our work will stimulate future work on space surveillance and maintenance of a space object catalog.
General Relativity without paradigm of space-time covariance, and resolution of the problem of time
NASA Astrophysics Data System (ADS)
Soo, Chopin; Yu, Hoi-Lai
2014-01-01
The framework of a theory of gravity from the quantum to the classical regime is presented. The paradigm shift from full space-time covariance to spatial diffeomorphism invariance, together with clean decomposition of the canonical structure, yield transparent physical dynamics and a resolution of the problem of time. The deep divide between quantum mechanics and conventional canonical formulations of quantum gravity is overcome with a Schrödinger equation for quantum geometrodynamics that describes evolution in intrinsic time. Unitary time development with gauge-invariant temporal ordering is also viable. All Kuchar observables become physical; and classical space-time, with direct correlation between its proper times and intrinsic time intervals, emerges from constructive interference. The framework not only yields a physical Hamiltonian for Einstein's theory, but also prompts natural extensions and improvements towards a well behaved quantum theory of gravity. It is a consistent canonical scheme to discuss Horava-Lifshitz theories with intrinsic time evolution, and of the many possible alternatives that respect 3-covariance (rather than the more restrictive 4-covariance of Einstein's theory), Horava's "detailed balance" form of the Hamiltonian constraint is essentially pinned down by this framework. Issues in quantum gravity that depend on radiative corrections and the rigorous definition and regularization of the Hamiltonian operator are not addressed in this work.
Some Pitfalls in Special Relativity
ERIC Educational Resources Information Center
Chai, An-Ti
1973-01-01
The problem of observing a collision between two spaceships is utilized to illustrate the misuse of time dilation and length contraction, and the incorrect description of space-time coordinates in special relativity. (Author/DF)
Borgia, G C; Brown, R J; Fantazzini, P
2000-12-01
The basic method of UPEN (uniform penalty inversion of multiexponential decay data) is given in an earlier publication (Borgia et al., J. Magn. Reson. 132, 65-77 (1998)), which also discusses the effects of noise, constraints, and smoothing on the resolution or apparent resolution of features of a computed distribution of relaxation times. UPEN applies negative feedback to a regularization penalty, allowing stronger smoothing for a broad feature than for a sharp line. This avoids unnecessarily broadening the sharp line and/or breaking the wide peak or tail into several peaks that the relaxation data do not demand to be separate. The experimental and artificial data presented earlier were T(1) data, and all had fixed data spacings, uniform in log-time. However, for T(2) data, usually spaced uniformly in linear time, or for data spaced in any manner, we have found that the data spacing does not enter explicitly into the computation. The present work shows the extension of UPEN to T(2) data, including the averaging of data in windows and the use of the corresponding weighting factors in the computation. Measures are implemented to control portions of computed distributions extending beyond the data range. The input smoothing parameters in UPEN are normally fixed, rather than data dependent. A major problem arises, especially at high signal-to-noise ratios, when UPEN is applied to data sets with systematic errors due to instrumental nonidealities or adjustment problems. For instance, a relaxation curve for a wide line can be narrowed by an artificial downward bending of the relaxation curve. Diagnostic parameters are generated to help identify data problems, and the diagnostics are applied in several examples, with particular attention to the meaningful resolution of two closely spaced peaks in a distribution of relaxation times. Where feasible, processing with UPEN in nearly real time should help identify data problems while further instrument adjustments can still be made. The need for the nonnegative constraint is greatly reduced in UPEN, and preliminary processing without this constraint helps identify data sets for which application of the nonnegative constraint is too expensive in terms of error of fit for the data set to represent sums of decaying positive exponentials plus random noise. Copyright 2000 Academic Press.
Space-time-modulated stochastic processes
NASA Astrophysics Data System (ADS)
Giona, Massimiliano
2017-10-01
Starting from the physical problem associated with the Lorentzian transformation of a Poisson-Kac process in inertial frames, the concept of space-time-modulated stochastic processes is introduced for processes possessing finite propagation velocity. This class of stochastic processes provides a two-way coupling between the stochastic perturbation acting on a physical observable and the evolution of the physical observable itself, which in turn influences the statistical properties of the stochastic perturbation during its evolution. The definition of space-time-modulated processes requires the introduction of two functions: a nonlinear amplitude modulation, controlling the intensity of the stochastic perturbation, and a time-horizon function, which modulates its statistical properties, providing irreducible feedback between the stochastic perturbation and the physical observable influenced by it. The latter property is the peculiar fingerprint of this class of models that makes them suitable for extension to generic curved-space times. Considering Poisson-Kac processes as prototypical examples of stochastic processes possessing finite propagation velocity, the balance equations for the probability density functions associated with their space-time modulations are derived. Several examples highlighting the peculiarities of space-time-modulated processes are thoroughly analyzed.
AI techniques for a space application scheduling problem
NASA Technical Reports Server (NTRS)
Thalman, N.; Sparn, T.; Jaffres, L.; Gablehouse, D.; Judd, D.; Russell, C.
1991-01-01
Scheduling is a very complex optimization problem which can be categorized as an NP-complete problem. NP-complete problems are quite diverse, as are the algorithms used in searching for an optimal solution. In most cases, the best solutions that can be derived for these combinatorial explosive problems are near-optimal solutions. Due to the complexity of the scheduling problem, artificial intelligence (AI) can aid in solving these types of problems. Some of the factors are examined which make space application scheduling problems difficult and presents a fairly new AI-based technique called tabu search as applied to a real scheduling application. the specific problem is concerned with scheduling application. The specific problem is concerned with scheduling solar and stellar observations for the SOLar-STellar Irradiance Comparison Experiment (SOLSTICE) instrument in a constrained environment which produces minimum impact on the other instruments and maximizes target observation times. The SOLSTICE instrument will gly on-board the Upper Atmosphere Research Satellite (UARS) in 1991, and a similar instrument will fly on the earth observing system (Eos).
Space development and space science together, an historic opportunity
NASA Astrophysics Data System (ADS)
Metzger, P. T.
2016-11-01
The national space programs have an historic opportunity to help solve the global-scale economic and environmental problems of Earth while becoming more effective at science through the use of space resources. Space programs will be more cost-effective when they work to establish a supply chain in space, mining and manufacturing then replicating the assets of the supply chain so it grows to larger capacity. This has become achievable because of advances in robotics and artificial intelligence. It is roughly estimated that developing a lunar outpost that relies upon and also develops the supply chain will cost about 1/3 or less of the existing annual budgets of the national space programs. It will require a sustained commitment of several decades to complete, during which time science and exploration become increasingly effective. At the end, this space industry will capable of addressing global-scale challenges including limited resources, clean energy, economic development, and preservation of the environment. Other potential solutions, including nuclear fusion and terrestrial renewable energy sources, do not address the root problem of our limited globe and there are real questions whether they will be inadequate or too late. While industry in space likewise cannot provide perfect assurance, it is uniquely able to solve the root problem, and it gives us an important chance that we should grasp. What makes this such an historic opportunity is that the space-based solution is obtainable as a side-benefit of doing space science and exploration within their existing budgets. Thinking pragmatically, it may take some time for policymakers to agree that setting up a complete supply chain is an achievable goal, so this paper describes a strategy of incremental progress. The most crucial part of this strategy is establishing a water economy by mining on the Moon and asteroids to manufacture rocket propellant. Technologies that support a water economy will play an important role leading toward space development.
NASA Astrophysics Data System (ADS)
Prechtel, Alexander; Ray, Nadja; Rupp, Andreas
2017-04-01
We want to present an approach for the mathematical, mechanistic modeling and numerical treatment of processes leading to the formation, stability, and turnover of soil micro-aggregates. This aims at deterministic aggregation models including detailed mechanistic pore-scale descriptions to account for the interplay of geochemistry and microbiology, and the link to soil functions as, e.g., the porosity. We therefore consider processes at the pore scale and the mesoscale (laboratory scale). At the pore scale transport by diffusion, advection, and drift emerging from electric forces can be taken into account, in addition to homogeneous and heterogeneous reactions of species. In the context of soil micro-aggregates the growth of biofilms or other glueing substances as EPS (extracellular polymeric substances) is important and affects the structure of the pore space in space and time. This model is upscaled mathematically in the framework of (periodic) homogenization to transfer it to the mesoscale resulting in effective coefficients/parameters there. This micro-macro model thus couples macroscopic equations that describe the transport and fluid flow at the scale of the porous medium (mesoscale) with averaged time- and space-dependent coefficient functions. These functions may be explicitly computed by means of auxiliary cell problems (microscale). Finally, the pore space in which the cell problems are defined is time and space dependent and its geometry inherits information from the transport equation's solutions. The microscale problems rely on versatile combinations of cellular automata and discontiuous Galerkin methods while on the mesoscale mixed finite elements are used. The numerical simulations allow to study the interplay between these processes.
NASA Astrophysics Data System (ADS)
Figueiredo, Danilo Zucolli; Costa, Oswaldo Luiz do Valle
2017-10-01
This paper deals with the H2 optimal control problem of discrete-time Markov jump linear systems (MJLS) considering the case in which the Markov chain takes values in a general Borel space ?. It is assumed that the controller has access only to an output variable and to the jump parameter. The goal, in this case, is to design a dynamic Markov jump controller such that the H2-norm of the closed-loop system is minimised. It is shown that the H2-norm can be written as the sum of two H2-norms, such that one of them does not depend on the control, and the other one is obtained from the optimal filter for an infinite-horizon filtering problem. This result can be seen as a separation principle for MJLS with Markov chain in a Borel space ? considering the infinite time horizon case.
Helmholtz, Riemann, and the Sirens: Sound, Color, and the "Problem of Space"
NASA Astrophysics Data System (ADS)
Pesic, Peter
2013-09-01
Emerging from music and the visual arts, questions about hearing and seeing deeply affected Hermann Helmholtz's and Bernhard Riemann's contributions to what became called the "problem of space [ Raumproblem]," which in turn influenced Albert Einstein's approach to general relativity. Helmholtz's physiological investigations measured the time dependence of nerve conduction and mapped the three-dimensional manifold of color sensation. His concurrent studies on hearing illuminated musical evidence through experiments with mechanical sirens that connect audible with visible phenomena, especially how the concept of frequency unifies motion, velocity, and pitch. Riemann's critique of Helmholtz's work on hearing led Helmholtz to respond and study Riemann's then-unpublished lecture on the foundations of geometry. During 1862-1870, Helmholtz applied his findings on the manifolds of hearing and seeing to the Raumproblem by supporting the quadratic distance relation Riemann had assumed as his fundamental hypothesis about geometrical space. Helmholtz also drew a "close analogy … in all essential relations between the musical scale and space." These intersecting studies of hearing and seeing thus led to reconsideration and generalization of the very concept of "space," which Einstein shaped into the general manifold of relativistic space-time.
Space radiation research in the new millenium--from where we come and where we go.
Kiefer, J
2001-01-01
Space radiation research had a significant impact in the past. The physical interaction of heavy charged particles with living matter and the development of models, including microdosimetry, were stimulated by problems encountered in space. New phenomena were discovered. Advanced dosimetric techniques had to be developed and computational methods to describe the radiation field in space. The understanding of the radiobiology of heavy ions, necessary for a well-founded risk assessment and prompted by space radiation research, constitutes also the basis for heavy ion radiotherapy. So far unknown areas like the interaction of microgravity and radiation were opened. The space station will give even more opportunities. For the first time it will be possible to investigate animals for a longer time under the influence of both microgravity and radiation. Living systems can be exposed under well defined conditions with parallel physical measurements. Solar particle events are still an unsolved problem. Significant improvement in their predictability and quantitative description can be expected. All this will not only give exciting opportunities for research but will also translate into immediate benefit for human beings. This paper will attempt to give an overview of the past achievements and glance into the future.
Approximate solution of space and time fractional higher order phase field equation
NASA Astrophysics Data System (ADS)
Shamseldeen, S.
2018-03-01
This paper is concerned with a class of space and time fractional partial differential equation (STFDE) with Riesz derivative in space and Caputo in time. The proposed STFDE is considered as a generalization of a sixth-order partial phase field equation. We describe the application of the optimal homotopy analysis method (OHAM) to obtain an approximate solution for the suggested fractional initial value problem. An averaged-squared residual error function is defined and used to determine the optimal convergence control parameter. Two numerical examples are studied, considering periodic and non-periodic initial conditions, to justify the efficiency and the accuracy of the adopted iterative approach. The dependence of the solution on the order of the fractional derivative in space and time and model parameters is investigated.
NASA Astrophysics Data System (ADS)
Alpers, Andreas; Gritzmann, Peter
2018-03-01
We consider the problem of reconstructing the paths of a set of points over time, where, at each of a finite set of moments in time the current positions of points in space are only accessible through some small number of their x-rays. This particular particle tracking problem, with applications, e.g. in plasma physics, is the basic problem in dynamic discrete tomography. We introduce and analyze various different algorithmic models. In particular, we determine the computational complexity of the problem (and various of its relatives) and derive algorithms that can be used in practice. As a byproduct we provide new results on constrained variants of min-cost flow and matching problems.
2007-12-09
KENNEDY SPACE CENTER, FLA. -- Bill Gerstenmaier, associate administrator for Space Operations, participates in a news briefing following the conclusion of a Mission Management Team, or MMT, meeting. The meeting followed the morning's launch scrub of the space shuttle Atlantis STS-122 mission caused by problems experienced with the external tank's engine cutoff sensor system during tanking for the second launch attempt. An announcement was made during the briefing that the STS-122 launch is postponed to no earlier than Jan. 2, 2008, to give the team time to resolve the system's problems. Atlantis will carry the Columbus Laboratory, the European Space Agency's largest contribution to the construction of the International Space Station. It will support scientific and technological research in a microgravity environment. Permanently attached to the Harmony node of the space station, the laboratory will carry out experiments in materials science, fluid physics and biosciences, as well as perform a number of technological applications. Photo credit: NASA/Kim Shiflett
NASA Technical Reports Server (NTRS)
Gibson, J. S.; Rosen, I. G.
1986-01-01
An abstract approximation framework is developed for the finite and infinite time horizon discrete-time linear-quadratic regulator problem for systems whose state dynamics are described by a linear semigroup of operators on an infinite dimensional Hilbert space. The schemes included the framework yield finite dimensional approximations to the linear state feedback gains which determine the optimal control law. Convergence arguments are given. Examples involving hereditary and parabolic systems and the vibration of a flexible beam are considered. Spline-based finite element schemes for these classes of problems, together with numerical results, are presented and discussed.
Stochastic Adaptive Estimation and Control.
1994-10-26
Marcus, "Language Stability and Stabilizability of Discrete Event Dynamical Systems ," SIAM Journal on Control and Optimization, 31, September 1993...in the hierarchical control of flexible manufacturing systems ; in this problem, the model involves a hybrid process in continuous time whose state is...of the average cost control problem for discrete- time Markov processes. Our exposition covers from finite to Borel state and action spaces and
Time Distribution Using SpaceWire in the SCaN Testbed on ISS
NASA Technical Reports Server (NTRS)
Lux, James P.
2012-01-01
A paper describes an approach for timekeeping and time transfer among the devices on the CoNNeCT project s SCaN Testbed. It also describes how the clocks may be synchronized with an external time reference; e.g., time tags from the International Space Station (ISS) or RF signals received by a radio (TDRSS time service or GPS). All the units have some sort of counter that is fed by an oscillator at some convenient frequency. The basic problem in timekeeping is relating the counter value to some external time standard such as UTC. With SpaceWire, there are two approaches possible: one is to just use SpaceWire to send a message, and use an external wire for the sync signal. This is much the same as with the RS- 232 messages and l pps line from a GPS receiver. However, SpaceWire has an additional capability that was added to make it easier - it can insert and receive a special "timecode" word in the data stream.
NASA Technical Reports Server (NTRS)
Bandurski, R. S.; Singh, N.
1983-01-01
A major laboratory dedicated to biological-medical research is proposed for the Space Platform. The laboratory would focus on sensor physiology and biochemistry since sensory physiology represents the first impact of the new space environment on living organisms. Microgravity and the high radiation environment of space would be used to help solve the problems of prolonged sojourns in space but, more importantly, to help solve terrestrial problems of human health and agricultural productivity. The emphasis would be on experimental use of microorganisms and small plants and small animals to minimize the space and time required to use the Space Platform for maximum human betterment. The Alpha Helix Concept, that is, the use of the Space Platform to bring experimental biomedicine to a new and extreme frontier is introduced so as to better understand the worldly environment. Staffing and instrumenting the Space Platform biomedical laboratory in a manner patterned after successful terrestrial sensory physiology laboratories is also proposed.
NASA Astrophysics Data System (ADS)
Bandurski, R. S.; Singh, N.
1983-10-01
A major laboratory dedicated to biological-medical research is proposed for the Space Platform. The laboratory would focus on sensor physiology and biochemistry since sensory physiology represents the first impact of the new space environment on living organisms. Microgravity and the high radiation environment of space would be used to help solve the problems of prolonged sojourns in space but, more importantly, to help solve terrestrial problems of human health and agricultural productivity. The emphasis would be on experimental use of microorganisms and small plants and small animals to minimize the space and time required to use the Space Platform for maximum human betterment. The Alpha Helix Concept, that is, the use of the Space Platform to bring experimental biomedicine to a new and extreme frontier is introduced so as to better understand the worldly environment. Staffing and instrumenting the Space Platform biomedical laboratory in a manner patterned after successful terrestrial sensory physiology laboratories is also proposed.
Perspectives on the geographic stability and mobility of people in cities
Hanson, Susan
2005-01-01
A class of questions in the human environment sciences focuses on the relationship between individual or household behavior and local geographic context. Central to these questions is the nature of people's geographic mobility as well as the duration of their locational stability at varying spatial and temporal scales. The problem for researchers is that the processes of mobility/stability are temporally and spatially dynamic and therefore difficult to measure. Whereas time and space are continuous, analysts must select levels of aggregation for both length of time in place and spatial scale of place that fit with the problem in question. Previous work has emphasized mobility and suppressed stability as an analytic category. I focus here on stability and show how analyzing individuals' stability requires also analyzing their mobility. Through an empirical example centered on the relationship between entrepreneurship and place, I demonstrate how a spotlight on stability illuminates a resolution to the measurement problem by highlighting the interdependence between the time and space dimensions of stability/mobility. PMID:16230616
Identification of time-varying structural dynamic systems - An artificial intelligence approach
NASA Technical Reports Server (NTRS)
Glass, B. J.; Hanagud, S.
1992-01-01
An application of the artificial intelligence-derived methodologies of heuristic search and object-oriented programming to the problem of identifying the form of the model and the associated parameters of a time-varying structural dynamic system is presented in this paper. Possible model variations due to changes in boundary conditions or configurations of a structure are organized into a taxonomy of models, and a variant of best-first search is used to identify the model whose simulated response best matches that of the current physical structure. Simulated model responses are verified experimentally. An output-error approach is used in a discontinuous model space, and an equation-error approach is used in the parameter space. The advantages of the AI methods used, compared with conventional programming techniques for implementing knowledge structuring and inheritance, are discussed. Convergence conditions and example problems have been discussed. In the example problem, both the time-varying model and its new parameters have been identified when changes occur.
Three-Axis Time-Optimal Attitude Maneuvers of a Rigid-Body
NASA Astrophysics Data System (ADS)
Wang, Xijing; Li, Jisheng
With the development trends for modern satellites towards macro-scale and micro-scale, new demands are requested for its attitude adjustment. Precise pointing control and rapid maneuvering capabilities have long been part of many space missions. While the development of computer technology enables new optimal algorithms being used continuously, a powerful tool for solving problem is provided. Many papers about attitude adjustment have been published, the configurations of the spacecraft are considered rigid body with flexible parts or gyrostate-type systems. The object function always include minimum time or minimum fuel. During earlier satellite missions, the attitude acquisition was achieved by using the momentum ex change devices, performed by a sequential single-axis slewing strategy. Recently, the simultaneous three-axis minimum-time maneuver(reorientation) problems have been studied by many researchers. It is important to research the minimum-time maneuver of a rigid spacecraft within onboard power limits, because of potential space application such as surveying multiple targets in space and academic value. The minimum-time maneuver of a rigid spacecraft is a basic problem because the solutions for maneuvering flexible spacecraft are based on the solution to the rigid body slew problem. A new method for the open-loop solution for a rigid spacecraft maneuver is presented. Having neglected all perturbation torque, the necessary conditions of spacecraft from one state to another state can be determined. There is difference between single-axis with multi-axis. For single- axis analytical solution is possible and the switching line passing through the state-space origin belongs to parabolic. For multi-axis, it is impossible to get analytical solution due to the dynamic coupling between the axes and must be solved numerically. Proved by modern research, Euler axis rotations are quasi-time-optimal in general. On the basis of minimum value principles, a research for reorienting an inertial syrnmetric spacecraft with time cost function from an initial state of rest to a final state of rest is deduced. And the solution to it is stated below: Firstly, the essential condition for solving the problem is deduced with the minimum value principle. The necessary conditions for optimality yield a two point boundary-value problem (TPBVP), which, when solved, produces the control history that minimize time performance index. In the nonsingular control, the solution is the' bang-bang maneuver. The control profile is characterized by Saturated controls for the entire maneuver. The singular control maybe existed. It is only singular in mathematics. According to physical principle, the bigger the mode of the control torque is, the shorter the time is. So saturated controls are used in singular control. Secondly, the control parameters are always in maximum, so the key problem is to determine switch point thus original problem is changed to find the changing time. By the use of adjusting the switch on/off time, the genetic algorithm, which is a new robust method is optimized to determine the switch features without the gyroscopic coupling. There is improvement upon the traditional GA in this research. The homotopy method to find the nonlinear algebra is based on rigorous topology continuum theory. Based on the idea of the homotopy, the relaxation parameters are introduced, and the switch point is figured out with simulated annealing. Computer simulation results using a rigid body show that the new method is feasible and efficient. A practical method of computing approximate solutions to the time-optimal control- switch times for rigid body reorientation has been developed.
Facing the PA team growth problem in space programs: The Hermes approach
NASA Astrophysics Data System (ADS)
Bernede, G.
1991-08-01
The increasing need for product assurance (PA) in space programs makes finding skilled and trained engineers willing to work in the PA field a top level difficulty. Few engineers are available and exchanging them from one project to another is not a solution. No reservoir of talent exists for the time being in Europe. Educational efforts have started in several countries to have product assurance techniques become part of basic engineering knowledge. These efforts will bring strong improvement compared to today's situation but not in time for programs such as Hermes. Peculiar solutions must be developed. Approaches used to solve this problem within the Hermes program are outlined.
On the emergence of the structure of physics
NASA Astrophysics Data System (ADS)
Majid, S.
2018-04-01
We consider Hilbert's problem of the axioms of physics at a qualitative or conceptual level. This is more pressing than ever as we seek to understand how both general relativity and quantum theory could emerge from some deeper theory of quantum gravity, and in this regard I have previously proposed a principle of self-duality or quantum Born reciprocity as a key structure. Here, I outline some of my recent work around the idea of quantum space-time as motivated by this non-standard philosophy, including a new toy model of gravity on a space-time consisting of four points forming a square. This article is part of the theme issue `Hilbert's sixth problem'.
On the emergence of the structure of physics.
Majid, S
2018-04-28
We consider Hilbert's problem of the axioms of physics at a qualitative or conceptual level. This is more pressing than ever as we seek to understand how both general relativity and quantum theory could emerge from some deeper theory of quantum gravity, and in this regard I have previously proposed a principle of self-duality or quantum Born reciprocity as a key structure. Here, I outline some of my recent work around the idea of quantum space-time as motivated by this non-standard philosophy, including a new toy model of gravity on a space-time consisting of four points forming a square.This article is part of the theme issue 'Hilbert's sixth problem'. © 2018 The Author(s).
NASA Astrophysics Data System (ADS)
Simoni, L.; Secchi, S.; Schrefler, B. A.
2008-12-01
This paper analyses the numerical difficulties commonly encountered in solving fully coupled numerical models and proposes a numerical strategy apt to overcome them. The proposed procedure is based on space refinement and time adaptivity. The latter, which in mainly studied here, is based on the use of a finite element approach in the space domain and a Discontinuous Galerkin approximation within each time span. Error measures are defined for the jump of the solution at each time station. These constitute the parameters allowing for the time adaptivity. Some care is however, needed for a useful definition of the jump measures. Numerical tests are presented firstly to demonstrate the advantages and shortcomings of the method over the more traditional use of finite differences in time, then to assess the efficiency of the proposed procedure for adapting the time step. The proposed method reveals its efficiency and simplicity to adapt the time step in the solution of coupled field problems.
The astronaut and the banana peel: An EVA retriever scenario
NASA Technical Reports Server (NTRS)
Shapiro, Daniel G.
1989-01-01
To prepare for the problem of accidents in Space Station activities, the Extravehicular Activity Retriever (EVAR) robot is being constructed, whose purpose is to retrieve astronauts and tools that float free of the Space Station. Advanced Decision Systems is at the beginning of a project to develop research software capable of guiding EVAR through the retrieval process. This involves addressing problems in machine vision, dexterous manipulation, real time construction of programs via speech input, and reactive execution of plans despite the mishaps and unexpected conditions that arise in uncontrolled domains. The problem analysis phase of this work is presented. An EVAR scenario is used to elucidate major domain and technical problems. An overview of the technical approach to prototyping an EVAR system is also presented.
Learning characteristics of a space-time neural network as a tether skiprope observer
NASA Technical Reports Server (NTRS)
Lea, Robert N.; Villarreal, James A.; Jani, Yashvant; Copeland, Charles
1993-01-01
The Software Technology Laboratory at the Johnson Space Center is testing a Space Time Neural Network (STNN) for observing tether oscillations present during retrieval of a tethered satellite. Proper identification of tether oscillations, known as 'skiprope' motion, is vital to safe retrieval of the tethered satellite. Our studies indicate that STNN has certain learning characteristics that must be understood properly to utilize this type of neural network for the tethered satellite problem. We present our findings on the learning characteristics including a learning rate versus momentum performance table.
Child- and elder-friendly urban public places in Fatahillah Square Historical District
NASA Astrophysics Data System (ADS)
Srinaga, F.; LKatoppo, M.; Hidayat, J.
2018-03-01
Fatahillah square as an important historical urban square in Jakarta has problems in eye level area integrative processing. Visitors cannot enjoy their time while in the square regarding their visuals, feelings, space, and bodies comfort. These also lead to other problems in which the square is lack of friendly and convenient places for children, the elderly and also the disabled, especially people with limited moving space. The research will attempt in proposing design inception for the Fatahillah Square that is using inclusive user-centered design approach, while in the same time incorporate theoretical studies of children and elderly-design considerations. The first stage of this research was building inclusive design parameter; begin with a context-led research which assesses the quality of Fatahillah square through three basic components of urban space: hardware, software and orgware. The second stage of this research is to propose inclusive design inception for the Fatahillah square.
Electro-Optic Time-to-Space Converter for Optical Detector Jitter Mitigation
NASA Technical Reports Server (NTRS)
Birnbaum, Kevin; Farr, William
2013-01-01
A common problem in optical detection is determining the arrival time of a weak optical pulse that may comprise only one to a few photons. Currently, this problem is solved by using a photodetector to convert the optical signal to an electronic signal. The timing of the electrical signal is used to infer the timing of the optical pulse, but error is introduced by random delay between the absorption of the optical pulse and the creation of the electrical one. To eliminate this error, a time-to-space converter separates a sequence of optical pulses and sends them to different photodetectors, depending on their arrival time. The random delay, called jitter, is at least 20 picoseconds for the best detectors capable of detecting the weakest optical pulses, a single photon, and can be as great as 500 picoseconds. This limits the resolution with which the timing of the optical pulse can be measured. The time-to-space converter overcomes this limitation. Generally, the time-to-space converter imparts a time-dependent momentum shift to the incoming optical pulses, followed by an optical system that separates photons of different momenta. As an example, an electro-optic phase modulator can be used to apply longitudinal momentum changes (frequency changes) that vary in time, followed by an optical spectrometer (such as a diffraction grating), which separates photons with different momenta into different paths and directs them to impinge upon an array of photodetectors. The pulse arrival time is then inferred by measuring which photodetector receives the pulse. The use of a time-to-space converter mitigates detector jitter and improves the resolution with which the timing of an optical pulse is determined. Also, the application of the converter enables the demodulation of a pulse position modulated signal (PPM) at higher bandwidths than using previous photodetector technology. This allows the creation of a receiver for a communication system with high bandwidth and high bits/photon efficiency.
Engineering Risk Assessment of Space Thruster Challenge Problem
NASA Technical Reports Server (NTRS)
Mathias, Donovan L.; Mattenberger, Christopher J.; Go, Susie
2014-01-01
The Engineering Risk Assessment (ERA) team at NASA Ames Research Center utilizes dynamic models with linked physics-of-failure analyses to produce quantitative risk assessments of space exploration missions. This paper applies the ERA approach to the baseline and extended versions of the PSAM Space Thruster Challenge Problem, which investigates mission risk for a deep space ion propulsion system with time-varying thruster requirements and operations schedules. The dynamic mission is modeled using a combination of discrete and continuous-time reliability elements within the commercially available GoldSim software. Loss-of-mission (LOM) probability results are generated via Monte Carlo sampling performed by the integrated model. Model convergence studies are presented to illustrate the sensitivity of integrated LOM results to the number of Monte Carlo trials. A deterministic risk model was also built for the three baseline and extended missions using the Ames Reliability Tool (ART), and results are compared to the simulation results to evaluate the relative importance of mission dynamics. The ART model did a reasonable job of matching the simulation models for the baseline case, while a hybrid approach using offline dynamic models was required for the extended missions. This study highlighted that state-of-the-art techniques can adequately adapt to a range of dynamic problems.
Evolutionary space platform concept study. Volume 2, part A: SASP special emphasis trade studies
NASA Technical Reports Server (NTRS)
1982-01-01
Efforts are in progress to define an approach to provide a simple and cost effective solution to the problem of long duration space flight. This approach involves a Space Platform in low Earth orbit, which can be tended by the Space Shuttle and which will provide, for extended periods of time, stability, utilities and access for a variety of replaceable payloads. The feasibility of an evolutionary space system which would cost effectively support unmanned payloads in groups, using a Space Platform which provides centralized basic subsystems is addressed.
A space-efficient algorithm for local similarities.
Huang, X Q; Hardison, R C; Miller, W
1990-10-01
Existing dynamic-programming algorithms for identifying similar regions of two sequences require time and space proportional to the product of the sequence lengths. Often this space requirement is more limiting than the time requirement. We describe a dynamic-programming local-similarity algorithm that needs only space proportional to the sum of the sequence lengths. The method can also find repeats within a single long sequence. To illustrate the algorithm's potential, we discuss comparison of a 73,360 nucleotide sequence containing the human beta-like globin gene cluster and a corresponding 44,594 nucleotide sequence for rabbit, a problem well beyond the capabilities of other dynamic-programming software.
Schmidhuber, Jürgen
2013-01-01
Most of computer science focuses on automatically solving given computational problems. I focus on automatically inventing or discovering problems in a way inspired by the playful behavior of animals and humans, to train a more and more general problem solver from scratch in an unsupervised fashion. Consider the infinite set of all computable descriptions of tasks with possibly computable solutions. Given a general problem-solving architecture, at any given time, the novel algorithmic framework PowerPlay (Schmidhuber, 2011) searches the space of possible pairs of new tasks and modifications of the current problem solver, until it finds a more powerful problem solver that provably solves all previously learned tasks plus the new one, while the unmodified predecessor does not. Newly invented tasks may require to achieve a wow-effect by making previously learned skills more efficient such that they require less time and space. New skills may (partially) re-use previously learned skills. The greedy search of typical PowerPlay variants uses time-optimal program search to order candidate pairs of tasks and solver modifications by their conditional computational (time and space) complexity, given the stored experience so far. The new task and its corresponding task-solving skill are those first found and validated. This biases the search toward pairs that can be described compactly and validated quickly. The computational costs of validating new tasks need not grow with task repertoire size. Standard problem solver architectures of personal computers or neural networks tend to generalize by solving numerous tasks outside the self-invented training set; PowerPlay’s ongoing search for novelty keeps breaking the generalization abilities of its present solver. This is related to Gödel’s sequence of increasingly powerful formal theories based on adding formerly unprovable statements to the axioms without affecting previously provable theorems. The continually increasing repertoire of problem-solving procedures can be exploited by a parallel search for solutions to additional externally posed tasks. PowerPlay may be viewed as a greedy but practical implementation of basic principles of creativity (Schmidhuber, 2006a, 2010). A first experimental analysis can be found in separate papers (Srivastava et al., 2012a,b, 2013). PMID:23761771
CHLOROPHYLL A DISTRIBUTION IN NARRAGANSETT BAY, RI: USING A SPECTRAL CURVATURE ALGORITHM
Chlorophyll a, a primary indicator of eutrophication in estuarine waters, varies enough in time and space to create spatial problems when monitored by satellite, and temporal problems when measured with in situ field programs. Using aircraft to sense ocean color of local waters, ...
On the usefulness of relativistic space-times for the description of the Earth's gravitational field
NASA Astrophysics Data System (ADS)
Soffel, Michael; Frutos, Francisco
2016-12-01
The usefulness of relativistic space-times for the description of the Earth's gravitational field is investigated. A variety of exact vacuum solutions of Einstein's field equations (Schwarzschild, Erez and Rosen, Gutsunayev and Manko, Hernández-Pastora and Martín, Kerr, Quevedo, and Mashhoon) are investigated in that respect. It is argued that because of their multipole structure and influences from external bodies, all these exact solutions are not really useful for the central problem. Then, approximate space-times resulting from an MPM or post-Newtonian approximation are considered. Only in the DSX formalism that is of the first post-Newtonian order, all aspects of the problem can be tackled: a relativistic description (a) of the Earth's gravity field in a well-defined geocentric reference system (GCRS), (b) of the motion of solar system bodies in a barycentric reference system (BCRS), and (c) of inertial and tidal terms in the geocentric metric describing the external gravitational field. A relativistic SLR theory is also discussed with respect to our central problem. Orders of magnitude of many effects related to the Earth's gravitational field and SLR are given. It is argued that a formalism with accuracies better than of the first post-Newtonian order is not yet available.
Representation of magnetic fields in space
NASA Technical Reports Server (NTRS)
Stern, D. P.
1975-01-01
Several methods by which a magnetic field in space can be represented are reviewed with particular attention to problems of the observed geomagnetic field. Time dependence is assumed to be negligible, and five main classes of representation are described by vector potential, scalar potential, orthogonal vectors, Euler potentials, and expanded magnetic field.
Space transportation system shuttle turnabout analysis report
NASA Technical Reports Server (NTRS)
Reedy, R. E.
1979-01-01
The progress made and the problems encountered by the various program elements of the shuttle program in achieving the 160 hour ground turnaround goal are presented and evaluated. Task assessment time is measured against the program allocation time.
NASA Technical Reports Server (NTRS)
1998-01-01
A Space Act Agreement between Kennedy Space Center and Surtreat Southeast, Inc., resulted in a new treatment that keeps buildings from corroding away over time. Structural corrosion is a multi-billion dollar problem in the United States. The agreement merged Kennedy Space Center's research into electrical treatments of structural corrosion with chemical processes developed by Surtreat. Combining NASA and Surtreat technologies has resulted in a unique process with broad corrosion-control applications.
NASA Astrophysics Data System (ADS)
Zalaletdinov, R. M.
1998-04-01
The averaging problem in general relativity is briefly discussed. A new setting of the problem as that of macroscopic description of gravitation is proposed. A covariant space-time averaging procedure is described. The structure of the geometry of macroscopic space-time, which follows from averaging Cartan's structure equations, is described and the correlation tensors present in the theory are discussed. The macroscopic field equations (averaged Einstein's equations) derived in the framework of the approach are presented and their structure is analysed. The correspondence principle for macroscopic gravity is formulated and a definition of the stress-energy tensor for the macroscopic gravitational field is proposed. It is shown that the physical meaning of using Einstein's equations with a hydrodynamic stress-energy tensor in looking for cosmological models means neglecting all gravitational field correlations. The system of macroscopic gravity equations to be solved when the correlations are taken into consideration is given and described.
A fully Sinc-Galerkin method for Euler-Bernoulli beam models
NASA Technical Reports Server (NTRS)
Smith, R. C.; Bowers, K. L.; Lund, J.
1990-01-01
A fully Sinc-Galerkin method in both space and time is presented for fourth-order time-dependent partial differential equations with fixed and cantilever boundary conditions. The Sinc discretizations for the second-order temporal problem and the fourth-order spatial problems are presented. Alternate formulations for variable parameter fourth-order problems are given which prove to be especially useful when applying the forward techniques to parameter recovery problems. The discrete system which corresponds to the time-dependent partial differential equations of interest are then formulated. Computational issues are discussed and a robust and efficient algorithm for solving the resulting matrix system is outlined. Numerical results which highlight the method are given for problems with both analytic and singular solutions as well as fixed and cantilever boundary conditions.
The Cauchy Problem in Local Spaces for the Complex Ginzburg-Landau EquationII. Contraction Methods
NASA Astrophysics Data System (ADS)
Ginibre, J.; Velo, G.
We continue the study of the initial value problem for the complex Ginzburg-Landau equation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Angel, Jordan B.; Banks, Jeffrey W.; Henshaw, William D.
High-order accurate upwind approximations for the wave equation in second-order form on overlapping grids are developed. Although upwind schemes are well established for first-order hyperbolic systems, it was only recently shown by Banks and Henshaw how upwinding could be incorporated into the second-order form of the wave equation. This new upwind approach is extended here to solve the time-domain Maxwell's equations in second-order form; schemes of arbitrary order of accuracy are formulated for general curvilinear grids. Taylor time-stepping is used to develop single-step space-time schemes, and the upwind dissipation is incorporated by embedding the exact solution of a local Riemannmore » problem into the discretization. Second-order and fourth-order accurate schemes are implemented for problems in two and three space dimensions, and overlapping grids are used to treat complex geometry and problems with multiple materials. Stability analysis of the upwind-scheme on overlapping grids is performed using normal mode theory. The stability analysis and computations confirm that the upwind scheme remains stable on overlapping grids, including the difficult case of thin boundary grids when the traditional non-dissipative scheme becomes unstable. The accuracy properties of the scheme are carefully evaluated on a series of classical scattering problems for both perfect conductors and dielectric materials in two and three space dimensions. Finally, the upwind scheme is shown to be robust and provide high-order accuracy.« less
Angel, Jordan B.; Banks, Jeffrey W.; Henshaw, William D.
2017-09-28
High-order accurate upwind approximations for the wave equation in second-order form on overlapping grids are developed. Although upwind schemes are well established for first-order hyperbolic systems, it was only recently shown by Banks and Henshaw how upwinding could be incorporated into the second-order form of the wave equation. This new upwind approach is extended here to solve the time-domain Maxwell's equations in second-order form; schemes of arbitrary order of accuracy are formulated for general curvilinear grids. Taylor time-stepping is used to develop single-step space-time schemes, and the upwind dissipation is incorporated by embedding the exact solution of a local Riemannmore » problem into the discretization. Second-order and fourth-order accurate schemes are implemented for problems in two and three space dimensions, and overlapping grids are used to treat complex geometry and problems with multiple materials. Stability analysis of the upwind-scheme on overlapping grids is performed using normal mode theory. The stability analysis and computations confirm that the upwind scheme remains stable on overlapping grids, including the difficult case of thin boundary grids when the traditional non-dissipative scheme becomes unstable. The accuracy properties of the scheme are carefully evaluated on a series of classical scattering problems for both perfect conductors and dielectric materials in two and three space dimensions. Finally, the upwind scheme is shown to be robust and provide high-order accuracy.« less
NASA Astrophysics Data System (ADS)
Angel, Jordan B.; Banks, Jeffrey W.; Henshaw, William D.
2018-01-01
High-order accurate upwind approximations for the wave equation in second-order form on overlapping grids are developed. Although upwind schemes are well established for first-order hyperbolic systems, it was only recently shown by Banks and Henshaw [1] how upwinding could be incorporated into the second-order form of the wave equation. This new upwind approach is extended here to solve the time-domain Maxwell's equations in second-order form; schemes of arbitrary order of accuracy are formulated for general curvilinear grids. Taylor time-stepping is used to develop single-step space-time schemes, and the upwind dissipation is incorporated by embedding the exact solution of a local Riemann problem into the discretization. Second-order and fourth-order accurate schemes are implemented for problems in two and three space dimensions, and overlapping grids are used to treat complex geometry and problems with multiple materials. Stability analysis of the upwind-scheme on overlapping grids is performed using normal mode theory. The stability analysis and computations confirm that the upwind scheme remains stable on overlapping grids, including the difficult case of thin boundary grids when the traditional non-dissipative scheme becomes unstable. The accuracy properties of the scheme are carefully evaluated on a series of classical scattering problems for both perfect conductors and dielectric materials in two and three space dimensions. The upwind scheme is shown to be robust and provide high-order accuracy.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Xia, Yidong; Liu, Xiaodong; Luo, Hong
2015-06-01
Here, a space and time third-order discontinuous Galerkin method based on a Hermite weighted essentially non-oscillatory reconstruction is presented for the unsteady compressible Euler and Navier–Stokes equations. At each time step, a lower-upper symmetric Gauss–Seidel preconditioned generalized minimal residual solver is used to solve the systems of linear equations arising from an explicit first stage, single diagonal coefficient, diagonally implicit Runge–Kutta time integration scheme. The performance of the developed method is assessed through a variety of unsteady flow problems. Numerical results indicate that this method is able to deliver the designed third-order accuracy of convergence in both space and time,more » while requiring remarkably less storage than the standard third-order discontinous Galerkin methods, and less computing time than the lower-order discontinous Galerkin methods to achieve the same level of temporal accuracy for computing unsteady flow problems.« less
Before the long journey: Development of Soviet space biology and medicine
NASA Technical Reports Server (NTRS)
Gazenko, O. G.
1978-01-01
Academician O. Gazenko, Chief of the Institute of Biomedical Problems, USSR Ministry of Public Health, reviewed the short but intense history of Soviet research in space biology and medicine. The solid academic approach of the Soviet Academy of Sciences in giving a good start at the very beginning of the space age is stressed and key people and institutions who initiated these studies are named. The basic feature of the first period of space biology is seen as the search for answers to a few fundamental questions of survival in space. It is pointed out that the initiated investigations were replaced by refined, in-depth studies of the biological, biophysical, and biochemical processes in human organism in the space environment and the search for methods which should enable cosmonaut crews to live in space for several years during interplanetary journeys. Discussing the typical problems of this effort, Gazenko each time showed how they benefit medical science and practice in general.
NASA Technical Reports Server (NTRS)
1971-01-01
The optimal allocation of resources to the national space program over an extended time period requires the solution of a large combinatorial problem in which the program elements are interdependent. The computer model uses an accelerated search technique to solve this problem. The model contains a large number of options selectable by the user to provide flexible input and a broad range of output for use in sensitivity analyses of all entering elements. Examples of these options are budget smoothing under varied appropriation levels, entry of inflation and discount effects, and probabilistic output which provides quantified degrees of certainty that program costs will remain within planned budget. Criteria and related analytic procedures were established for identifying potential new space program directions. Used in combination with the optimal resource allocation model, new space applications can be analyzed in realistic perspective, including the advantage gain from existing space program plant and on-going programs such as the space transportation system.
An evaluation of superminicomputers for thermal analysis
NASA Technical Reports Server (NTRS)
Storaasli, O. O.; Vidal, J. B.; Jones, G. K.
1982-01-01
The use of superminicomputers for solving a series of increasingly complex thermal analysis problems is investigated. The approach involved (1) installation and verification of the SPAR thermal analyzer software on superminicomputers at Langley Research Center and Goddard Space Flight Center, (2) solution of six increasingly complex thermal problems on this equipment, and (3) comparison of solution (accuracy, CPU time, turnaround time, and cost) with solutions on large mainframe computers.
NASA Technical Reports Server (NTRS)
Melkumov, T. M.
1977-01-01
The research for more efficient methods of propelling a spacecraft, than can be achieved with chemical energy, was studied. During a time when rockets for space flight had not actually been built pioneers in rocket technology were already concerned with this problem. Alternative sources proposed at that time, were nuclear and solar energy. Basic engineering problems of each source were investigated.
HEMP 3D: A finite difference program for calculating elastic-plastic flow, appendix B
NASA Astrophysics Data System (ADS)
Wilkins, Mark L.
1993-05-01
The HEMP 3D program can be used to solve problems in solid mechanics involving dynamic plasticity and time dependent material behavior and problems in gas dynamics. The equations of motion, the conservation equations, and the constitutive relations listed below are solved by finite difference methods following the format of the HEMP computer simulation program formulated in two space dimensions and time.
Exact Magnetic Diffusion Solutions for Magnetohydrodynamic Code Verification
DOE Office of Scientific and Technical Information (OSTI.GOV)
Miller, D S
In this paper, the authors present several new exact analytic space and time dependent solutions to the problem of magnetic diffusion in R-Z geometry. These problems serve to verify several different elements of an MHD implementation: magnetic diffusion, external circuit time integration, current and voltage energy sources, spatially dependent conductivities, and ohmic heating. The exact solutions are shown in comparison with 2D simulation results from the Ares code.
Deep Learning for Flow Sculpting: Insights into Efficient Learning using Scientific Simulation Data
NASA Astrophysics Data System (ADS)
Stoecklein, Daniel; Lore, Kin Gwn; Davies, Michael; Sarkar, Soumik; Ganapathysubramanian, Baskar
2017-04-01
A new technique for shaping microfluid flow, known as flow sculpting, offers an unprecedented level of passive fluid flow control, with potential breakthrough applications in advancing manufacturing, biology, and chemistry research at the microscale. However, efficiently solving the inverse problem of designing a flow sculpting device for a desired fluid flow shape remains a challenge. Current approaches struggle with the many-to-one design space, requiring substantial user interaction and the necessity of building intuition, all of which are time and resource intensive. Deep learning has emerged as an efficient function approximation technique for high-dimensional spaces, and presents a fast solution to the inverse problem, yet the science of its implementation in similarly defined problems remains largely unexplored. We propose that deep learning methods can completely outpace current approaches for scientific inverse problems while delivering comparable designs. To this end, we show how intelligent sampling of the design space inputs can make deep learning methods more competitive in accuracy, while illustrating their generalization capability to out-of-sample predictions.
The urban public space betterment and land use sustainability Under the human behavior
NASA Astrophysics Data System (ADS)
Zhao, Xiaofan; Ji, Yanning
2018-02-01
This paper analyzes the differences between Chinese and western public life and environmental behavior habits. Identify specific needs for Chinese urban public Spaces. At the same time, the paper analyzes the problems related to urban construction in China, including micro-land use, transportation and urban pattern. The solution of Chinese urban public space layout is proposed and the prospects of sustainable urban public space. Urban betterment are prospected in the future.
NASA Astrophysics Data System (ADS)
Glazyrina, O. V.; Pavlova, M. F.
2016-11-01
We consider the parabolic inequality with monotone with respect to a gradient space operator, which is depended on integral with respect to space variables solution characteristic. We construct a two-layer differential scheme for this problem with use of penalty method, semidiscretization with respect to time variable method and the finite element method (FEM) with respect to space variables. We proved a convergence of constructed mothod.
16S rRNA beacons for bacterial monitoring during human space missions.
Larios-Sanz, Maia; Kourentzi, Katerina D; Warmflash, David; Jones, Jeffrey; Pierson, Duane L; Willson, Richard C; Fox, George E
2007-04-01
Microorganisms are unavoidable in space environments and their presence has, at times, been a source of problems. Concerns about disease during human space missions are particularly important considering the significant changes the immune system incurs during spaceflight and the history of microbial contamination aboard the Mir space station. Additionally, these contaminants may have adverse effects on instrumentation and life-support systems. A sensitive, highly specific system to detect, characterize, and monitor these microbial populations is essential. Herein we describe a monitoring approach that uses 16S rRNA targeted molecular beacons to successfully detect several specific bacterial groupings. This methodology will greatly simplify in-flight monitoring by minimizing sample handling and processing. We also address and provide solutions to target accessibility problems encountered in hybridizations that target 16S rRNA.
Phase space simulation of collisionless stellar systems on the massively parallel processor
NASA Technical Reports Server (NTRS)
White, Richard L.
1987-01-01
A numerical technique for solving the collisionless Boltzmann equation describing the time evolution of a self gravitating fluid in phase space was implemented on the Massively Parallel Processor (MPP). The code performs calculations for a two dimensional phase space grid (with one space and one velocity dimension). Some results from calculations are presented. The execution speed of the code is comparable to the speed of a single processor of a Cray-XMP. Advantages and disadvantages of the MPP architecture for this type of problem are discussed. The nearest neighbor connectivity of the MPP array does not pose a significant obstacle. Future MPP-like machines should have much more local memory and easier access to staging memory and disks in order to be effective for this type of problem.
Crew collaboration in space: a naturalistic decision-making perspective
NASA Technical Reports Server (NTRS)
Orasanu, Judith
2005-01-01
Successful long-duration space missions will depend on the ability of crewmembers to respond promptly and effectively to unanticipated problems that arise under highly stressful conditions. Naturalistic decision making (NDM) exploits the knowledge and experience of decision makers in meaningful work domains, especially complex sociotechnical systems, including aviation and space. Decision making in these ambiguous, dynamic, high-risk environments is a complex task that involves defining the nature of the problem and crafting a response to achieve one's goals. Goal conflicts, time pressures, and uncertain outcomes may further complicate the process. This paper reviews theory and research pertaining to the NDM model and traces some of the implications for space crews and other groups that perform meaningful work in extreme environments. It concludes with specific recommendations for preparing exploration crews to use NDM effectively.
Stability and chaos in Kustaanheimo-Stiefel space induced by the Hopf fibration
NASA Astrophysics Data System (ADS)
Roa, Javier; Urrutxua, Hodei; Peláez, Jesús
2016-07-01
The need for the extra dimension in Kustaanheimo-Stiefel (KS) regularization is explained by the topology of the Hopf fibration, which defines the geometry and structure of KS space. A trajectory in Cartesian space is represented by a four-dimensional manifold called the fundamental manifold. Based on geometric and topological aspects classical concepts of stability are translated to KS language. The separation between manifolds of solutions generalizes the concept of Lyapunov stability. The dimension-raising nature of the fibration transforms fixed points, limit cycles, attractive sets, and Poincaré sections to higher dimensional subspaces. From these concepts chaotic systems are studied. In strongly perturbed problems, the numerical error can break the topological structure of KS space: points in a fibre are no longer transformed to the same point in Cartesian space. An observer in three dimensions will see orbits departing from the same initial conditions but diverging in time. This apparent randomness of the integration can only be understood in four dimensions. The concept of topological stability results in a simple method for estimating the time-scale in which numerical simulations can be trusted. Ideally, all trajectories departing from the same fibre should be KS transformed to a unique trajectory in three-dimensional space, because the fundamental manifold that they constitute is unique. By monitoring how trajectories departing from one fibre separate from the fundamental manifold a critical time, equivalent to the Lyapunov time, is estimated. These concepts are tested on N-body examples: the Pythagorean problem, and an example of field stars interacting with a binary.
NASA--has its biological groundwork for a trip to Mars improved?
Haddy, Francis J
2007-03-01
In a 1991 editorial in The FASEB Journal, Robert W. Krauss commented on a recent report of the Presidential Advisory Committee on the Future of the U.S. Space Program (Augustine report). He concluded that, although a manned mission to Mars with life sciences as the priority was endorsed by the Committee, it failed to deal realistically with one huge gap; biological sciences have never been given high priority. According to Krauss, this left a void that will cripple, perhaps fatally, any early effort to ensure long-term survival on any mission of extended duration. The gap included insufficient flight time for fundamental biological space research and insufficient funds. Krauss expressed his opinions 15 years ago. Have we better knowledge of space biology now? This question becomes more acute now that President George W. Bush recently proposed a manned return to the moon by 2015 or 2020, with the moon to become our staging post for manned missions to Mars. Will we be ready so soon? A review of the progress in the last 15 years suggests that we will not. Because of the Columbia disaster, flight opportunities for biological sciences in shuttle spacelabs and in Space Station laboratories compete with time for engineering problems and construction. Thus, research on gravity, radiation, and isolation loses out to problems deemed to be of higher priority. Radiation in deep space and graded gravity in space with on board centrifuges are areas that must be studied before we undertake prolonged space voyages. Very recent budgetary changes within National Aeronautics and Space Administration threaten to greatly reduce the fundamental space biology funds. Are we ready for a trip to Mars? Like Krauss 15 years ago, I think not for some time.
On Information Storage Models.
ERIC Educational Resources Information Center
Leimkuhler, Ferdinand F.
The transfer of information through space and time in communication systems is often accompanied by significant delays which give rise to meaningful storage problems. Mathematical models have been developed for the study of these kinds of problems which are applicable to the design of manual, library-type, or mechanized information storage and…
NASA Astrophysics Data System (ADS)
Boscheri, Walter; Dumbser, Michael
2014-10-01
In this paper we present a new family of high order accurate Arbitrary-Lagrangian-Eulerian (ALE) one-step ADER-WENO finite volume schemes for the solution of nonlinear systems of conservative and non-conservative hyperbolic partial differential equations with stiff source terms on moving tetrahedral meshes in three space dimensions. A WENO reconstruction technique is used to achieve high order of accuracy in space, while an element-local space-time Discontinuous Galerkin finite element predictor on moving curved meshes is used to obtain a high order accurate one-step time discretization. Within the space-time predictor the physical element is mapped onto a reference element using a high order isoparametric approach, where the space-time basis and test functions are given by the Lagrange interpolation polynomials passing through a predefined set of space-time nodes. Since our algorithm is cell-centered, the final mesh motion is computed by using a suitable node solver algorithm. A rezoning step as well as a flattener strategy are used in some of the test problems to avoid mesh tangling or excessive element deformations that may occur when the computation involves strong shocks or shear waves. The ALE algorithm presented in this article belongs to the so-called direct ALE methods because the final Lagrangian finite volume scheme is based directly on a space-time conservation formulation of the governing PDE system, with the rezoned geometry taken already into account during the computation of the fluxes. We apply our new high order unstructured ALE schemes to the 3D Euler equations of compressible gas dynamics, for which a set of classical numerical test problems has been solved and for which convergence rates up to sixth order of accuracy in space and time have been obtained. We furthermore consider the equations of classical ideal magnetohydrodynamics (MHD) as well as the non-conservative seven-equation Baer-Nunziato model of compressible multi-phase flows with stiff relaxation source terms.
Performance and state-space analyses of systems using Petri nets
NASA Technical Reports Server (NTRS)
Watson, James Francis, III
1992-01-01
The goal of any modeling methodology is to develop a mathematical description of a system that is accurate in its representation and also permits analysis of structural and/or performance properties. Inherently, trade-offs exist between the level detail in the model and the ease with which analysis can be performed. Petri nets (PN's), a highly graphical modeling methodology for Discrete Event Dynamic Systems, permit representation of shared resources, finite capacities, conflict, synchronization, concurrency, and timing between state changes. By restricting the state transition time delays to the family of exponential density functions, Markov chain analysis of performance problems is possible. One major drawback of PN's is the tendency for the state-space to grow rapidly (exponential complexity) compared to increases in the PN constructs. It is the state space, or the Markov chain obtained from it, that is needed in the solution of many problems. The theory of state-space size estimation for PN's is introduced. The problem of state-space size estimation is defined, its complexities are examined, and estimation algorithms are developed. Both top-down and bottom-up approaches are pursued, and the advantages and disadvantages of each are described. Additionally, the author's research in non-exponential transition modeling for PN's is discussed. An algorithm for approximating non-exponential transitions is developed. Since only basic PN constructs are used in the approximation, theory already developed for PN's remains applicable. Comparison to results from entropy theory show the transition performance is close to the theoretic optimum. Inclusion of non-exponential transition approximations improves performance results at the expense of increased state-space size. The state-space size estimation theory provides insight and algorithms for evaluating this trade-off.
NASA Astrophysics Data System (ADS)
Dib, Alain; Kavvas, M. Levent
2018-03-01
The characteristic form of the Saint-Venant equations is solved in a stochastic setting by using a newly proposed Fokker-Planck Equation (FPE) methodology. This methodology computes the ensemble behavior and variability of the unsteady flow in open channels by directly solving for the flow variables' time-space evolutionary probability distribution. The new methodology is tested on a stochastic unsteady open-channel flow problem, with an uncertainty arising from the channel's roughness coefficient. The computed statistical descriptions of the flow variables are compared to the results obtained through Monte Carlo (MC) simulations in order to evaluate the performance of the FPE methodology. The comparisons show that the proposed methodology can adequately predict the results of the considered stochastic flow problem, including the ensemble averages, variances, and probability density functions in time and space. Unlike the large number of simulations performed by the MC approach, only one simulation is required by the FPE methodology. Moreover, the total computational time of the FPE methodology is smaller than that of the MC approach, which could prove to be a particularly crucial advantage in systems with a large number of uncertain parameters. As such, the results obtained in this study indicate that the proposed FPE methodology is a powerful and time-efficient approach for predicting the ensemble average and variance behavior, in both space and time, for an open-channel flow process under an uncertain roughness coefficient.
Propagating Qualitative Values Through Quantitative Equations
NASA Technical Reports Server (NTRS)
Kulkarni, Deepak
1992-01-01
In most practical problems where traditional numeric simulation is not adequate, one need to reason about a system with both qualitative and quantitative equations. In this paper, we address the problem of propagating qualitative values represented as interval values through quantitative equations. Previous research has produced exponential-time algorithms for approximate solution of the problem. These may not meet the stringent requirements of many real time applications. This paper advances the state of art by producing a linear-time algorithm that can propagate a qualitative value through a class of complex quantitative equations exactly and through arbitrary algebraic expressions approximately. The algorithm was found applicable to Space Shuttle Reaction Control System model.
2007-12-09
KENNEDY SPACE CENTER, FLA. -- Doug Lyons, STS-122 launch director, participates in a news briefing following the conclusion of a Mission Management Team, or MMT, meeting. The meeting followed the morning's launch scrub of the space shuttle Atlantis STS-122 mission caused by problems experienced with the external tank's engine cutoff sensor system during tanking for the second launch attempt. An announcement was made during the briefing that the STS-122 launch is postponed to no earlier than Jan. 2, 2008, to give the team time to resolve the system's problems. Atlantis will carry the Columbus Laboratory, the European Space Agency's largest contribution to the construction of the International Space Station. It will support scientific and technological research in a microgravity environment. Permanently attached to the Harmony node of the space station, the laboratory will carry out experiments in materials science, fluid physics and biosciences, as well as perform a number of technological applications. Photo credit: NASA/Kim Shiflett
2007-12-09
KENNEDY SPACE CENTER, FLA. -- LeRoy Cain, the Mission Management Team chairman, participates in a news briefing following the conclusion of a team meeting. The meeting followed the morning's launch scrub caused by problems experienced with the space shuttle Atlantis STS-122 external tank's engine cutoff sensor system during tanking for the second launch attempt. An announcement was made during the briefing that the STS-122 launch is postponed to no earlier than Jan. 2, 2008, to give the team time to resolve the system's problems. Atlantis will carry the Columbus Laboratory, the European Space Agency's largest contribution to the construction of the International Space Station. It will support scientific and technological research in a microgravity environment. Permanently attached to the Harmony node of the space station, the laboratory will carry out experiments in materials science, fluid physics and biosciences, as well as perform a number of technological applications. Photo credit: NASA/Kim Shiflett
DOE Office of Scientific and Technical Information (OSTI.GOV)
Plimak, L.I., E-mail: Lev.Plimak@mbi-berlin.de; Olsen, M.K.
2014-12-15
In this work we present the formal background used to develop the methods used in earlier works to extend the truncated Wigner representation of quantum and atom optics in order to address multi-time problems. Analogs of Wick’s theorem for the Weyl ordering are verified. Using the Bose–Hubbard chain as an example, we show how these may be applied to constructing a mapping of the system in question to phase space. Regularisation issues and the reordering problem for the Heisenberg operators are addressed.
NASA Technical Reports Server (NTRS)
Englander, Jacob; Vavrina, Matthew
2015-01-01
The customer (scientist or project manager) most often does not want just one point solution to the mission design problem Instead, an exploration of a multi-objective trade space is required. For a typical main-belt asteroid mission the customer might wish to see the trade-space of: Launch date vs. Flight time vs. Deliverable mass, while varying the destination asteroid, planetary flybys, launch year, etcetera. To address this question we use a multi-objective discrete outer-loop which defines many single objective real-valued inner-loop problems.
Space-time least-squares Petrov-Galerkin projection in nonlinear model reduction.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Choi, Youngsoo; Carlberg, Kevin Thomas
Our work proposes a space-time least-squares Petrov-Galerkin (ST-LSPG) projection method for model reduction of nonlinear dynamical systems. In contrast to typical nonlinear model-reduction methods that first apply Petrov-Galerkin projection in the spatial dimension and subsequently apply time integration to numerically resolve the resulting low-dimensional dynamical system, the proposed method applies projection in space and time simultaneously. To accomplish this, the method first introduces a low-dimensional space-time trial subspace, which can be obtained by computing tensor decompositions of state-snapshot data. The method then computes discrete-optimal approximations in this space-time trial subspace by minimizing the residual arising after time discretization over allmore » space and time in a weighted ℓ 2-norm. This norm can be de ned to enable complexity reduction (i.e., hyper-reduction) in time, which leads to space-time collocation and space-time GNAT variants of the ST-LSPG method. Advantages of the approach relative to typical spatial-projection-based nonlinear model reduction methods such as Galerkin projection and least-squares Petrov-Galerkin projection include: (1) a reduction of both the spatial and temporal dimensions of the dynamical system, (2) the removal of spurious temporal modes (e.g., unstable growth) from the state space, and (3) error bounds that exhibit slower growth in time. Numerical examples performed on model problems in fluid dynamics demonstrate the ability of the method to generate orders-of-magnitude computational savings relative to spatial-projection-based reduced-order models without sacrificing accuracy.« less
Some aspects of algorithm performance and modeling in transient analysis of structures
NASA Technical Reports Server (NTRS)
Adelman, H. M.; Haftka, R. T.; Robinson, J. C.
1981-01-01
The status of an effort to increase the efficiency of calculating transient temperature fields in complex aerospace vehicle structures is described. The advantages and disadvantages of explicit algorithms with variable time steps, known as the GEAR package, is described. Four test problems, used for evaluating and comparing various algorithms, were selected and finite-element models of the configurations are described. These problems include a space shuttle frame component, an insulated cylinder, a metallic panel for a thermal protection system, and a model of the wing of the space shuttle orbiter. Results generally indicate a preference for implicit over explicit algorithms for solution of transient structural heat transfer problems when the governing equations are stiff (typical of many practical problems such as insulated metal structures).
Additive schemes for certain operator-differential equations
NASA Astrophysics Data System (ADS)
Vabishchevich, P. N.
2010-12-01
Unconditionally stable finite difference schemes for the time approximation of first-order operator-differential systems with self-adjoint operators are constructed. Such systems arise in many applied problems, for example, in connection with nonstationary problems for the system of Stokes (Navier-Stokes) equations. Stability conditions in the corresponding Hilbert spaces for two-level weighted operator-difference schemes are obtained. Additive (splitting) schemes are proposed that involve the solution of simple problems at each time step. The results are used to construct splitting schemes with respect to spatial variables for nonstationary Navier-Stokes equations for incompressible fluid. The capabilities of additive schemes are illustrated using a two-dimensional model problem as an example.
Yurtkuran, Alkın; Emel, Erdal
2014-01-01
The traveling salesman problem with time windows (TSPTW) is a variant of the traveling salesman problem in which each customer should be visited within a given time window. In this paper, we propose an electromagnetism-like algorithm (EMA) that uses a new constraint handling technique to minimize the travel cost in TSPTW problems. The EMA utilizes the attraction-repulsion mechanism between charged particles in a multidimensional space for global optimization. This paper investigates the problem-specific constraint handling capability of the EMA framework using a new variable bounding strategy, in which real-coded particle's boundary constraints associated with the corresponding time windows of customers, is introduced and combined with the penalty approach to eliminate infeasibilities regarding time window violations. The performance of the proposed algorithm and the effectiveness of the constraint handling technique have been studied extensively, comparing it to that of state-of-the-art metaheuristics using several sets of benchmark problems reported in the literature. The results of the numerical experiments show that the EMA generates feasible and near-optimal results within shorter computational times compared to the test algorithms.
Yurtkuran, Alkın
2014-01-01
The traveling salesman problem with time windows (TSPTW) is a variant of the traveling salesman problem in which each customer should be visited within a given time window. In this paper, we propose an electromagnetism-like algorithm (EMA) that uses a new constraint handling technique to minimize the travel cost in TSPTW problems. The EMA utilizes the attraction-repulsion mechanism between charged particles in a multidimensional space for global optimization. This paper investigates the problem-specific constraint handling capability of the EMA framework using a new variable bounding strategy, in which real-coded particle's boundary constraints associated with the corresponding time windows of customers, is introduced and combined with the penalty approach to eliminate infeasibilities regarding time window violations. The performance of the proposed algorithm and the effectiveness of the constraint handling technique have been studied extensively, comparing it to that of state-of-the-art metaheuristics using several sets of benchmark problems reported in the literature. The results of the numerical experiments show that the EMA generates feasible and near-optimal results within shorter computational times compared to the test algorithms. PMID:24723834
An application of computer aided requirements analysis to a real time deep space system
NASA Technical Reports Server (NTRS)
Farny, A. M.; Morris, R. V.; Hartsough, C.; Callender, E. D.; Teichroew, D.; Chikofsky, E.
1981-01-01
The entire procedure of incorporating the requirements and goals of a space flight project into integrated, time ordered sequences of spacecraft commands, is called the uplink process. The Uplink Process Control Task (UPCT) was created to examine the uplink process and determine ways to improve it. The Problem Statement Language/Problem Statement Analyzer (PSL/PSA) designed to assist the designer/analyst/engineer in the preparation of specifications of an information system is used as a supporting tool to aid in the analysis. Attention is given to a definition of the uplink process, the definition of PSL/PSA, the construction of a PSA database, the value of analysis to the study of the uplink process, and the PSL/PSA lessons learned.
Spatial and Temporal Scaling of Thermal Infrared Remote Sensing Data
NASA Technical Reports Server (NTRS)
Quattrochi, Dale A.; Goel, Narendra S.
1995-01-01
Although remote sensing has a central role to play in the acquisition of synoptic data obtained at multiple spatial and temporal scales to facilitate our understanding of local and regional processes as they influence the global climate, the use of thermal infrared (TIR) remote sensing data in this capacity has received only minimal attention. This results from some fundamental challenges that are associated with employing TIR data collected at different space and time scales, either with the same or different sensing systems, and also from other problems that arise in applying a multiple scaled approach to the measurement of surface temperatures. In this paper, we describe some of the more important problems associated with using TIR remote sensing data obtained at different spatial and temporal scales, examine why these problems appear as impediments to using multiple scaled TIR data, and provide some suggestions for future research activities that may address these problems. We elucidate the fundamental concept of scale as it relates to remote sensing and explore how space and time relationships affect TIR data from a problem-dependency perspective. We also describe how linearity and non-linearity observation versus parameter relationships affect the quantitative analysis of TIR data. Some insight is given on how the atmosphere between target and sensor influences the accurate measurement of surface temperatures and how these effects will be compounded in analyzing multiple scaled TIR data. Last, we describe some of the challenges in modeling TIR data obtained at different space and time scales and discuss how multiple scaled TIR data can be used to provide new and important information for measuring and modeling land-atmosphere energy balance processes.
Critical spaces for quasilinear parabolic evolution equations and applications
NASA Astrophysics Data System (ADS)
Prüss, Jan; Simonett, Gieri; Wilke, Mathias
2018-02-01
We present a comprehensive theory of critical spaces for the broad class of quasilinear parabolic evolution equations. The approach is based on maximal Lp-regularity in time-weighted function spaces. It is shown that our notion of critical spaces coincides with the concept of scaling invariant spaces in case that the underlying partial differential equation enjoys a scaling invariance. Applications to the vorticity equations for the Navier-Stokes problem, convection-diffusion equations, the Nernst-Planck-Poisson equations in electro-chemistry, chemotaxis equations, the MHD equations, and some other well-known parabolic equations are given.
Controller Chips Preserve Microprocessor Function
NASA Technical Reports Server (NTRS)
2012-01-01
Above the Atlantic Ocean, off the coast of Brazil, there is a dip in the Earth s surrounding magnetic field called the South Atlantic Anomaly. Here, space radiation can reach into Earth s upper atmosphere to interfere with the functioning of satellites, aircraft, and even the International Space Station. "The South Atlantic Anomaly is a hot spot of radiation that the space station goes through at a certain point in orbit," Miria Finckenor, a physicist at Marshall Space Flight Center, describes, "If there s going to be a problem with the electronics, 90 percent of that time, it is going to be in that spot." Space radiation can cause physical damage to microchips and can actually change the software commands in computers. When high-energy particles penetrate a satellite or other spacecraft, the electrical components can absorb the energy and temporarily switch off. If the energy is high enough, it can cause the device to enter a hung state, which can only be addressed by restarting the system. When space radiation affects the operational status of microprocessors, the occurrence is called single event functional interrupt (SEFI). SEFI happens not only to the computers onboard spacecraft in Earth orbit, but to the computers on spacecraft throughout the solar system. "One of the Mars rovers had this problem in the radiation environment and was rebooting itself several times a day. On one occasion, it rebooted 40 times in one day," Finckenor says. "It s hard to obtain any data when you have to constantly reboot and start over."
Future Visions for Scientific Human Exploration
NASA Technical Reports Server (NTRS)
Garvin, James
2005-01-01
Today, humans explore deep-space locations such as Mars, asteroids, and beyond, vicariously here on Earth, with noteworthy success. However, to achieve the revolutionary breakthroughs that have punctuated the history of science since the dawn of the Space Age has always required humans as "the discoverers," as Daniel Boorstin contends in this book of the same name. During Apollo 17, human explorers on the lunar surface discovered the "genesis rock," orange glass, and humans in space revamped the optically crippled Hubble Space Telescope to enable some of the greatest astronomical discoveries of all time. Science-driven human exploration is about developing the opportunities for such events, perhaps associated with challenging problems such as whether we can identify life beyond Earth within the universe. At issue, however, is how to safely insert humans and the spaceflight systems required to allow humans to operate as they do best in the hostile environment of deep space. The first issue is minimizing the problems associated with human adaptation to the most challenging aspects of deep space space radiation and microgravity (or non-Earth gravity). One solution path is to develop technologies that allow for minimization of the exposure time of people to deep space, as was accomplished in Apollo. For a mission to the planet Mars, this might entail new technological solutions for in-space propulsion that would make possible time-minimized transfers to and from Mars. The problem of rapid, reliable in-space transportation is challenged by the celestial mechanics of moving in space and the so-called "rocket equation." To travel to Mars from Earth in less than the time fuel-minimizing trajectories allow (i.e., Hohmann transfers) requires an exponential increase in the amount of fuel. Thus, month-long transits would require a mass of fuel as large as the dry mass of the ISS, assuming the existence of continuous acceleration engines. This raises the largest technological stumbling block to moving humans on site as deep-space explorers, delivering the masses required for human spaceflight systems to LEO or other Earth orbital vantage points using the existing or projected fleet of Earth-to-orbit (ETO) launch vehicles. Without a return to Saturn V-class boosters or an alternate path, one cannot imagine emplacing the masses that would be required for any deep-space voyage without a prohibitive number of Shuttle-class launches. One futurist solution might involve mass launch systems that could be used to move the consumables, including fuel, water, food, and building materials, to LEO in pieces rather than launching integrated systems. This approach would necessitate the development of robotic assembly and fuel-storage systems in Earth orbit, but could provide for a natural separation of low-value cargo (e.g., fuel, water).
Modular space station phase B extension program master plan
NASA Technical Reports Server (NTRS)
Munsey, E. H.
1971-01-01
The project is defined for design, development, fabrication, test, and pre-mission and mission operations of a shuttle-launched modular space station. The project management approach is described in terms of organization, management requirements, work breakdown structure, schedule, time-phased logic, implementation plans, manpower, and funding. The programmatic and technical problems are identified.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Samet Y. Kadioglu
2011-12-01
We present a computational gas dynamics method based on the Spectral Deferred Corrections (SDC) time integration technique and the Piecewise Parabolic Method (PPM) finite volume method. The PPM framework is used to define edge averaged quantities which are then used to evaluate numerical flux functions. The SDC technique is used to integrate solution in time. This kind of approach was first taken by Anita et al in [17]. However, [17] is problematic when it is implemented to certain shock problems. Here we propose significant improvements to [17]. The method is fourth order (both in space and time) for smooth flows,more » and provides highly resolved discontinuous solutions. We tested the method by solving variety of problems. Results indicate that the fourth order of accuracy in both space and time has been achieved when the flow is smooth. Results also demonstrate the shock capturing ability of the method.« less
NASA Technical Reports Server (NTRS)
Chang, Sin-Chung; Himansu, Ananda; Loh, Ching-Yuen; Wang, Xiao-Yen; Yu, Shang-Tao
2003-01-01
This paper reports on a significant advance in the area of non-reflecting boundary conditions (NRBCs) for unsteady flow computations. As a part of the development of the space-time conservation element and solution element (CE/SE) method, sets of NRBCs for 1D Euler problems are developed without using any characteristics-based techniques. These conditions are much simpler than those commonly reported in the literature, yet so robust that they are applicable to subsonic, transonic and supersonic flows even in the presence of discontinuities. In addition, the straightforward multidimensional extensions of the present 1D NRBCs have been shown numerically to be equally simple and robust. The paper details the theoretical underpinning of these NRBCs, and explains their unique robustness and accuracy in terms of the conservation of space-time fluxes. Some numerical results for an extended Sod's shock-tube problem, illustrating the effectiveness of the present NRBCs are included, together with an associated simple Fortran computer program. As a preliminary to the present development, a review of the basic CE/SE schemes is also included.
Cell transmission model of dynamic assignment for urban rail transit networks.
Xu, Guangming; Zhao, Shuo; Shi, Feng; Zhang, Feilian
2017-01-01
For urban rail transit network, the space-time flow distribution can play an important role in evaluating and optimizing the space-time resource allocation. For obtaining the space-time flow distribution without the restriction of schedules, a dynamic assignment problem is proposed based on the concept of continuous transmission. To solve the dynamic assignment problem, the cell transmission model is built for urban rail transit networks. The priority principle, queuing process, capacity constraints and congestion effects are considered in the cell transmission mechanism. Then an efficient method is designed to solve the shortest path for an urban rail network, which decreases the computing cost for solving the cell transmission model. The instantaneous dynamic user optimal state can be reached with the method of successive average. Many evaluation indexes of passenger flow can be generated, to provide effective support for the optimization of train schedules and the capacity evaluation for urban rail transit network. Finally, the model and its potential application are demonstrated via two numerical experiments using a small-scale network and the Beijing Metro network.
Cultural-based particle swarm for dynamic optimisation problems
NASA Astrophysics Data System (ADS)
Daneshyari, Moayed; Yen, Gary G.
2012-07-01
Many practical optimisation problems are with the existence of uncertainties, among which a significant number belong to the dynamic optimisation problem (DOP) category in which the fitness function changes through time. In this study, we propose the cultural-based particle swarm optimisation (PSO) to solve DOP problems. A cultural framework is adopted incorporating the required information from the PSO into five sections of the belief space, namely situational, temporal, domain, normative and spatial knowledge. The stored information will be adopted to detect the changes in the environment and assists response to the change through a diversity-based repulsion among particles and migration among swarms in the population space, and also helps in selecting the leading particles in three different levels, personal, swarm and global levels. Comparison of the proposed heuristics over several difficult dynamic benchmark problems demonstrates the better or equal performance with respect to most of other selected state-of-the-art dynamic PSO heuristics.
NASA Technical Reports Server (NTRS)
Van Dongen, H. P.; Olofsen, E.; VanHartevelt, J. H.; Kruyt, E. W.; Dinges, D. F. (Principal Investigator)
1999-01-01
Periodogram analysis of unequally spaced time-series, as part of many biological rhythm investigations, is complicated. The mathematical framework is scattered over the literature, and the interpretation of results is often debatable. In this paper, we show that the Lomb-Scargle method is the appropriate tool for periodogram analysis of unequally spaced data. A unique procedure of multiple period searching is derived, facilitating the assessment of the various rhythms that may be present in a time-series. All relevant mathematical and statistical aspects are considered in detail, and much attention is given to the correct interpretation of results. The use of the procedure is illustrated by examples, and problems that may be encountered are discussed. It is argued that, when following the procedure of multiple period searching, we can even benefit from the unequal spacing of a time-series in biological rhythm research.
Dynamics of Pure Shape, Relativity, and the Problem of Time
NASA Astrophysics Data System (ADS)
Barbour, Julian
A new approach to the dynamics of the universe based on work by Ó Murchadha, Foster, Anderson and the author is presented. The only kinematics presupposed is the spatial geometry needed to define configuration spaces in purely relational terms. A new formulation of the relativity principle based on Poincarés analysis of the problem of absolute and relative motion (Machs principle) is given. The entire dynamics is based on shape and nothing else. It leads to much stronger predictions than standard Newtonian theory. For the dynamics of Riemannian 3-geometries on which matter fields also evolve, implementation of the new relativity principle establishes unexpected links between special relativity, general relativity and the gauge principle. They all emerge together as a self-consistent complex from a unified and completely relational approach to dynamics. A connection between time and scale invariance is established. In particular, the representation of general relativity as evolution of the shape of space leads to a unique dynamical definition of simultaneity. This opens up the prospect of a solution of the problem of time in quantum gravity on the basis of a fundamental dynamical principle.
NASA Technical Reports Server (NTRS)
Oeftering, Richard C.; Bradish, Martin A.; Juergens, Jeffrey R.; Lewis, Michael J.
2011-01-01
The NASA Constellation Program is investigating and developing technologies to support human exploration of the Moon and Mars. The Component-Level Electronic-Assembly Repair (CLEAR) task is part of the Supportability Project managed by the Exploration Technology Development Program. CLEAR is aimed at enabling a flight crew to diagnose and repair electronic circuits in space yet minimize logistics spares, equipment, and crew time and training. For insight into actual space repair needs, in early 2008 the project examined the operational experience of the International Space Station (ISS) program. CLEAR examined the ISS on-orbit Problem Reporting and Corrective Action database for electrical and electronic system problems. The ISS has higher than predicted reliability yet, as expected, it has persistent problems. A goal was to identify which on-orbit electrical problems could be resolved by a component-level replacement. A further goal was to identify problems that could benefit from the additional diagnostic and test capability that a component-level repair capability could provide. The study indicated that many problems stem from a small set of root causes that also represent distinct component problems. The study also determined that there are certain recurring problems where the current telemetry instrumentation and built-in tests are unable to completely resolve the problem. As a result, the root cause is listed as unknown. Overall, roughly 42 percent of on-orbit electrical problems on ISS could be addressed with a component-level repair. Furthermore, 63 percent of on-orbit electrical problems on ISS could benefit from additional external diagnostic and test capability. These results indicate that in situ component-level repair in combination with diagnostic and test capability can be expected to increase system availability and reduce logistics. The CLEAR approach can increase the flight crew s ability to act decisively to resolve problems while reducing dependency on Earth-supplied logistics for future Constellation Program missions.
Insight and search in Katona's five-square problem.
Ollinger, Michael; Jones, Gary; Knoblich, Günther
2014-01-01
Insights are often productive outcomes of human thinking. We provide a cognitive model that explains insight problem solving by the interplay of problem space search and representational change, whereby the problem space is constrained or relaxed based on the problem representation. By introducing different experimental conditions that either constrained the initial search space or helped solvers to initiate a representational change, we investigated the interplay of problem space search and representational change in Katona's five-square problem. Testing 168 participants, we demonstrated that independent hints relating to the initial search space and to representational change had little effect on solution rates. However, providing both hints caused a significant increase in solution rates. Our results show the interplay between problem space search and representational change in insight problem solving: The initial problem space can be so large that people fail to encounter impasse, but even when representational change is achieved the resulting problem space can still provide a major obstacle to finding the solution.
Interacting dark sector and the coincidence problem within the scope of LRS Bianchi type I model
NASA Astrophysics Data System (ADS)
Muharlyamov, Ruslan K.; Pankratyeva, Tatiana N.
2018-05-01
It is shown that a suitable interaction between dark energy and dark matter in locally rotationally symmetric (LRS) Bianchi-I space-time can solve the coincidence problem and not contradict the accelerated expansion of present Universe. The interaction parameters are estimated from observational data.
NASA Technical Reports Server (NTRS)
Iida, H. T.
1966-01-01
Computational procedure reduces the numerical effort whenever the method of finite differences is used to solve ablation problems for which the surface recession is large relative to the initial slab thickness. The number of numerical operations required for a given maximum space mesh size is reduced.
Wei, Qinglai; Liu, Derong; Lin, Qiao
In this paper, a novel local value iteration adaptive dynamic programming (ADP) algorithm is developed to solve infinite horizon optimal control problems for discrete-time nonlinear systems. The focuses of this paper are to study admissibility properties and the termination criteria of discrete-time local value iteration ADP algorithms. In the discrete-time local value iteration ADP algorithm, the iterative value functions and the iterative control laws are both updated in a given subset of the state space in each iteration, instead of the whole state space. For the first time, admissibility properties of iterative control laws are analyzed for the local value iteration ADP algorithm. New termination criteria are established, which terminate the iterative local ADP algorithm with an admissible approximate optimal control law. Finally, simulation results are given to illustrate the performance of the developed algorithm.In this paper, a novel local value iteration adaptive dynamic programming (ADP) algorithm is developed to solve infinite horizon optimal control problems for discrete-time nonlinear systems. The focuses of this paper are to study admissibility properties and the termination criteria of discrete-time local value iteration ADP algorithms. In the discrete-time local value iteration ADP algorithm, the iterative value functions and the iterative control laws are both updated in a given subset of the state space in each iteration, instead of the whole state space. For the first time, admissibility properties of iterative control laws are analyzed for the local value iteration ADP algorithm. New termination criteria are established, which terminate the iterative local ADP algorithm with an admissible approximate optimal control law. Finally, simulation results are given to illustrate the performance of the developed algorithm.
Ullah, Sami; Daud, Hanita; Dass, Sarat C; Khan, Habib Nawaz; Khalil, Alamgir
2017-11-06
Ability to detect potential space-time clusters in spatio-temporal data on disease occurrences is necessary for conducting surveillance and implementing disease prevention policies. Most existing techniques use geometrically shaped (circular, elliptical or square) scanning windows to discover disease clusters. In certain situations, where the disease occurrences tend to cluster in very irregularly shaped areas, these algorithms are not feasible in practise for the detection of space-time clusters. To address this problem, a new algorithm is proposed, which uses a co-clustering strategy to detect prospective and retrospective space-time disease clusters with no restriction on shape and size. The proposed method detects space-time disease clusters by tracking the changes in space-time occurrence structure instead of an in-depth search over space. This method was utilised to detect potential clusters in the annual and monthly malaria data in Khyber Pakhtunkhwa Province, Pakistan from 2012 to 2016 visualising the results on a heat map. The results of the annual data analysis showed that the most likely hotspot emerged in three sub-regions in the years 2013-2014. The most likely hotspots in monthly data appeared in the month of July to October in each year and showed a strong periodic trend.
NASA Technical Reports Server (NTRS)
Goldman, Nathan C.
1992-01-01
Space industrialization is confronting space law with problems that are changing old and shaping new legal principles. The return to the Moon, the next logical step beyond the space station, will establish a permanent human presence there. Science and engineering, manufacturing and mining will involve the astronauts in the settlement of the solar system. These pioneers, from many nations, will need a legal, political, and social framework to structure their lives and interactions. International and even domestic space law are only the beginning of this framework. Dispute resolution and simple experience will be needed in order to develop, over time, a new social system for the new regime of space.
q-Space Upsampling Using x-q Space Regularization.
Chen, Geng; Dong, Bin; Zhang, Yong; Shen, Dinggang; Yap, Pew-Thian
2017-09-01
Acquisition time in diffusion MRI increases with the number of diffusion-weighted images that need to be acquired. Particularly in clinical settings, scan time is limited and only a sparse coverage of the vast q -space is possible. In this paper, we show how non-local self-similar information in the x - q space of diffusion MRI data can be harnessed for q -space upsampling. More specifically, we establish the relationships between signal measurements in x - q space using a patch matching mechanism that caters to unstructured data. We then encode these relationships in a graph and use it to regularize an inverse problem associated with recovering a high q -space resolution dataset from its low-resolution counterpart. Experimental results indicate that the high-resolution datasets reconstructed using the proposed method exhibit greater quality, both quantitatively and qualitatively, than those obtained using conventional methods, such as interpolation using spherical radial basis functions (SRBFs).
Global, finite energy, weak solutions for the NLS with rough, time-dependent magnetic potentials
NASA Astrophysics Data System (ADS)
Antonelli, Paolo; Michelangeli, Alessandro; Scandone, Raffaele
2018-04-01
We prove the existence of weak solutions in the space of energy for a class of nonlinear Schrödinger equations in the presence of a external, rough, time-dependent magnetic potential. Under our assumptions, it is not possible to study the problem by means of usual arguments like resolvent techniques or Fourier integral operators, for example. We use a parabolic regularisation, and we solve the approximating Cauchy problem. This is achieved by obtaining suitable smoothing estimates for the dissipative evolution. The total mass and energy bounds allow to extend the solution globally in time. We then infer sufficient compactness properties in order to produce a global-in-time finite energy weak solution to our original problem.
Scalable learning method for feedforward neural networks using minimal-enclosing-ball approximation.
Wang, Jun; Deng, Zhaohong; Luo, Xiaoqing; Jiang, Yizhang; Wang, Shitong
2016-06-01
Training feedforward neural networks (FNNs) is one of the most critical issues in FNNs studies. However, most FNNs training methods cannot be directly applied for very large datasets because they have high computational and space complexity. In order to tackle this problem, the CCMEB (Center-Constrained Minimum Enclosing Ball) problem in hidden feature space of FNN is discussed and a novel learning algorithm called HFSR-GCVM (hidden-feature-space regression using generalized core vector machine) is developed accordingly. In HFSR-GCVM, a novel learning criterion using L2-norm penalty-based ε-insensitive function is formulated and the parameters in the hidden nodes are generated randomly independent of the training sets. Moreover, the learning of parameters in its output layer is proved equivalent to a special CCMEB problem in FNN hidden feature space. As most CCMEB approximation based machine learning algorithms, the proposed HFSR-GCVM training algorithm has the following merits: The maximal training time of the HFSR-GCVM training is linear with the size of training datasets and the maximal space consumption is independent of the size of training datasets. The experiments on regression tasks confirm the above conclusions. Copyright © 2016 Elsevier Ltd. All rights reserved.
An evolutionary strategy based on partial imitation for solving optimization problems
NASA Astrophysics Data System (ADS)
Javarone, Marco Alberto
2016-12-01
In this work we introduce an evolutionary strategy to solve combinatorial optimization tasks, i.e. problems characterized by a discrete search space. In particular, we focus on the Traveling Salesman Problem (TSP), i.e. a famous problem whose search space grows exponentially, increasing the number of cities, up to becoming NP-hard. The solutions of the TSP can be codified by arrays of cities, and can be evaluated by fitness, computed according to a cost function (e.g. the length of a path). Our method is based on the evolution of an agent population by means of an imitative mechanism, we define 'partial imitation'. In particular, agents receive a random solution and then, interacting among themselves, may imitate the solutions of agents with a higher fitness. Since the imitation mechanism is only partial, agents copy only one entry (randomly chosen) of another array (i.e. solution). In doing so, the population converges towards a shared solution, behaving like a spin system undergoing a cooling process, i.e. driven towards an ordered phase. We highlight that the adopted 'partial imitation' mechanism allows the population to generate solutions over time, before reaching the final equilibrium. Results of numerical simulations show that our method is able to find, in a finite time, both optimal and suboptimal solutions, depending on the size of the considered search space.
Real-time control for manufacturing space shuttle main engines: Work in progress
NASA Technical Reports Server (NTRS)
Ruokangas, Corinne C.
1988-01-01
During the manufacture of space-based assemblies such as Space Shuttle Main Engines, flexibility is required due to the high-cost and low-volume nature of the end products. Various systems have been developed pursuing the goal of adaptive, flexible manufacturing for several space applications, including an Advanced Robotic Welding System for the manufacture of complex components of the Space Shuttle Main Engines. The Advanced Robotic Welding System (AROWS) is an on-going joint effort, funded by NASA, between NASA/Marshall Space Flight Center, and two divisions of Rockwell International: Rocketdyne and the Science Center. AROWS includes two levels of flexible control of both motion and process parameters: Off-line programming using both geometric and weld-process data bases, and real-time control incorporating multiple sensors during weld execution. Both control systems were implemented using conventional hardware and software architectures. The feasibility of enhancing the real-time control system using the problem-solving architecture of Schemer is investigated and described.
1984-02-01
I . . . . . . An Introduction to Geometric Programming Patrick D. Allen and David W. Baker . . . . . . , . . . . . . . Space and Time...Zarwyn, US-Army Electronics R & D Comhiand GEOMETRIC PROGRAMING SPACE AND TIFFE ANALYSIS IN DYNAMIC PROGRAMING ALGORITHMS Renne..tf Stizti, AkeanXa...physical and parameter space can be connected by asymptotic matching. The purpose of the asymptotic analysis is to define the simplest problems
NASA Astrophysics Data System (ADS)
Deswal, Sunita; Kalkal, Kapil Kumar; Sheoran, Sandeep Singh
2016-09-01
A mathematical model of fractional order two-temperature generalized thermoelasticity with diffusion and initial stress is proposed to analyze the transient wave phenomenon in an infinite thermoelastic half-space. The governing equations are derived in cylindrical coordinates for a two dimensional axi-symmetric problem. The analytical solution is procured by employing the Laplace and Hankel transforms for time and space variables respectively. The solutions are investigated in detail for a time dependent heat source. By using numerical inversion method of integral transforms, we obtain the solutions for displacement, stress, temperature and diffusion fields in physical domain. Computations are carried out for copper material and displayed graphically. The effect of fractional order parameter, two-temperature parameter, diffusion, initial stress and time on the different thermoelastic and diffusion fields is analyzed on the basis of analytical and numerical results. Some special cases have also been deduced from the present investigation.
Optical solver of combinatorial problems: nanotechnological approach.
Cohen, Eyal; Dolev, Shlomi; Frenkel, Sergey; Kryzhanovsky, Boris; Palagushkin, Alexandr; Rosenblit, Michael; Zakharov, Victor
2013-09-01
We present an optical computing system to solve NP-hard problems. As nano-optical computing is a promising venue for the next generation of computers performing parallel computations, we investigate the application of submicron, or even subwavelength, computing device designs. The system utilizes a setup of exponential sized masks with exponential space complexity produced in polynomial time preprocessing. The masks are later used to solve the problem in polynomial time. The size of the masks is reduced to nanoscaled density. Simulations were done to choose a proper design, and actual implementations show the feasibility of such a system.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pourmatin, Hossein, E-mail: mpourmat@andrew.cmu.edu; Dayal, Kaushik, E-mail: kaushik@cmu.edu
2016-10-15
Graphical abstract: - Abstract: We consider the scattering of incident plane-wave electrons from a defect in a crystal modeled by the time-harmonic Schrödinger equation. While the defect potential is localized, the far-field potential is periodic, unlike standard free-space scattering problems. Previous work on the Schrödinger equation has been almost entirely in free-space conditions; a few works on crystals have been in one-dimension. We construct absorbing boundary conditions for this problem using perfectly matched layers in a tight-binding formulation. Using the example of a point defect in graphene, we examine the efficiency and convergence of the proposed absorbing boundary condition.
The two-dimensional Stefan problem with slightly varying heat flux
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gammon, J.; Howarth, J.A.
1995-09-01
The authors solve the two-dimensional stefan problem of solidification in a half-space, where the heat flux at the wall is a slightly varying function of positioning along the wall, by means of a large Stefan number approximation (which turns out to be equivalent to a small time solution), and then by means of the Heat Balance Integral Method, which is valid for all time, and which agrees with the large Stefan number solution for small times. A representative solution is given for a particular form of the heat flux perturbation.
NASA Astrophysics Data System (ADS)
Magee, Daniel J.; Niemeyer, Kyle E.
2018-03-01
The expedient design of precision components in aerospace and other high-tech industries requires simulations of physical phenomena often described by partial differential equations (PDEs) without exact solutions. Modern design problems require simulations with a level of resolution difficult to achieve in reasonable amounts of time-even in effectively parallelized solvers. Though the scale of the problem relative to available computing power is the greatest impediment to accelerating these applications, significant performance gains can be achieved through careful attention to the details of memory communication and access. The swept time-space decomposition rule reduces communication between sub-domains by exhausting the domain of influence before communicating boundary values. Here we present a GPU implementation of the swept rule, which modifies the algorithm for improved performance on this processing architecture by prioritizing use of private (shared) memory, avoiding interblock communication, and overwriting unnecessary values. It shows significant improvement in the execution time of finite-difference solvers for one-dimensional unsteady PDEs, producing speedups of 2 - 9 × for a range of problem sizes, respectively, compared with simple GPU versions and 7 - 300 × compared with parallel CPU versions. However, for a more sophisticated one-dimensional system of equations discretized with a second-order finite-volume scheme, the swept rule performs 1.2 - 1.9 × worse than a standard implementation for all problem sizes.
Parameter estimation in nonlinear distributed systems - Approximation theory and convergence results
NASA Technical Reports Server (NTRS)
Banks, H. T.; Reich, Simeon; Rosen, I. G.
1988-01-01
An abstract approximation framework and convergence theory is described for Galerkin approximations applied to inverse problems involving nonlinear distributed parameter systems. Parameter estimation problems are considered and formulated as the minimization of a least-squares-like performance index over a compact admissible parameter set subject to state constraints given by an inhomogeneous nonlinear distributed system. The theory applies to systems whose dynamics can be described by either time-independent or nonstationary strongly maximal monotonic operators defined on a reflexive Banach space which is densely and continuously embedded in a Hilbert space. It is demonstrated that if readily verifiable conditions on the system's dependence on the unknown parameters are satisfied, and the usual Galerkin approximation assumption holds, then solutions to the approximating problems exist and approximate a solution to the original infinite-dimensional identification problem.
NASA Astrophysics Data System (ADS)
Roverso, Davide
2003-08-01
Many-class learning is the problem of training a classifier to discriminate among a large number of target classes. Together with the problem of dealing with high-dimensional patterns (i.e. a high-dimensional input space), the many class problem (i.e. a high-dimensional output space) is a major obstacle to be faced when scaling-up classifier systems and algorithms from small pilot applications to large full-scale applications. The Autonomous Recursive Task Decomposition (ARTD) algorithm is here proposed as a solution to the problem of many-class learning. Example applications of ARTD to neural classifier training are also presented. In these examples, improvements in training time are shown to range from 4-fold to more than 30-fold in pattern classification tasks of both static and dynamic character.
Contact in an expanding universe: an instructive exercise in dynamic geometry
NASA Astrophysics Data System (ADS)
Zimmerman, Seth
2010-11-01
The particular problem solved in this paper is that of calculating the time required to overtake a distant object receding under cosmic expansion, and the speed at which that object is passed. This is a rarely investigated problem leading to some interesting apparent paradoxes. We employ the problem to promote a deeper understanding of the dynamic geometry behind the expansion of space in three eras, especially for physics undergraduates. We do not utilize the standard cosmological formulae, but deliberately take a simpler approach, comprehensible to any student comfortable with differentials. This should offer an intuitive preparation for later courses in general relativity. The paper can be read straight through, or offered to a class in segments as problems to investigate. The overall intention is to leave students with a more tangible grasp of expanding space.
Steps Toward Optimal Competitive Scheduling
NASA Technical Reports Server (NTRS)
Frank, Jeremy; Crawford, James; Khatib, Lina; Brafman, Ronen
2006-01-01
This paper is concerned with the problem of allocating a unit capacity resource to multiple users within a pre-defined time period. The resource is indivisible, so that at most one user can use it at each time instance. However, different users may use it at different times. The users have independent, se@sh preferences for when and for how long they are allocated this resource. Thus, they value different resource access durations differently, and they value different time slots differently. We seek an optimal allocation schedule for this resource. This problem arises in many institutional settings where, e.g., different departments, agencies, or personal, compete for a single resource. We are particularly motivated by the problem of scheduling NASA's Deep Space Satellite Network (DSN) among different users within NASA. Access to DSN is needed for transmitting data from various space missions to Earth. Each mission has different needs for DSN time, depending on satellite and planetary orbits. Typically, the DSN is over-subscribed, in that not all missions will be allocated as much time as they want. This leads to various inefficiencies - missions spend much time and resource lobbying for their time, often exaggerating their needs. NASA, on the other hand, would like to make optimal use of this resource, ensuring that the good for NASA is maximized. This raises the thorny problem of how to measure the utility to NASA of each allocation. In the typical case, it is difficult for the central agency, NASA in our case, to assess the value of each interval to each user - this is really only known to the users who understand their needs. Thus, our problem is more precisely formulated as follows: find an allocation schedule for the resource that maximizes the sum of users preferences, when the preference values are private information of the users. We bypass this problem by making the assumptions that one can assign money to customers. This assumption is reasonable; a committee is usually in charge of deciding the priority of each mission competing for access to the DSN within a time period while scheduling. Instead, we can assume that the committee assigns a budget to each mission.This paper is concerned with the problem of allocating a unit capacity resource to multiple users within a pre-defined time period. The resource is indivisible, so that at most one user can use it at each time instance. However, different users may use it at different times. The users have independent, se@sh preferences for when and for how long they are allocated this resource. Thus, they value different resource access durations differently, and they value different time slots differently. We seek an optimal allocation schedule for this resource. This problem arises in many institutional settings where, e.g., different departments, agencies, or personal, compete for a single resource. We are particularly motivated by the problem of scheduling NASA's Deep Space Satellite Network (DSN) among different users within NASA. Access to DSN is needed for transmitting data from various space missions to Earth. Each mission has different needs for DSN time, depending on satellite and planetary orbits. Typically, the DSN is over-subscribed, in that not all missions will be allocated as much time as they want. This leads to various inefficiencies - missions spend much time and resource lobbying for their time, often exaggerating their needs. NASA, on the other hand, would like to make optimal use of this resource, ensuring that the good for NASA is maximized. This raises the thorny problem of how to measure the utility to NASA of each allocation. In the typical case, it is difficult for the central agency, NASA in our case, to assess the value of each interval to each user - this is really only known to the users who understand their needs. Thus, our problem is more precisely formulated as follows: find an allocation schedule for the resource that maximizes the sum ofsers preferences, when the preference values are private information of the users. We bypass this problem by making the assumptions that one can assign money to customers. This assumption is reasonable; a committee is usually in charge of deciding the priority of each mission competing for access to the DSN within a time period while scheduling. Instead, we can assume that the committee assigns a budget to each mission.
Error analysis of multipoint flux domain decomposition methods for evolutionary diffusion problems
NASA Astrophysics Data System (ADS)
Arrarás, A.; Portero, L.; Yotov, I.
2014-01-01
We study space and time discretizations for mixed formulations of parabolic problems. The spatial approximation is based on the multipoint flux mixed finite element method, which reduces to an efficient cell-centered pressure system on general grids, including triangles, quadrilaterals, tetrahedra, and hexahedra. The time integration is performed by using a domain decomposition time-splitting technique combined with multiterm fractional step diagonally implicit Runge-Kutta methods. The resulting scheme is unconditionally stable and computationally efficient, as it reduces the global system to a collection of uncoupled subdomain problems that can be solved in parallel without the need for Schwarz-type iteration. Convergence analysis for both the semidiscrete and fully discrete schemes is presented.
Solid rocket motor certification to meet space shuttle requirements from challenge to achievement
NASA Technical Reports Server (NTRS)
Miller, J. Q.; Kilminster, J. C.
1985-01-01
Three solid rocket motor (SRM) design requirements for the Space Shuttle were discussed. No existing solid rocket motor experience was available for the requirement for a thrust-time trace, twenty uses for the principle hardware, and a moveable nozzle with an 8 deg. omnivaxial vectoring capability. The solutions to these problems are presented.
Short-term capture of the Earth-Moon system
NASA Astrophysics Data System (ADS)
Qi, Yi; de Ruiter, Anton
2018-06-01
In this paper, the short-term capture (STC) of an asteroid in the Earth-Moon system is proposed and investigated. First, the space condition of STC is analysed and five subsets of the feasible region are defined and discussed. Then, the time condition of STC is studied by parameter scanning in the Sun-Earth-Moon-asteroid restricted four-body problem. Numerical results indicate that there is a clear association between the distributions of the time probability of STC and the five subsets. Next, the influence of the Jacobi constant on STC is examined using the space and time probabilities of STC. Combining the space and time probabilities of STC, we propose a STC index to evaluate the probability of STC comprehensively. Finally, three potential STC asteroids are found and analysed.
NASA Astrophysics Data System (ADS)
Wang, Lusheng; Yang, Yong; Lin, Guohui
Finding the closest object for a query in a database is a classical problem in computer science. For some modern biological applications, computing the similarity between two objects might be very time consuming. For example, it takes a long time to compute the edit distance between two whole chromosomes and the alignment cost of two 3D protein structures. In this paper, we study the nearest neighbor search problem in metric space, where the pair-wise distance between two objects in the database is known and we want to minimize the number of distances computed on-line between the query and objects in the database in order to find the closest object. We have designed two randomized approaches for indexing metric space databases, where objects are purely described by their distances with each other. Analysis and experiments show that our approaches only need to compute O(logn) objects in order to find the closest object, where n is the total number of objects in the database.
Biomedical program at Space Biospheres Ventures
NASA Technical Reports Server (NTRS)
Walford, Roy
1990-01-01
There are many similarities and some important differences between potential health problems of Biosphere 2 and those of which might be anticipated for a space station or a major outpost on Mars. The demands of time, expense, and equipment would not readily allow medical evacuation from deep space for a serious illness or major trauma, whereas personnel can easily be evacuated from Biosphere 2 if necessary. Treatment facilities can be somewhat less inclusive, since distance would not compel the undertaking of heroic measures or highly complicated surgical procedures on site, and with personnel not fully trained for these procedures. The similarities are given between medical requirements of Biosphere 2 and the complex closed ecological systems of biospheres in space or on Mars. The major problems common to all these would seem to be trauma, infection, and toxicity. It is planned that minor and moderate degrees of trauma, including debridement and suturing of wounds, x ray study of fractures, will be done within Biosphere 2. Bacteriologic and fungal infections, and possibly allergies to pollen or spores are expected to be the commonest medical problem within Biosphere 2.
Method for using global optimization to the estimation of surface-consistent residual statics
Reister, David B.; Barhen, Jacob; Oblow, Edward M.
2001-01-01
An efficient method for generating residual statics corrections to compensate for surface-consistent static time shifts in stacked seismic traces. The method includes a step of framing the residual static corrections as a global optimization problem in a parameter space. The method also includes decoupling the global optimization problem involving all seismic traces into several one-dimensional problems. The method further utilizes a Stochastic Pijavskij Tunneling search to eliminate regions in the parameter space where a global minimum is unlikely to exist so that the global minimum may be quickly discovered. The method finds the residual statics corrections by maximizing the total stack power. The stack power is a measure of seismic energy transferred from energy sources to receivers.
Microbiological testing of Skylab foods.
NASA Technical Reports Server (NTRS)
Heidelbaugh, N. D.; Mcqueen, J. L.; Rowley, D. B.; Powers , E. M.; Bourland, C. T.
1973-01-01
Review of some of the unique food microbiology problems and problem-generating circumstances the Skylab manned space flight program involves. The situations these problems arise from include: extended storage times, variations in storage temperatures, no opportunity to resupply or change foods after launch of the Skylab Workshop, first use of frozen foods in space, first use of a food-warming device in weightlessness, relatively small size of production lots requiring statistically valid sampling plans, and use of food as an accurately controlled part in a set of sophisticated life science experiments. Consideration of all of these situations produced the need for definite microbiological tests and test limits. These tests are described along with the rationale for their selection. Reported test results show good compliance with the test limits.
Deuterium microbomb rocket propulsion
NASA Astrophysics Data System (ADS)
Winterberg, F.
2010-01-01
Large scale manned space flight within the solar system is still confronted with the solution of two problems: (1) A propulsion system to transport large payloads with short transit times between different planetary orbits. (2) A cost effective lifting of large payloads into earth orbit. For the solution of the first problem a deuterium fusion bomb propulsion system is proposed where a thermonuclear detonation wave is ignited in a small cylindrical assembly of deuterium with a gigavolt-multimegaampere proton beam, drawn from the magnetically insulated spacecraft acting in the ultrahigh vacuum of space as a gigavolt capacitor. For the solution of the second problem, the ignition is done by argon ion lasers driven by high explosives, with the lasers destroyed in the fusion explosion and becoming part of the exhaust.
AX-5 space suit bearing torque investigation
NASA Technical Reports Server (NTRS)
Loewenthal, Stuart; Vykukal, Vic; Mackendrick, Robert; Culbertson, Philip, Jr.
1990-01-01
The symptoms and eventual resolution of a torque increase problem occurring with ball bearings in the joints of the AX-5 space suit are described. Starting torques that rose 5 to 10 times initial levels were observed in crew evaluation tests of the suit in a zero-g water tank. This bearing problem was identified as a blocking torque anomaly, observed previously in oscillatory gimbal bearings. A large matrix of lubricants, ball separator designs and materials were evaluated. None of these combinations showed sufficient tolerance to lubricant washout when repeatedly cycled in water. The problem was resolved by retrofitting a pressure compensated, water exclusion seal to the outboard side of the bearing cavity. The symptoms and possible remedies to blocking are discussed.
[A review on the urban green space cooling effect based on field measurement of air temperature].
Liu, Feng Feng; Yan, Wei Jiao; Kong, Fan Hua; Yin, Hai Wei; Ban, Yu Long; Xu, Wen Bin
2017-04-18
With the development of urbanization, the effect of urban heat island has become increasingly evident. As an essential component of the urban natural landscapes, urban green space plays an important role in mitigating the effect of urban heat island. However, facing the rapid urbanization and changing environment, how to rationally plan and design the green space and realize its best cooling effect which can improve the urban environment and microclimate is still an urgent problem to be solved. So there is a strong need for mulitiscale researches on the cooling effect of urban green space. This paper systematically gave a review on the cooling effect of urban green space based on field measurement of air temperature, the main factors that influenced the cooling effect of green space were explored from three aspects including the area and shape characteristics of urban green space, the structure characteristics of vegetation and the external factors which affected the cooling effect, and the characteristics of the cooling effect of the green space were summarized from the aspect of time variation and distance decay. Then, the main problems and future research prospects of urban green space cooling effect were put forward.
NASA Astrophysics Data System (ADS)
Melas, Evangelos
2017-07-01
The original Bondi-Metzner-Sachs (BMS) group B is the common asymptotic symmetry group of all asymptotically flat Lorentzian radiating 4-dim space-times. As such, B is the best candidate for the universal symmetry group of General Relativity (G.R.). In 1973, with this motivation, McCarthy classified all relativistic B-invariant systems in terms of strongly continuous irreducible unitary representations (IRS) of B. Here we introduce the analogue B(2, 1) of the BMS group B in 3 space-time dimensions. B(2, 1) itself admits thirty-four analogues both real in all signatures and in complex space-times. In order to find the IRS of both B(2, 1) and its analogues, we need to extend Wigner-Mackey's theory of induced representations. The necessary extension is described and is reduced to the solution of three problems. These problems are solved in the case where B(2, 1) and its analogues are equipped with the Hilbert topology. The extended theory is necessary in order to construct the IRS of both B and its analogues in any number d of space-time dimensions, d ≥3 , and also in order to construct the IRS of their supersymmetric counterparts. We use the extended theory to obtain the necessary data in order to construct the IRS of B(2, 1). The main results of the representation theory are as follows: The IRS are induced from "little groups" which are compact. The finite "little groups" are cyclic groups of even order. The inducing construction is exhaustive notwithstanding the fact that B(2, 1) is not locally compact in the employed Hilbert topology.
Deep Learning for Flow Sculpting: Insights into Efficient Learning using Scientific Simulation Data
Stoecklein, Daniel; Lore, Kin Gwn; Davies, Michael; Sarkar, Soumik; Ganapathysubramanian, Baskar
2017-01-01
A new technique for shaping microfluid flow, known as flow sculpting, offers an unprecedented level of passive fluid flow control, with potential breakthrough applications in advancing manufacturing, biology, and chemistry research at the microscale. However, efficiently solving the inverse problem of designing a flow sculpting device for a desired fluid flow shape remains a challenge. Current approaches struggle with the many-to-one design space, requiring substantial user interaction and the necessity of building intuition, all of which are time and resource intensive. Deep learning has emerged as an efficient function approximation technique for high-dimensional spaces, and presents a fast solution to the inverse problem, yet the science of its implementation in similarly defined problems remains largely unexplored. We propose that deep learning methods can completely outpace current approaches for scientific inverse problems while delivering comparable designs. To this end, we show how intelligent sampling of the design space inputs can make deep learning methods more competitive in accuracy, while illustrating their generalization capability to out-of-sample predictions. PMID:28402332
Topological defects in alternative theories to cosmic inflation and string cosmology
NASA Astrophysics Data System (ADS)
Alexander, Stephon H. S.
The physics of the Early Universe is described in terms of the inflationary paradigm, which is based on a marriage between Einstein's general theory of relativity minimally coupled to quantum field theory. Inflation was posed to solve some of the outstanding problems of the Standard Big Bang Cosmology (SBB) such as the horizon, formation of structure and monopole problems. Despite its observational and theoretical successes, inflation is plagued with fine tuning and initial singularity problems. On the other hand, superstring/M theory, a theory of quantum gravity, possesses symmetries which naturally avoid space-time singularities. This thesis investigates alternative theories to cosmic inflation for solving the initial singularity, horizon and monopole problems, making use of topological defects. It was proposed by Dvali, Liu and Vaschaspati that the monopole problem can be solved without inflation if domain walls "sweep" up the monopoles in the early universe, thus reducing their number density significantly. Necessary for this mechanism to work is the presence of an attractive force between the monopole and the domain wall as well as a channel for the monopole's unwinding. We show numerically and analytically in two field theory models that for global defects the attraction is a universal result but the unwinding is model specific. The second part of this thesis investigates a string/M theory inspired model for solving the horizon problem. It was proposed by Moffat, Albrecht and Magueijo that the horizon problem is solved with a "phase transition" associated with a varying speed of light before the surface of last scattering. We provide a string/M theory mechanism based on assuming that our space-time is a D-3 brane probing a bulk supergravity black hole bulk background. This mechanism provides the necessary time variation of the velocity of light to solve the horizon problem. We suggest a mechanism which stablilizes the speed of light on the D-3 brane. We finally address the cosmological initial singularity problem using the target space duality inherent in string/M theory. It was suggested by Brandenberger and Vafa that superstring theory can solve the singularity problem and in addition explain why only three spatial dimensions can become large. We show that under specific conditions this mechanism still persists when including the effects of D-branes.
Unified space--time trigonometry and its applications to relativistic kinematics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jaccarini, A.
1973-06-15
A geometrical approach to relativistic kinematics is presented. Owing to a unified space-time trigonometry, the spherical trigonometry formalism may be used to describe and study the kinematics of any collision process. Lorentz transformations may thus lie treated as purely geometrical problems. A different way to define a unified trigonometry is also proposed, which is based on the spinor representation of the Lorentz group. This leads to a different and more general formalism than the former one. (auth)
Long time existence from interior gluing
NASA Astrophysics Data System (ADS)
Chruściel, Piotr T.
2017-07-01
We prove completeness-to-the-future of null hypersurfaces emanating outwards from large spheres, in vacuum space-times evolving from general asymptotically flat data with well-defined energy-momentum. The proof uses scaling and a gluing construction to reduce the problem to Bieri’s stability theorem.
A quasi-Lagrangian finite element method for the Navier-Stokes equations in a time-dependent domain
NASA Astrophysics Data System (ADS)
Lozovskiy, Alexander; Olshanskii, Maxim A.; Vassilevski, Yuri V.
2018-05-01
The paper develops a finite element method for the Navier-Stokes equations of incompressible viscous fluid in a time-dependent domain. The method builds on a quasi-Lagrangian formulation of the problem. The paper provides stability and convergence analysis of the fully discrete (finite-difference in time and finite-element in space) method. The analysis does not assume any CFL time-step restriction, it rather needs mild conditions of the form $\\Delta t\\le C$, where $C$ depends only on problem data, and $h^{2m_u+2}\\le c\\,\\Delta t$, $m_u$ is polynomial degree of velocity finite element space. Both conditions result from a numerical treatment of practically important non-homogeneous boundary conditions. The theoretically predicted convergence rate is confirmed by a set of numerical experiments. Further we apply the method to simulate a flow in a simplified model of the left ventricle of a human heart, where the ventricle wall dynamics is reconstructed from a sequence of contrast enhanced Computed Tomography images.
Solving the MHD equations by the space time conservation element and solution element method
NASA Astrophysics Data System (ADS)
Zhang, Moujin; John Yu, S.-T.; Henry Lin, S.-C.; Chang, Sin-Chung; Blankson, Isaiah
2006-05-01
We apply the space-time conservation element and solution element (CESE) method to solve the ideal MHD equations with special emphasis on satisfying the divergence free constraint of magnetic field, i.e., ∇ · B = 0. In the setting of the CESE method, four approaches are employed: (i) the original CESE method without any additional treatment, (ii) a simple corrector procedure to update the spatial derivatives of magnetic field B after each time marching step to enforce ∇ · B = 0 at all mesh nodes, (iii) a constraint-transport method by using a special staggered mesh to calculate magnetic field B, and (iv) the projection method by solving a Poisson solver after each time marching step. To demonstrate the capabilities of these methods, two benchmark MHD flows are calculated: (i) a rotated one-dimensional MHD shock tube problem and (ii) a MHD vortex problem. The results show no differences between different approaches and all results compare favorably with previously reported data.
NASA Astrophysics Data System (ADS)
Wang, Hongyan
2017-04-01
This paper addresses the waveform optimization problem for improving the detection performance of multi-input multioutput (MIMO) orthogonal frequency division multiplexing (OFDM) radar-based space-time adaptive processing (STAP) in the complex environment. By maximizing the output signal-to-interference-and-noise-ratio (SINR) criterion, the waveform optimization problem for improving the detection performance of STAP, which is subjected to the constant modulus constraint, is derived. To tackle the resultant nonlinear and complicated optimization issue, a diagonal loading-based method is proposed to reformulate the issue as a semidefinite programming one; thereby, this problem can be solved very efficiently. In what follows, the optimized waveform can be obtained to maximize the output SINR of MIMO-OFDM such that the detection performance of STAP can be improved. The simulation results show that the proposed method can improve the output SINR detection performance considerably as compared with that of uncorrelated waveforms and the existing MIMO-based STAP method.
NASA Astrophysics Data System (ADS)
Lee, H.; Seo, D.; McKee, P.; Corby, R.
2009-12-01
One of the large challenges in data assimilation (DA) into distributed hydrologic models is to reduce the large degrees of freedom involved in the inverse problem to avoid overfitting. To assess the sensitivity of the performance of DA to the dimensionality of the inverse problem, we design and carry out real-world experiments in which the control vector in variational DA (VAR) is solved at different scales in space and time, e.g., lumped, semi-distributed, and fully-distributed in space, and hourly, 6 hourly, etc., in time. The size of the control vector is related to the degrees of freedom in the inverse problem. For the assessment, we use the prototype 4-dimenational variational data assimilator (4DVAR) that assimilates streamflow, precipitation and potential evaporation data into the NWS Hydrology Laboratory’s Research Distributed Hydrologic Model (HL-RDHM). In this talk, we present the initial results for a number of basins in Oklahoma and Texas.
Path Flow Estimation Using Time Varying Coefficient State Space Model
NASA Astrophysics Data System (ADS)
Jou, Yow-Jen; Lan, Chien-Lun
2009-08-01
The dynamic path flow information is very crucial in the field of transportation operation and management, i.e., dynamic traffic assignment, scheduling plan, and signal timing. Time-dependent path information, which is important in many aspects, is nearly impossible to be obtained. Consequently, researchers have been seeking estimation methods for deriving valuable path flow information from less expensive traffic data, primarily link traffic counts of surveillance systems. This investigation considers a path flow estimation problem involving the time varying coefficient state space model, Gibbs sampler, and Kalman filter. Numerical examples with part of a real network of the Taipei Mass Rapid Transit with real O-D matrices is demonstrated to address the accuracy of proposed model. Results of this study show that this time-varying coefficient state space model is very effective in the estimation of path flow compared to time-invariant model.
Dynamic State Estimation of Terrestrial and Solar Plasmas
NASA Astrophysics Data System (ADS)
Kamalabadi, Farzad
A pervasive problem in virtually all branches of space science is the estimation of multi-dimensional state parameters of a dynamical system from a collection of indirect, often incomplete, and imprecise measurements. Subsequent scientific inference is predicated on rigorous analysis, interpretation, and understanding of physical observations and on the reliability of the associated quantitative statistical bounds and performance characteristics of the algorithms used. In this work, we focus on these dynamic state estimation problems and illustrate their importance in the context of two timely activities in space remote sensing. First, we discuss the estimation of multi-dimensional ionospheric state parameters from UV spectral imaging measurements anticipated to be acquired the recently selected NASA Heliophysics mission, Ionospheric Connection Explorer (ICON). Next, we illustrate that similar state-space formulations provide the means for the estimation of 3D, time-dependent densities and temperatures in the solar corona from a series of white-light and EUV measurements. We demonstrate that, while a general framework for the stochastic formulation of the state estimation problem is suited for systematic inference of the parameters of a hidden Markov process, several challenges must be addressed in the assimilation of an increasing volume and diversity of space observations. These challenges are: (1) the computational tractability when faced with voluminous and multimodal data, (2) the inherent limitations of the underlying models which assume, often incorrectly, linear dynamics and Gaussian noise, and (3) the unavailability or inaccuracy of transition probabilities and noise statistics. We argue that pursuing answers to these questions necessitates cross-disciplinary research that enables progress toward systematically reconciling observational and theoretical understanding of the space environment.
Periodic trim solutions with hp-version finite elements in time
NASA Technical Reports Server (NTRS)
Peters, David A.; Hou, Lin-Jun
1990-01-01
Finite elements in time as an alternative strategy for rotorcraft trim problems are studied. The research treats linear flap and linearized flap-lag response both for quasi-trim and trim cases. The connection between Fourier series analysis and hp-finite elements for periodic a problem is also examined. It is proved that Fourier series is a special case of space-time finite elements in which one element is used with a strong displacement formulation. Comparisons are made with respect to accuracy among Fourier analysis, displacement methods, and mixed methods over a variety parameters. The hp trade-off is studied for the periodic trim problem to provide an optimum step size and order of polynomial for a given error criteria. It is found that finite elements in time can outperform Fourier analysis for periodic problems, and for some given error criteria. The mixed method provides better results than does the displacement method.
Individuation in Quantum Mechanics and Space-Time
NASA Astrophysics Data System (ADS)
Jaeger, Gregg
2010-10-01
Two physical approaches—as distinct, under the classification of Mittelstaedt, from formal approaches—to the problem of individuation of quantum objects are considered, one formulated in spatiotemporal terms and one in quantum mechanical terms. The spatiotemporal approach itself has two forms: one attributed to Einstein and based on the ontology of space-time points, and the other proposed by Howard and based on intersections of world lines. The quantum mechanical approach is also provided here in two forms, one based on interference and another based on a new Quantum Principle of Individuation (QPI). It is argued that the space-time approach to individuation fails and that the quantum approach offers several advantages over it, including consistency with Leibniz’s Principle of Identity of Indiscernibles.
Position space analysis of the AdS (in)stability problem
NASA Astrophysics Data System (ADS)
Dimitrakopoulos, Fotios V.; Freivogel, Ben; Lippert, Matthew; Yang, I.-Sheng
2015-08-01
We investigate whether arbitrarily small perturbations in global AdS space are generically unstable and collapse into black holes on the time scale set by gravitational interactions. We argue that current evidence, combined with our analysis, strongly suggests that a set of nonzero measure in the space of initial conditions does not collapse on this time scale. We perform an analysis in position space to study this puzzle, and our formalism allows us to directly study the vanishing-amplitude limit. We show that gravitational self-interaction leads to tidal deformations which are equally likely to focus or defocus energy, and we sketch the phase diagram accordingly. We also clarify the connection between gravitational evolution in global AdS and holographic thermalization.
A lifelong learning hyper-heuristic method for bin packing.
Sim, Kevin; Hart, Emma; Paechter, Ben
2015-01-01
We describe a novel hyper-heuristic system that continuously learns over time to solve a combinatorial optimisation problem. The system continuously generates new heuristics and samples problems from its environment; and representative problems and heuristics are incorporated into a self-sustaining network of interacting entities inspired by methods in artificial immune systems. The network is plastic in both its structure and content, leading to the following properties: it exploits existing knowledge captured in the network to rapidly produce solutions; it can adapt to new problems with widely differing characteristics; and it is capable of generalising over the problem space. The system is tested on a large corpus of 3,968 new instances of 1D bin-packing problems as well as on 1,370 existing problems from the literature; it shows excellent performance in terms of the quality of solutions obtained across the datasets and in adapting to dynamically changing sets of problem instances compared to previous approaches. As the network self-adapts to sustain a minimal repertoire of both problems and heuristics that form a representative map of the problem space, the system is further shown to be computationally efficient and therefore scalable.
NASA Technical Reports Server (NTRS)
Rodgers, E. B.; Seale, D. B.; Boraas, M. E.; Sommer, C. V.
1989-01-01
The probable sources and implications of microbial contamination on the proposed Space Station are discussed. Because of the limited availability of material, facilities and time on the Space Station, we are exploring the feasibility of replacing traditional incubation methods for assessing microbial contamination with rapid, automated methods. Some possibilities include: ATP measurement, microscopy and telecommunications, and molecular techniques such as DNA probes or monoclonal antibodies. Some of the important ecological factors that could alter microbes in space include microgravity, exposure to radiation, and antibiotic resistance.
Connectivity Restoration in Wireless Sensor Networks via Space Network Coding.
Uwitonze, Alfred; Huang, Jiaqing; Ye, Yuanqing; Cheng, Wenqing
2017-04-20
The problem of finding the number and optimal positions of relay nodes for restoring the network connectivity in partitioned Wireless Sensor Networks (WSNs) is Non-deterministic Polynomial-time hard (NP-hard) and thus heuristic methods are preferred to solve it. This paper proposes a novel polynomial time heuristic algorithm, namely, Relay Placement using Space Network Coding (RPSNC), to solve this problem, where Space Network Coding, also called Space Information Flow (SIF), is a new research paradigm that studies network coding in Euclidean space, in which extra relay nodes can be introduced to reduce the cost of communication. Unlike contemporary schemes that are often based on Minimum Spanning Tree (MST), Euclidean Steiner Minimal Tree (ESMT) or a combination of MST with ESMT, RPSNC is a new min-cost multicast space network coding approach that combines Delaunay triangulation and non-uniform partitioning techniques for generating a number of candidate relay nodes, and then linear programming is applied for choosing the optimal relay nodes and computing their connection links with terminals. Subsequently, an equilibrium method is used to refine the locations of the optimal relay nodes, by moving them to balanced positions. RPSNC can adapt to any density distribution of relay nodes and terminals, as well as any density distribution of terminals. The performance and complexity of RPSNC are analyzed and its performance is validated through simulation experiments.
Effective degrees of freedom of a random walk on a fractal
NASA Astrophysics Data System (ADS)
Balankin, Alexander S.
2015-12-01
We argue that a non-Markovian random walk on a fractal can be treated as a Markovian process in a fractional dimensional space with a suitable metric. This allows us to define the fractional dimensional space allied to the fractal as the ν -dimensional space Fν equipped with the metric induced by the fractal topology. The relation between the number of effective spatial degrees of freedom of walkers on the fractal (ν ) and fractal dimensionalities is deduced. The intrinsic time of random walk in Fν is inferred. The Laplacian operator in Fν is constructed. This allows us to map physical problems on fractals into the corresponding problems in Fν. In this way, essential features of physics on fractals are revealed. Particularly, subdiffusion on path-connected fractals is elucidated. The Coulomb potential of a point charge on a fractal embedded in the Euclidean space is derived. Intriguing attributes of some types of fractals are highlighted.
Self-similar space-time evolution of an initial density discontinuity
NASA Astrophysics Data System (ADS)
Rekaa, V. L.; Pécseli, H. L.; Trulsen, J. K.
2013-07-01
The space-time evolution of an initial step-like plasma density variation is studied. We give particular attention to formulate the problem in a way that opens for the possibility of realizing the conditions experimentally. After a short transient time interval of the order of the electron plasma period, the solution is self-similar as illustrated by a video where the space-time evolution is reduced to be a function of the ratio x/t. Solutions of this form are usually found for problems without characteristic length and time scales, in our case the quasi-neutral limit. By introducing ion collisions with neutrals into the numerical analysis, we introduce a length scale, the collisional mean free path. We study the breakdown of the self-similarity of the solution as the mean free path is made shorter than the system length. Analytical results are presented for charge exchange collisions, demonstrating a short time collisionless evolution with an ensuing long time diffusive relaxation of the initial perturbation. For large times, we find a diffusion equation as the limiting analytical form for a charge-exchange collisional plasma, with a diffusion coefficient defined as the square of the ion sound speed divided by the (constant) ion collision frequency. The ion-neutral collision frequency acts as a parameter that allows a collisionless result to be obtained in one limit, while the solution of a diffusion equation is recovered in the opposite limit of large collision frequencies.
A space-efficient quantum computer simulator suitable for high-speed FPGA implementation
NASA Astrophysics Data System (ADS)
Frank, Michael P.; Oniciuc, Liviu; Meyer-Baese, Uwe H.; Chiorescu, Irinel
2009-05-01
Conventional vector-based simulators for quantum computers are quite limited in the size of the quantum circuits they can handle, due to the worst-case exponential growth of even sparse representations of the full quantum state vector as a function of the number of quantum operations applied. However, this exponential-space requirement can be avoided by using general space-time tradeoffs long known to complexity theorists, which can be appropriately optimized for this particular problem in a way that also illustrates some interesting reformulations of quantum mechanics. In this paper, we describe the design and empirical space/time complexity measurements of a working software prototype of a quantum computer simulator that avoids excessive space requirements. Due to its space-efficiency, this design is well-suited to embedding in single-chip environments, permitting especially fast execution that avoids access latencies to main memory. We plan to prototype our design on a standard FPGA development board.
A New Approach for Solving the Generalized Traveling Salesman Problem
NASA Astrophysics Data System (ADS)
Pop, P. C.; Matei, O.; Sabo, C.
The generalized traveling problem (GTSP) is an extension of the classical traveling salesman problem. The GTSP is known to be an NP-hard problem and has many interesting applications. In this paper we present a local-global approach for the generalized traveling salesman problem. Based on this approach we describe a novel hybrid metaheuristic algorithm for solving the problem using genetic algorithms. Computational results are reported for Euclidean TSPlib instances and compared with the existing ones. The obtained results point out that our hybrid algorithm is an appropriate method to explore the search space of this complex problem and leads to good solutions in a reasonable amount of time.
Green and Blue Spaces and Behavioral Development in Barcelona Schoolchildren: The BREATHE Project
Amoly, Elmira; Forns, Joan; López-Vicente, Mónica; Basagaña, Xavier; Julvez, Jordi; Alvarez-Pedrerol, Mar; Nieuwenhuijsen, Mark J.; Sunyer, Jordi
2014-01-01
Background: Green spaces have been associated with improved mental health in children; however, available epidemiological evidence on their impact on child behavioral development is scarce. Objectives: We investigated the impact of contact with green spaces and blue spaces (beaches) on indicators of behavioral development and symptoms of attention deficit/hyperactivity disorder (ADHD) in schoolchildren. Methods: This study was based on a sample of 2,111 schoolchildren (7–10 years of age) from 36 schools in Barcelona in 2012. We obtained data on time spent in green spaces and beaches and Strengths and Difficulties Questionnaires (SDQ) from parents, and ADHD/DSM-IV questionnaires from teachers. Surrounding greenness was abstracted as the average Normalized Difference Vegetation Index (NDVI) in buffers of 100 m, 250 m, and 500 m around each home address. Proximity to green spaces was defined as living within 300 m of a major green space (≥ 0.05 km2). We applied quasi-Poisson mixed-effects models (with school random effect) to separately estimate associations between indicators of contact with green spaces and SDQ and ADHD total and subscale scores. Results: We generally estimated beneficial associations between behavioral indicators and longer time spent in green spaces and beaches, and with residential surrounding greenness. Specifically, we found statistically significant inverse associations between green space playing time and SDQ total difficulties, emotional symptoms, and peer relationship problems; between residential surrounding greenness and SDQ total difficulties and hyperactivity/inattention and ADHD/DSM-IV total and inattention scores; and between annual beach attendance and SDQ total difficulties, peer relationship problems, and prosocial behavior. For proximity to major green spaces, the results were not conclusive. Conclusion: Our findings support beneficial impacts of contact with green and blue spaces on behavioral development in schoolchildren. Citation: Amoly E, Dadvand P, Forns J, López-Vicente M, Basagaña X, Julvez J, Alvarez-Pedrerol M, Nieuwenhuijsen MJ, Sunyer J. 2014. Green and blue spaces and behavioral development in Barcelona schoolchildren: the BREATHE Project. Environ Health Perspect 122:1351–1358; http://dx.doi.org/10.1289/ehp.1408215 PMID:25204008
On non-autonomous dynamical systems
NASA Astrophysics Data System (ADS)
Anzaldo-Meneses, A.
2015-04-01
In usual realistic classical dynamical systems, the Hamiltonian depends explicitly on time. In this work, a class of classical systems with time dependent nonlinear Hamiltonians is analyzed. This type of problems allows to find invariants by a family of Veronese maps. The motivation to develop this method results from the observation that the Poisson-Lie algebra of monomials in the coordinates and momenta is clearly defined in terms of its brackets and leads naturally to an infinite linear set of differential equations, under certain circumstances. To perform explicit analytic and numerical calculations, two examples are presented to estimate the trajectories, the first given by a nonlinear problem and the second by a quadratic Hamiltonian with three time dependent parameters. In the nonlinear problem, the Veronese approach using jets is shown to be equivalent to a direct procedure using elliptic functions identities, and linear invariants are constructed. For the second example, linear and quadratic invariants as well as stability conditions are given. Explicit solutions are also obtained for stepwise constant forces. For the quadratic Hamiltonian, an appropriated set of coordinates relates the geometric setting to that of the three dimensional manifold of central conic sections. It is shown further that the quantum mechanical problem of scattering in a superlattice leads to mathematically equivalent equations for the wave function, if the classical time is replaced by the space coordinate along a superlattice. The mathematical method used to compute the trajectories for stepwise constant parameters can be applied to both problems. It is the standard method in quantum scattering calculations, as known for locally periodic systems including a space dependent effective mass.
Computational Study for Planar Connected Dominating Set Problem
NASA Astrophysics Data System (ADS)
Marzban, Marjan; Gu, Qian-Ping; Jia, Xiaohua
The connected dominating set (CDS) problem is a well studied NP-hard problem with many important applications. Dorn et al. [ESA2005, LNCS3669,pp95-106] introduce a new technique to generate 2^{O(sqrt{n})} time and fixed-parameter algorithms for a number of non-local hard problems, including the CDS problem in planar graphs. The practical performance of this algorithm is yet to be evaluated. We perform a computational study for such an evaluation. The results show that the size of instances can be solved by the algorithm mainly depends on the branchwidth of the instances, coinciding with the theoretical result. For graphs with small or moderate branchwidth, the CDS problem instances with size up to a few thousands edges can be solved in a practical time and memory space. This suggests that the branch-decomposition based algorithms can be practical for the planar CDS problem.
A real-space approach to the X-ray phase problem
NASA Astrophysics Data System (ADS)
Liu, Xiangan
Over the past few decades, the phase problem of X-ray crystallography has been explored in reciprocal space in the so called direct methods . Here we investigate the problem using a real-space approach that bypasses the laborious procedure of frequent Fourier synthesis and peak picking. Starting from a completely random structure, we move the atoms around in real space to minimize a cost function. A Monte Carlo method named simulated annealing (SA) is employed to search the global minimum of the cost function which could be constructed in either real space or reciprocal space. In the hybrid minimal principle, we combine the dual space costs together. One part of the cost function monitors the probability distribution of the phase triplets, while the other is a real space cost function which represents the discrepancy between measured and calculated intensities. Compared to the single space cost functions, the dual space cost function has a greatly improved landscape and therefore could prevent the system from being trapped in metastable states. Thus, the structures of large molecules such as virginiamycin (C43H 49N7O10 · 3CH0OH), isoleucinomycin (C60H102N 6O18) and hexadecaisoleucinomycin (HEXIL) (C80H136 N8O24) can now be solved, whereas it would not be possible using the single cost function. When a molecule gets larger, the configurational space becomes larger, and the requirement of CPU time increases exponentially. The method of improved Monte Carlo sampling has demonstrated its capability to solve large molecular structures. The atoms are encouraged to sample the high density regions in space determined by an approximate density map which in turn is updated and modified by averaging and Fourier synthesis. This type of biased sampling has led to considerable reduction of the configurational space. It greatly improves the algorithm compared to the previous uniform sampling. Hence, for instance, 90% of computer run time could be cut in solving the complex structure of isoleucinomycin. Successful trial calculations include larger molecular structures such as HEXIL and a collagen-like peptide (PPG). Moving chemical fragment is proposed to reduce the degrees of freedom. Furthermore, stereochemical parameters are considered for geometric constraints and for a cost function related to chemical energy.
Lörincz, András; Póczos, Barnabás
2003-06-01
In optimizations the dimension of the problem may severely, sometimes exponentially increase optimization time. Parametric function approximatiors (FAPPs) have been suggested to overcome this problem. Here, a novel FAPP, cost component analysis (CCA) is described. In CCA, the search space is resampled according to the Boltzmann distribution generated by the energy landscape. That is, CCA converts the optimization problem to density estimation. Structure of the induced density is searched by independent component analysis (ICA). The advantage of CCA is that each independent ICA component can be optimized separately. In turn, (i) CCA intends to partition the original problem into subproblems and (ii) separating (partitioning) the original optimization problem into subproblems may serve interpretation. Most importantly, (iii) CCA may give rise to high gains in optimization time. Numerical simulations illustrate the working of the algorithm.
NASA Astrophysics Data System (ADS)
Beyhaghi, Pooriya
2016-11-01
This work considers the problem of the efficient minimization of the infinite time average of a stationary ergodic process in the space of a handful of independent parameters which affect it. Problems of this class, derived from physical or numerical experiments which are sometimes expensive to perform, are ubiquitous in turbulence research. In such problems, any given function evaluation, determined with finite sampling, is associated with a quantifiable amount of uncertainty, which may be reduced via additional sampling. This work proposes the first algorithm of this type. Our algorithm remarkably reduces the overall cost of the optimization process for problems of this class. Further, under certain well-defined conditions, rigorous proof of convergence is established to the global minimum of the problem considered.
Enhanced Sampling Methods for the Computation of Conformational Kinetics in Macromolecules
NASA Astrophysics Data System (ADS)
Grazioli, Gianmarc
Calculating the kinetics of conformational changes in macromolecules, such as proteins and nucleic acids, is still very much an open problem in theoretical chemistry and computational biophysics. If it were feasible to run large sets of molecular dynamics trajectories that begin in one configuration and terminate when reaching another configuration of interest, calculating kinetics from molecular dynamics simulations would be simple, but in practice, configuration spaces encompassing all possible configurations for even the simplest of macromolecules are far too vast for such a brute force approach. In fact, many problems related to searches of configuration spaces, such as protein structure prediction, are considered to be NP-hard. Two approaches to addressing this problem are to either develop methods for enhanced sampling of trajectories that confine the search to productive trajectories without loss of temporal information, or coarse-grained methodologies that recast the problem in reduced spaces that can be exhaustively searched. This thesis will begin with a description of work carried out in the vein of the second approach, where a Smoluchowski diffusion equation model was developed that accurately reproduces the rate vs. force relationship observed in the mechano-catalytic disulphide bond cleavage observed in thioredoxin-catalyzed reduction of disulphide bonds. Next, three different novel enhanced sampling methods developed in the vein of the first approach will be described, which can be employed either separately or in conjunction with each other to autonomously define a set of energetically relevant subspaces in configuration space, accelerate trajectories between the interfaces dividing the subspaces while preserving the distribution of unassisted transition times between subspaces, and approximate time correlation functions from the kinetic data collected from the transitions between interfaces.
Metrics in Keplerian orbits quotient spaces
NASA Astrophysics Data System (ADS)
Milanov, Danila V.
2018-03-01
Quotient spaces of Keplerian orbits are important instruments for the modelling of orbit samples of celestial bodies on a large time span. We suppose that variations of the orbital eccentricities, inclinations and semi-major axes remain sufficiently small, while arbitrary perturbations are allowed for the arguments of pericentres or longitudes of the nodes, or both. The distance between orbits or their images in quotient spaces serves as a numerical criterion for such problems of Celestial Mechanics as search for common origin of meteoroid streams, comets, and asteroids, asteroid families identification, and others. In this paper, we consider quotient sets of the non-rectilinear Keplerian orbits space H. Their elements are identified irrespective of the values of pericentre arguments or node longitudes. We prove that distance functions on the quotient sets, introduced in Kholshevnikov et al. (Mon Not R Astron Soc 462:2275-2283, 2016), satisfy metric space axioms and discuss theoretical and practical importance of this result. Isometric embeddings of the quotient spaces into R^n, and a space of compact subsets of H with Hausdorff metric are constructed. The Euclidean representations of the orbits spaces find its applications in a problem of orbit averaging and computational algorithms specific to Euclidean space. We also explore completions of H and its quotient spaces with respect to corresponding metrics and establish a relation between elements of the extended spaces and rectilinear trajectories. Distance between an orbit and subsets of elliptic and hyperbolic orbits is calculated. This quantity provides an upper bound for the metric value in a problem of close orbits identification. Finally the invariance of the equivalence relations in H under coordinates change is discussed.
Creating a Learning Space in Problem-Based Learning
ERIC Educational Resources Information Center
Hmelo-Silver, Cindy E.
2013-01-01
An important aspect of PBL problems is the affordances that they hold for engaging students in discussion of important content knowledge. In this paper, I argue that one can analyze a problem in terms of a deep problem space and a broader learning space to identify the conceptual ideas for potential engagement. The problem space refers to the…
NASA Astrophysics Data System (ADS)
Noer, Fadhly; Matondang, A. Rahim; Sirojuzilam, Saleh, Sofyan M.
2017-11-01
Due to the shifting of city urban development causing the shift of city services center, so there is a change in space pattern and space structure in Banda Aceh, then resulting urban sprawl which can lead to congestion problem occurs on the arterial road in Banda Aceh, it can be seen from the increasing number of vehicles per year by 6%. Another issue occurs by urban sprawl is not well organized of settlement due to the uncontrolled use of space so that caused grouping or the differences in socioeconomic strata that can impact to the complexity of population mobility problem. From this background problem considered to be solved by a concept that is Transit Oriented Development (TOD), that is a concept of transportation development in co-operation with spatial. This research will get the model of transportation infrastructure development with TOD concept that can handle transportation problem in Banda Aceh, due to change of spatial structure, and to find whether TOD concept can use for the area that has a population in medium density range. The result that is obtained equation so the space structure is: Space Structure = 0.520 + 0.206X3 + 0.264X6 + 0.100X7 and Transportation Infrastructure Development = -1.457 + 0.652X1 + 0.388X5 + 0.235X6 + 0.222X7 + 0.327X8, So results obtained with path analysis method obtained variable influences, node ratio, network connectivity, travel frequency, travel destination, travel cost, and travel time, it has a lower value when direct effect with transportation infrastructure development, but if the indirect effect through the structure of space has a greater influence, can be seen from spatial structure path scheme - transportation infrastructure development.
Sparse RNA folding revisited: space-efficient minimum free energy structure prediction.
Will, Sebastian; Jabbari, Hosna
2016-01-01
RNA secondary structure prediction by energy minimization is the central computational tool for the analysis of structural non-coding RNAs and their interactions. Sparsification has been successfully applied to improve the time efficiency of various structure prediction algorithms while guaranteeing the same result; however, for many such folding problems, space efficiency is of even greater concern, particularly for long RNA sequences. So far, space-efficient sparsified RNA folding with fold reconstruction was solved only for simple base-pair-based pseudo-energy models. Here, we revisit the problem of space-efficient free energy minimization. Whereas the space-efficient minimization of the free energy has been sketched before, the reconstruction of the optimum structure has not even been discussed. We show that this reconstruction is not possible in trivial extension of the method for simple energy models. Then, we present the time- and space-efficient sparsified free energy minimization algorithm SparseMFEFold that guarantees MFE structure prediction. In particular, this novel algorithm provides efficient fold reconstruction based on dynamically garbage-collected trace arrows. The complexity of our algorithm depends on two parameters, the number of candidates Z and the number of trace arrows T; both are bounded by [Formula: see text], but are typically much smaller. The time complexity of RNA folding is reduced from [Formula: see text] to [Formula: see text]; the space complexity, from [Formula: see text] to [Formula: see text]. Our empirical results show more than 80 % space savings over RNAfold [Vienna RNA package] on the long RNAs from the RNA STRAND database (≥2500 bases). The presented technique is intentionally generalizable to complex prediction algorithms; due to their high space demands, algorithms like pseudoknot prediction and RNA-RNA-interaction prediction are expected to profit even stronger than "standard" MFE folding. SparseMFEFold is free software, available at http://www.bioinf.uni-leipzig.de/~will/Software/SparseMFEFold.
NASA Technical Reports Server (NTRS)
Chang, S.-C.; Himansu, A.; Loh, C.-Y.; Wang, X.-Y.; Yu, S.-T.J.
2005-01-01
This paper reports on a significant advance in the area of nonreflecting boundary conditions (NRBCs) for unsteady flow computations. As a part of t he development of t he space-time conservation element and solution element (CE/SE) method, sets of NRBCs for 1D Euler problems are developed without using any characteristics- based techniques. These conditions are much simpler than those commonly reported in the literature, yet so robust that they are applicable to subsonic, transonic and supersonic flows even in the presence of discontinuities. In addition, the straightforward multidimensional extensions of the present 1D NRBCs have been shown numerically to be equally simple and robust. The paper details the theoretical underpinning of these NRBCs, and explains t heir unique robustness and accuracy in terms of t he conservation of space-time fluxes. Some numerical results for an extended Sod's shock-tube problem, illustrating the effectiveness of the present NRBCs are included, together with an associated simple Fortran computer program. As a preliminary to the present development, a review of the basic CE/SE schemes is also included.
Tailoff thrust and impulse imbalance between pairs of Space Shuttle solid rocket motors
NASA Technical Reports Server (NTRS)
Jacobs, E. P.; Yeager, J. M.
1975-01-01
The tailoff thrust and impulse imbalance between pairs of solid rocket motors is of particular interest for the Space Shuttle Vehicle because of the potential control problems that exist with this asymmetric configuration. Although a similar arrangement of solid rocket motors was utilized for the Titan Program, they produced less than one-half the thrust level of the Space Shuttle at web action time, and the overall vehicle was symmetric. Since the Titan Program does provide the most applicable actual test data, 23 flight pairs were analyzed to determine the actual tailoff thrust and impulse imbalance experienced. The results were scaled up using the predicted web action time thrust and tailoff time to arrive at values for the Space Shuttle. These values were then statistically treated to obtain a prediction of the maximum imbalance one could expect to experience during the Shuttle Program.
NASA Technical Reports Server (NTRS)
Shubert, W. C.
1973-01-01
Transportation requirements are considered during the engine design layout reviews and maintenance engineering analyses. Where designs cannot be influenced to avoid transportation problems, the transportation representative is advised of the problems permitting remedies early in the program. The transportation representative will monitor and be involved in the shipment of development engine and GSE hardware between FRDC and vehicle manufacturing plant and thereby will be provided an early evaluation of the transportation plans, methods and procedures to be used in the space tug support program. Unanticipated problems discovered in the shipment of development hardware will be known early enough to permit changes in packaging designs and transportation plans before the start of production hardware and engine shipments. All conventional transport media can be used for the movement of space tug engines. However, truck transport is recommended for ready availability, variety of routes, short transit time, and low cost.
NASA Technical Reports Server (NTRS)
Venkatachari, Balaji Shankar; Streett, Craig L.; Chang, Chau-Lyan; Friedlander, David J.; Wang, Xiao-Yen; Chang, Sin-Chung
2016-01-01
Despite decades of development of unstructured mesh methods, high-fidelity time-accurate simulations are still predominantly carried out on structured, or unstructured hexahedral meshes by using high-order finite-difference, weighted essentially non-oscillatory (WENO), or hybrid schemes formed by their combinations. In this work, the space-time conservation element solution element (CESE) method is used to simulate several flow problems including supersonic jet/shock interaction and its impact on launch vehicle acoustics, and direct numerical simulations of turbulent flows using tetrahedral meshes. This paper provides a status report for the continuing development of the space-time conservation element solution element (CESE) numerical and software framework under the Revolutionary Computational Aerosciences (RCA) project. Solution accuracy and large-scale parallel performance of the numerical framework is assessed with the goal of providing a viable paradigm for future high-fidelity flow physics simulations.
Time-Tagged Risk/Reliability Assessment Program for Development and Operation of Space System
NASA Astrophysics Data System (ADS)
Kubota, Yuki; Takegahara, Haruki; Aoyagi, Junichiro
We have investigated a new method of risk/reliability assessment for development and operation of space system. It is difficult to evaluate risk of spacecraft, because of long time operation, maintenance free and difficulty of test under the ground condition. Conventional methods are FMECA, FTA, ETA and miscellaneous. These are not enough to assess chronological anomaly and there is a problem to share information during R&D. A new method of risk and reliability assessment, T-TRAP (Time-tagged Risk/Reliability Assessment Program) is proposed as a management tool for the development and operation of space system. T-TRAP consisting of time-resolved Fault Tree and Criticality Analyses, upon occurrence of anomaly in the system, facilitates the responsible personnel to quickly identify the failure cause and decide corrective actions. This paper describes T-TRAP method and its availability.
Accelerating molecular property calculations with nonorthonormal Krylov space methods
DOE Office of Scientific and Technical Information (OSTI.GOV)
Furche, Filipp; Krull, Brandon T.; Nguyen, Brian D.
Here, we formulate Krylov space methods for large eigenvalue problems and linear equation systems that take advantage of decreasing residual norms to reduce the cost of matrix-vector multiplication. The residuals are used as subspace basis without prior orthonormalization, which leads to generalized eigenvalue problems or linear equation systems on the Krylov space. These nonorthonormal Krylov space (nKs) algorithms are favorable for large matrices with irregular sparsity patterns whose elements are computed on the fly, because fewer operations are necessary as the residual norm decreases as compared to the conventional method, while errors in the desired eigenpairs and solution vectors remainmore » small. We consider real symmetric and symplectic eigenvalue problems as well as linear equation systems and Sylvester equations as they appear in configuration interaction and response theory. The nKs method can be implemented in existing electronic structure codes with minor modifications and yields speed-ups of 1.2-1.8 in typical time-dependent Hartree-Fock and density functional applications without accuracy loss. The algorithm can compute entire linear subspaces simultaneously which benefits electronic spectra and force constant calculations requiring many eigenpairs or solution vectors. The nKs approach is related to difference density methods in electronic ground state calculations, and particularly efficient for integral direct computations of exchange-type contractions. By combination with resolution-of-the-identity methods for Coulomb contractions, three- to fivefold speed-ups of hybrid time-dependent density functional excited state and response calculations are achieved.« less
Accelerating molecular property calculations with nonorthonormal Krylov space methods
Furche, Filipp; Krull, Brandon T.; Nguyen, Brian D.; ...
2016-05-03
Here, we formulate Krylov space methods for large eigenvalue problems and linear equation systems that take advantage of decreasing residual norms to reduce the cost of matrix-vector multiplication. The residuals are used as subspace basis without prior orthonormalization, which leads to generalized eigenvalue problems or linear equation systems on the Krylov space. These nonorthonormal Krylov space (nKs) algorithms are favorable for large matrices with irregular sparsity patterns whose elements are computed on the fly, because fewer operations are necessary as the residual norm decreases as compared to the conventional method, while errors in the desired eigenpairs and solution vectors remainmore » small. We consider real symmetric and symplectic eigenvalue problems as well as linear equation systems and Sylvester equations as they appear in configuration interaction and response theory. The nKs method can be implemented in existing electronic structure codes with minor modifications and yields speed-ups of 1.2-1.8 in typical time-dependent Hartree-Fock and density functional applications without accuracy loss. The algorithm can compute entire linear subspaces simultaneously which benefits electronic spectra and force constant calculations requiring many eigenpairs or solution vectors. The nKs approach is related to difference density methods in electronic ground state calculations, and particularly efficient for integral direct computations of exchange-type contractions. By combination with resolution-of-the-identity methods for Coulomb contractions, three- to fivefold speed-ups of hybrid time-dependent density functional excited state and response calculations are achieved.« less
NASA Technical Reports Server (NTRS)
Hein, C.; Meystel, A.
1994-01-01
There are many multi-stage optimization problems that are not easily solved through any known direct method when the stages are coupled. For instance, we have investigated the problem of planning a vehicle's control sequence to negotiate obstacles and reach a goal in minimum time. The vehicle has a known mass, and the controlling forces have finite limits. We have developed a technique that finds admissible control trajectories which tend to minimize the vehicle's transit time through the obstacle field. The immediate applications is that of a space robot which must rapidly traverse around 2-or-3 dimensional structures via application of a rotating thruster or non-rotating on-off for such vehicles is located at the Marshall Space Flight Center in Huntsville Alabama. However, it appears that the development method is applicable to a general set of optimization problems in which the cost function and the multi-dimensional multi-state system can be any nonlinear functions, which are continuous in the operating regions. Other applications included the planning of optimal navigation pathways through a transversability graph; the planning of control input for under-water maneuvering vehicles which have complex control state-space relationships; the planning of control sequences for milling and manufacturing robots; the planning of control and trajectories for automated delivery vehicles; and the optimization and athletic training in slalom sports.
NASA Astrophysics Data System (ADS)
Min, Huang; Na, Cai
2017-06-01
These years, ant colony algorithm has been widely used in solving the domain of discrete space optimization, while the research on solving the continuous space optimization was relatively little. Based on the original optimization for continuous space, the article proposes the improved ant colony algorithm which is used to Solve the optimization for continuous space, so as to overcome the ant colony algorithm’s disadvantages of searching for a long time in continuous space. The article improves the solving way for the total amount of information of each interval and the due number of ants. The article also introduces a function of changes with the increase of the number of iterations in order to enhance the convergence rate of the improved ant colony algorithm. The simulation results show that compared with the result in literature[5], the suggested improved ant colony algorithm that based on the information distribution function has a better convergence performance. Thus, the article provides a new feasible and effective method for ant colony algorithm to solve this kind of problem.
Multiplexed phase-space imaging for 3D fluorescence microscopy.
Liu, Hsiou-Yuan; Zhong, Jingshan; Waller, Laura
2017-06-26
Optical phase-space functions describe spatial and angular information simultaneously; examples of optical phase-space functions include light fields in ray optics and Wigner functions in wave optics. Measurement of phase-space enables digital refocusing, aberration removal and 3D reconstruction. High-resolution capture of 4D phase-space datasets is, however, challenging. Previous scanning approaches are slow, light inefficient and do not achieve diffraction-limited resolution. Here, we propose a multiplexed method that solves these problems. We use a spatial light modulator (SLM) in the pupil plane of a microscope in order to sequentially pattern multiplexed coded apertures while capturing images in real space. Then, we reconstruct the 3D fluorescence distribution of our sample by solving an inverse problem via regularized least squares with a proximal accelerated gradient descent solver. We experimentally reconstruct a 101 Megavoxel 3D volume (1010×510×500µm with NA 0.4), demonstrating improved acquisition time, light throughput and resolution compared to scanning aperture methods. Our flexible patterning scheme further allows sparsity in the sample to be exploited for reduced data capture.
On Time/Space Aggregation of Fine-Scale Error Estimates (Invited)
NASA Astrophysics Data System (ADS)
Huffman, G. J.
2013-12-01
Estimating errors inherent in fine time/space-scale satellite precipitation data sets is still an on-going problem and a key area of active research. Complicating features of these data sets include the intrinsic intermittency of the precipitation in space and time and the resulting highly skewed distribution of precipitation rates. Additional issues arise from the subsampling errors that satellites introduce, the errors due to retrieval algorithms, and the correlated error that retrieval and merger algorithms sometimes introduce. Several interesting approaches have been developed recently that appear to make progress on these long-standing issues. At the same time, the monthly averages over 2.5°x2.5° grid boxes in the Global Precipitation Climatology Project (GPCP) Satellite-Gauge (SG) precipitation data set follow a very simple sampling-based error model (Huffman 1997) with coefficients that are set using coincident surface and GPCP SG data. This presentation outlines the unsolved problem of how to aggregate the fine-scale errors (discussed above) to an arbitrary time/space averaging volume for practical use in applications, reducing in the limit to simple Gaussian expressions at the monthly 2.5°x2.5° scale. Scatter diagrams with different time/space averaging show that the relationship between the satellite and validation data improves due to the reduction in random error. One of the key, and highly non-linear, issues is that fine-scale estimates tend to have large numbers of cases with points near the axes on the scatter diagram (one of the values is exactly or nearly zero, while the other value is higher). Averaging 'pulls' the points away from the axes and towards the 1:1 line, which usually happens for higher precipitation rates before lower rates. Given this qualitative observation of how aggregation affects error, we observe that existing aggregation rules, such as the Steiner et al. (2003) power law, only depend on the aggregated precipitation rate. Is this sufficient, or is it necessary to aggregate the precipitation error estimates across the time/space data cube used for averaging? At least for small time/space data cubes it would seem that the detailed variables that affect each precipitation error estimate in the aggregation, such as sensor type, land/ocean surface type, convective/stratiform type, and so on, drive variations that must be accounted for explicitly.
NASA Astrophysics Data System (ADS)
Bode, F.; Reuschen, S.; Nowak, W.
2015-12-01
Drinking-water well catchments include many potential sources of contaminations like gas stations or agriculture. Finding optimal positions of early-warning monitoring wells is challenging because there are various parameters (and their uncertainties) that influence the reliability and optimality of any suggested monitoring location or monitoring network.The overall goal of this project is to develop and establish a concept to assess, design and optimize early-warning systems within well catchments. Such optimal monitoring networks need to optimize three competing objectives: a high detection probability, which can be reached by maximizing the "field of vision" of the monitoring network, a long early-warning time such that there is enough time left to install counter measures after first detection, and the overall operating costs of the monitoring network, which should ideally be reduced to a minimum. The method is based on numerical simulation of flow and transport in heterogeneous porous media coupled with geostatistics and Monte-Carlo, scenario analyses for real data, respectively, wrapped up within the framework of formal multi-objective optimization using a genetic algorithm.In order to speed up the optimization process and to better explore the Pareto-front, we developed a concept that forces the algorithm to search only in regions of the search space where promising solutions can be expected. We are going to show how to define these regions beforehand, using knowledge of the optimization problem, but also how to define them independently of problem attributes. With that, our method can be used with and/or without detailed knowledge of the objective functions.In summary, our study helps to improve optimization results in less optimization time by meaningful restrictions of the search space. These restrictions can be done independently of the optimization problem, but also in a problem-specific manner.
Computer simulation of multigrid body dynamics and control
NASA Technical Reports Server (NTRS)
Swaminadham, M.; Moon, Young I.; Venkayya, V. B.
1990-01-01
The objective is to set up and analyze benchmark problems on multibody dynamics and to verify the predictions of two multibody computer simulation codes. TREETOPS and DISCOS have been used to run three example problems - one degree-of-freedom spring mass dashpot system, an inverted pendulum system, and a triple pendulum. To study the dynamics and control interaction, an inverted planar pendulum with an external body force and a torsional control spring was modeled as a hinge connected two-rigid body system. TREETOPS and DISCOS affected the time history simulation of this problem. System state space variables and their time derivatives from two simulation codes were compared.
ERIC Educational Resources Information Center
Nichols, Maria
2014-01-01
What happens in classrooms when we create the time and space for authentic talk about texts? Extended, collaborative conversations that allow understanding to unfold over time can be messy and dynamic. As students wrestle with complex texts and ideas, talk can become lively--and predictable problems can arise. In this article, Marie Nichols uses…
ERIC Educational Resources Information Center
Christian, David
1991-01-01
Urges an approach to the teaching of history that takes the largest possible perspective, crossing time as well as space. Discusses the problems and advantages of such an approach. Describes a course on "big" history that begins with time, creation myths, and astronomy, and moves on to paleontology and evolution. (DK)
Designing a supply chain of ready-mix concrete using Voronoi diagrams
NASA Astrophysics Data System (ADS)
Kozniewski, E.; Orlowski, M.; Orlowski, Z.
2017-10-01
Voronoi diagrams are used to solve scientific and practical problems in many fields. In this paper Voronoi diagrams have been applied to logistic problems in construction, more specifically in the design of the ready-mix concrete supply chain. Apart from the Voronoi diagram, the so-called time-distance circle (circle of range), which in metric space terminology is simply a sphere, appears useful. It was introduced to solve the problem of supplying concrete-related goods.
NASA Astrophysics Data System (ADS)
Azarnavid, Babak; Parand, Kourosh; Abbasbandy, Saeid
2018-06-01
This article discusses an iterative reproducing kernel method with respect to its effectiveness and capability of solving a fourth-order boundary value problem with nonlinear boundary conditions modeling beams on elastic foundations. Since there is no method of obtaining reproducing kernel which satisfies nonlinear boundary conditions, the standard reproducing kernel methods cannot be used directly to solve boundary value problems with nonlinear boundary conditions as there is no knowledge about the existence and uniqueness of the solution. The aim of this paper is, therefore, to construct an iterative method by the use of a combination of reproducing kernel Hilbert space method and a shooting-like technique to solve the mentioned problems. Error estimation for reproducing kernel Hilbert space methods for nonlinear boundary value problems have yet to be discussed in the literature. In this paper, we present error estimation for the reproducing kernel method to solve nonlinear boundary value problems probably for the first time. Some numerical results are given out to demonstrate the applicability of the method.
Stabilization and control of distributed systems with time-dependent spatial domains
NASA Technical Reports Server (NTRS)
Wang, P. K. C.
1990-01-01
This paper considers the problem of the stabilization and control of distributed systems with time-dependent spatial domains. The evolution of the spatial domains with time is described by a finite-dimensional system of ordinary differential equations, while the distributed systems are described by first-order or second-order linear evolution equations defined on appropriate Hilbert spaces. First, results pertaining to the existence and uniqueness of solutions of the system equations are presented. Then, various optimal control and stabilization problems are considered. The paper concludes with some examples which illustrate the application of the main results.
First Semiannual Report of the National Aeronautics and Space Administration
NASA Technical Reports Server (NTRS)
Glennan, T. Keith
1959-01-01
The First Semiannual Report of the National Aeronautics and Space Administration (NASA) is submitted to Congress pursuant to section 206 (a) of the National Aeronautics and Space Act of 1958 (Public Law 85-568) to provide for research into problems of flight within and outside the Earth's atmosphere, which states: The Administration shall submit to the President for transmittal to Congress, semiannually and at such other times as it deems desirable, a report on its activities and accomplishments.
Instationary Generalized Stokes Equations in Partially Periodic Domains
NASA Astrophysics Data System (ADS)
Sauer, Jonas
2018-06-01
We consider an instationary generalized Stokes system with nonhomogeneous divergence data under a periodic condition in only some directions. The problem is set in the whole space, the half space or in (after an identification of the periodic directions with a torus) bounded domains with sufficiently regular boundary. We show unique solvability for all times in Muckenhoupt weighted Lebesgue spaces. The divergence condition is dealt with by analyzing the associated reduced Stokes system and in particular by showing maximal regularity of the partially periodic reduced Stokes operator.
STS-78 Space Shuttle Mission Report
NASA Technical Reports Server (NTRS)
Fricke, Robert W., Jr.
1996-01-01
The STS-78 Space Shuttle Program Mission Report summarizes the Payload activities as well as the Orbiter, External Tank (ET), Solid Rocket Booster (SRB), Reusable Solid Rocket Motor (RSRM), and the Space Shuttle main engine (SSME) systems performance during the seventy-eighth flight of the Space Shuttle Program, the fifty-third flight since the return-to-flight, and the twentieth flight of the Orbiter Columbia (OV-102). In addition to the Orbiter, the flight vehicle consisted of an ET that was designated ET-79; three SSME's that were designated as serial numbers 2041, 2039, and 2036 in positions 1, 2, and 3, respectively; and two SRB's that were designated BI-081. The RSRM's, designated RSRM-55, were installed in each SRB and the individual RSRM's were designated as 360L055A for the left SRB, and 360L055B for the right SRB. The STS-78 Space Shuttle Program Mission Report fulfills the Space Shuttle Program requirement as documented in NSTS 07700, Volume 7, Appendix E. The requirement stated in that document is that each organizational element supporting the Program will report the results of their hardware (and software) evaluation and mission performance plus identify all related in-flight anomalies. The primary objective of this flight was to successfully perform the planned operations of the Life and Microgravity Spacelab experiments. The secondary objectives of this flight were to complete the operations of the Orbital Acceleration Research Experiment (OARE), Biological Research in Canister Unit-Block II (BRIC), and the Shuttle Amateur Radio Experiment II-Configuration C (SAREX-II). The STS-78 mission was planned as a 16-day, plus one day flight plus two contingency days, which were available for weather avoidance or Orbiter contingency operations. The sequence of events for the STS-78 mission is shown in Table 1, and the Space Shuttle Vehicle Management Office Problem Tracking List is shown in Table 2. The Government Furnished Equipment/Flight Crew Equipment (GFE/FCE) Problem Tracking List is shown in Table 3. The Marshall Space Flight Center (MSFC) Problem Tracking List is shown in Table 4. Appendix A lists the sources of data, both formal and informal, that were used to prepare this report. Appendix B provides the definition of acronyms and abbreviations used throughout the report. All times during the flight are given in Greenwich mean time (G.m.t.) and mission elapsed time (MET).
ERIC Educational Resources Information Center
Harris, Cornelia C.
1969-01-01
Gives reasons for opposing kindergarten teachers, teaching double sessions because of resulting problems relating to close scheduling, lack of space, lack of time for parent conferences and, inability of teachers to adequately respond to individual children's needs. (DR)
Approaching German Culture: A Tentative Analysis
ERIC Educational Resources Information Center
Tinsley, Royal; Woloshin, David
1974-01-01
A comparative analysis of the five universal problems of cultural orientation: 1) human nature, 2) social relations, 3) man and nature, 4) time, 5) space, as they are reflected in German and American culture. (PP)
NASA Astrophysics Data System (ADS)
Kong, Fande; Cai, Xiao-Chuan
2017-07-01
Nonlinear fluid-structure interaction (FSI) problems on unstructured meshes in 3D appear in many applications in science and engineering, such as vibration analysis of aircrafts and patient-specific diagnosis of cardiovascular diseases. In this work, we develop a highly scalable, parallel algorithmic and software framework for FSI problems consisting of a nonlinear fluid system and a nonlinear solid system, that are coupled monolithically. The FSI system is discretized by a stabilized finite element method in space and a fully implicit backward difference scheme in time. To solve the large, sparse system of nonlinear algebraic equations at each time step, we propose an inexact Newton-Krylov method together with a multilevel, smoothed Schwarz preconditioner with isogeometric coarse meshes generated by a geometry preserving coarsening algorithm. Here "geometry" includes the boundary of the computational domain and the wet interface between the fluid and the solid. We show numerically that the proposed algorithm and implementation are highly scalable in terms of the number of linear and nonlinear iterations and the total compute time on a supercomputer with more than 10,000 processor cores for several problems with hundreds of millions of unknowns.
Kong, Fande; Cai, Xiao-Chuan
2017-03-24
Nonlinear fluid-structure interaction (FSI) problems on unstructured meshes in 3D appear many applications in science and engineering, such as vibration analysis of aircrafts and patient-specific diagnosis of cardiovascular diseases. In this work, we develop a highly scalable, parallel algorithmic and software framework for FSI problems consisting of a nonlinear fluid system and a nonlinear solid system, that are coupled monolithically. The FSI system is discretized by a stabilized finite element method in space and a fully implicit backward difference scheme in time. To solve the large, sparse system of nonlinear algebraic equations at each time step, we propose an inexactmore » Newton-Krylov method together with a multilevel, smoothed Schwarz preconditioner with isogeometric coarse meshes generated by a geometry preserving coarsening algorithm. Here ''geometry'' includes the boundary of the computational domain and the wet interface between the fluid and the solid. We show numerically that the proposed algorithm and implementation are highly scalable in terms of the number of linear and nonlinear iterations and the total compute time on a supercomputer with more than 10,000 processor cores for several problems with hundreds of millions of unknowns.« less
Computation of forces from deformed visco-elastic biological tissues
NASA Astrophysics Data System (ADS)
Muñoz, José J.; Amat, David; Conte, Vito
2018-04-01
We present a least-squares based inverse analysis of visco-elastic biological tissues. The proposed method computes the set of contractile forces (dipoles) at the cell boundaries that induce the observed and quantified deformations. We show that the computation of these forces requires the regularisation of the problem functional for some load configurations that we study here. The functional measures the error of the dynamic problem being discretised in time with a second-order implicit time-stepping and in space with standard finite elements. We analyse the uniqueness of the inverse problem and estimate the regularisation parameter by means of an L-curved criterion. We apply the methodology to a simple toy problem and to an in vivo set of morphogenetic deformations of the Drosophila embryo.
Space power system scheduling using an expert system
NASA Technical Reports Server (NTRS)
Bahrami, K. A.; Biefeld, E.; Costello, L.; Klein, J. W.
1986-01-01
A most pressing problem in space exploration is timely spacecraft power system sequence generation, which requires the scheduling of a set of loads given a set of resource constraints. This is particularly important after an anomaly or failure. This paper discusses the power scheduling problem and how the software program, Plan-It, can be used as a consultant for scheduling power system activities. Modeling of power activities, human interface, and two of the many strategies used by Plan-It are discussed. Preliminary results showing the development of a conflict-free sequence from an initial sequence with conflicts is presented. It shows that a 4-day schedule can be generated in a matter of a few minutes, which provides sufficient time in many cases to aid the crew in the replanning of loads and generation use following a failure or anomaly.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kaloper, N.
We discuss a particular stringy modular cosmology with two axion fields in seven space-time dimensions, decomposable as a time and two flat three-spaces. The effective equations of motion for the problem are those of the SU(3) Toda molecule and, hence, are integrable. We write down the solutions, and show that all of them are singular. They can be thought of as a generalization of the pre-big-bang cosmology with excited internal degrees of freedom, and still suffering from the graceful exit problem. Some of the solutions, however, show a rather unexpected property: some of their spatial sections shrink to a pointmore » in spite of winding modes wrapped around them. We also comment how more general, anisotropic solutions, with fewer Killing symmetries, can be obtained with the help of STU dualities. {copyright} {ital 1997} {ital The American Physical Society}« less
A test of the one-way isotropy of the speed of light from the T2L2 space experiment
NASA Astrophysics Data System (ADS)
Belli, A.; Exertier, P.; Samain, E.
2017-12-01
The Time Transfer by Laser Link (T2L2) space experiment that is currently flying on-board Jason-2 (1335 km of altitude) provides an opportunity to make a test of the isotropy of the speed of light using one-way propagation on a non-laboratory scale. Following the general framework given by te{Mansouri1977}, which permits violations of Einstein special relativity, we study the problem of deducing the isotropy of the speed between two clocks as the orientation path varies relative to an inertial reference frame. The short term stability of the T2L2 ground-to-space time transfer has been established at 5-6 ps at 60 seconds between a hydrogen maser and the on-board oscillator on use for the Jason-2 satellite. Nevertheless, during the satellite pass above a laser ranging station (of around 1000 seconds), the stability of the space oscillator is decreasing in τ^{3/2} that clearly impacts the expected performance of the present test. We thus give insights into certain modelling issues and processes, including time transfer problems which have a bearing on the global error budget. Our goal is to achieve an accuracy of {δc}/{c} ≈ 2-3.10^{-9} locally with a scope for improvement by cumulating numerous passes over the same laser ranging station.
Algebraically special space-time in relativity, black holes, and pulsar models
NASA Technical Reports Server (NTRS)
Adler, R. J.; Sheffield, C.
1973-01-01
The entire field of astronomy is in very rapid flux, and at the center of interest are problems relating to the very dense, rotating, neutron stars observed as pulsars. the hypothesized collapsed remains of stars known as black holes, and quasars. Degenerate metric form, or Kerr-Schild metric form, was used to study several problems related to intense gravitational fields.
2D and 3D Traveling Salesman Problem
ERIC Educational Resources Information Center
Haxhimusa, Yll; Carpenter, Edward; Catrambone, Joseph; Foldes, David; Stefanov, Emil; Arns, Laura; Pizlo, Zygmunt
2011-01-01
When a two-dimensional (2D) traveling salesman problem (TSP) is presented on a computer screen, human subjects can produce near-optimal tours in linear time. In this study we tested human performance on a real and virtual floor, as well as in a three-dimensional (3D) virtual space. Human performance on the real floor is as good as that on a…
Improved Fractal Space Filling Curves Hybrid Optimization Algorithm for Vehicle Routing Problem.
Yue, Yi-xiang; Zhang, Tong; Yue, Qun-xing
2015-01-01
Vehicle Routing Problem (VRP) is one of the key issues in optimization of modern logistics system. In this paper, a modified VRP model with hard time window is established and a Hybrid Optimization Algorithm (HOA) based on Fractal Space Filling Curves (SFC) method and Genetic Algorithm (GA) is introduced. By incorporating the proposed algorithm, SFC method can find an initial and feasible solution very fast; GA is used to improve the initial solution. Thereafter, experimental software was developed and a large number of experimental computations from Solomon's benchmark have been studied. The experimental results demonstrate the feasibility and effectiveness of the HOA.
Improved Fractal Space Filling Curves Hybrid Optimization Algorithm for Vehicle Routing Problem
Yue, Yi-xiang; Zhang, Tong; Yue, Qun-xing
2015-01-01
Vehicle Routing Problem (VRP) is one of the key issues in optimization of modern logistics system. In this paper, a modified VRP model with hard time window is established and a Hybrid Optimization Algorithm (HOA) based on Fractal Space Filling Curves (SFC) method and Genetic Algorithm (GA) is introduced. By incorporating the proposed algorithm, SFC method can find an initial and feasible solution very fast; GA is used to improve the initial solution. Thereafter, experimental software was developed and a large number of experimental computations from Solomon's benchmark have been studied. The experimental results demonstrate the feasibility and effectiveness of the HOA. PMID:26167171
Space-Time Dependent Transport, Activation, and Dose Rates for Radioactivated Fluids.
NASA Astrophysics Data System (ADS)
Gavazza, Sergio
Two methods are developed to calculate the space - and time-dependent mass transport of radionuclides, their production and decay, and the associated dose rates generated from the radioactivated fluids flowing through pipes. The work couples space- and time-dependent phenomena, treated as only space- or time-dependent in the open literature. The transport and activation methodology (TAM) is used to numerically calculate space- and time-dependent transport and activation of radionuclides in fluids flowing through pipes exposed to radiation fields, and volumetric radioactive sources created by radionuclide motions. The computer program Radionuclide Activation and Transport in Pipe (RNATPA1) performs the numerical calculations required in TAM. The gamma ray dose methodology (GAM) is used to numerically calculate space- and time-dependent gamma ray dose equivalent rates from the volumetric radioactive sources determined by TAM. The computer program Gamma Ray Dose Equivalent Rate (GRDOSER) performs the numerical calculations required in GAM. The scope of conditions considered by TAM and GAM herein include (a) laminar flow in straight pipe, (b)recirculating flow schemes, (c) time-independent fluid velocity distributions, (d) space-dependent monoenergetic neutron flux distribution, (e) space- and time-dependent activation process of a single parent nuclide and transport and decay of a single daughter radionuclide, and (f) assessment of space- and time-dependent gamma ray dose rates, outside the pipe, generated by the space- and time-dependent source term distributions inside of it. The methodologies, however, can be easily extended to include all the situations of interest for solving the phenomena addressed in this dissertation. A comparison is made from results obtained by the described calculational procedures with analytical expressions. The physics of the problems addressed by the new technique and the increased accuracy versus non -space and time-dependent methods are presented. The value of the methods is also discussed. It has been demonstrated that TAM and GAM can be used to enhance the understanding of the space- and time-dependent mass transport of radionuclides, their production and decay, and the associated dose rates related to radioactivated fluids flowing through pipes.
One Controller at a Time (1-CAT): A mimo design methodology
NASA Technical Reports Server (NTRS)
Mitchell, J. R.; Lucas, J. C.
1987-01-01
The One Controller at a Time (1-CAT) methodology for designing digital controllers for Large Space Structures (LSS's) is introduced and illustrated. The flexible mode problem is first discussed. Next, desirable features of a LSS control system design methodology are delineated. The 1-CAT approach is presented, along with an analytical technique for carrying out the 1-CAT process. Next, 1-CAT is used to design digital controllers for the proposed Space Based Laser (SBL). Finally, the SBL design is evaluated for dynamical performance, noise rejection, and robustness.
Learning characteristics of a space-time neural network as a tether skiprope observer
NASA Technical Reports Server (NTRS)
Lea, Robert N.; Villarreal, James A.; Jani, Yashvant; Copeland, Charles
1992-01-01
The Software Technology Laboratory at JSC is testing a Space Time Neural Network (STNN) for observing tether oscillations present during retrieval of a tethered satellite. Proper identification of tether oscillations, known as 'skiprope' motion, is vital to safe retrieval of the tethered satellite. Our studies indicate that STNN has certain learning characteristics that must be understood properly to utilize this type of neural network for the tethered satellite problem. We present our findings on the learning characteristics including a learning rate versus momentum performance table.
Semisupervised Support Vector Machines With Tangent Space Intrinsic Manifold Regularization.
Sun, Shiliang; Xie, Xijiong
2016-09-01
Semisupervised learning has been an active research topic in machine learning and data mining. One main reason is that labeling examples is expensive and time-consuming, while there are large numbers of unlabeled examples available in many practical problems. So far, Laplacian regularization has been widely used in semisupervised learning. In this paper, we propose a new regularization method called tangent space intrinsic manifold regularization. It is intrinsic to data manifold and favors linear functions on the manifold. Fundamental elements involved in the formulation of the regularization are local tangent space representations, which are estimated by local principal component analysis, and the connections that relate adjacent tangent spaces. Simultaneously, we explore its application to semisupervised classification and propose two new learning algorithms called tangent space intrinsic manifold regularized support vector machines (TiSVMs) and tangent space intrinsic manifold regularized twin SVMs (TiTSVMs). They effectively integrate the tangent space intrinsic manifold regularization consideration. The optimization of TiSVMs can be solved by a standard quadratic programming, while the optimization of TiTSVMs can be solved by a pair of standard quadratic programmings. The experimental results of semisupervised classification problems show the effectiveness of the proposed semisupervised learning algorithms.
Improving Resource Selection and Scheduling Using Predictions. Chapter 1
NASA Technical Reports Server (NTRS)
Smith, Warren
2003-01-01
The introduction of computational grids has resulted in several new problems in the area of scheduling that can be addressed using predictions. The first problem is selecting where to run an application on the many resources available in a grid. Our approach to help address this problem is to provide predictions of when an application would start to execute if submitted to specific scheduled computer systems. The second problem is gaining simultaneous access to multiple computer systems so that distributed applications can be executed. We help address this problem by investigating how to support advance reservations in local scheduling systems. Our approaches to both of these problems are based on predictions for the execution time of applications on space- shared parallel computers. As a side effect of this work, we also discuss how predictions of application run times can be used to improve scheduling performance.
2007-06-01
xc)−∇2g(x̃c)](x− xc). The second transformation is a space mapping function P that handles the change in variable dimensions (see Bandler et al. [11...17(2):188–217, 2004. 11. Bandler, J. W., Q. Cheng, S. Dakroury, A. S. Mohamed, M.H. Bakr, K. Madsen, J. Søndergaard. “ Space Mapping : The State of
Otorhinolaryngological problems in medical support of space flights.
Bryanov, I I; Matsnev, E I; Yakovleva, I Y
1975-01-01
At present the main trends among the most important problems of otorhinolaryngology in space medicine have become defined as vestibulology, audiology and clinical aspects (prophylaxis, diagnosis and treatment of ENT diseases in flight). The principal result of recent vestibular studies has probably been the establishment of an apparent relationship between the resistance of the vestibular system to adequate ground-based stimulation and tolerance to space flight. The findings of the studies formed the basis for the development of a new system of vestibular selection, as well as demonstrated the usefulness of special vestibular training of astronauts by active and passive methods. In audiology certain urgency is acquired by the problem of noise limitation in space cabins and auditory system reliability prediction for preserving a high work capability in crew members. The hemodynamic changes in weightlessness, as well as the possibility for allergic lesions, create conditions for distorted course of the ENT diseases and vaso-motor disorders. The prophylaxis of aspirations also deserves close attention since the possibilities of their onset increase in weightlessness. The rendering of immediate, timely aid will depend not only on the presence of the necessary medical equipment but also on the ability of the crew members to render the appropriate otorhinolaryngological aid.
NASA Technical Reports Server (NTRS)
Merlin, Peter W.
2006-01-01
The space shuttle orbiter was the first spacecraft designed with the aerodynamic characteristics and in-atmosphere handling qualities of a conventional airplane. In order to evaluate the orbiter's flight control systems and subsonic handling characteristics, a series of flight tests were undertaken at NASA Dryden Flight Research Center in 1977. A modified Boeing 747 Shuttle Carrier Aircraft carried the Enterprise, a prototype orbiter, during eight captive tests to determine how well the two vehicles flew together and to test some of the orbiter s systems. The free-flight phase of the ALT program allowed shuttle pilots to explore the orbiter's low-speed flight and landing characteristics. The Enterprise provided realistic, in-flight simulations of how subsequent space shuttles would be flown at the end of an orbital mission. The fifth free flight, with the Enterprise landing on a concrete runway for the first time, revealed a problem with the space shuttle flight control system that made it susceptible to pilot-induced oscillation, a potentially dangerous control problem. Further research using various aircraft, particularly NASA Dryden's F-8 Digital-Fly-By-Wire testbed, led to correction of the problem before the first Orbital Test Flight.
Continuous-time quantum random walks require discrete space
NASA Astrophysics Data System (ADS)
Manouchehri, K.; Wang, J. B.
2007-11-01
Quantum random walks are shown to have non-intuitive dynamics which makes them an attractive area of study for devising quantum algorithms for long-standing open problems as well as those arising in the field of quantum computing. In the case of continuous-time quantum random walks, such peculiar dynamics can arise from simple evolution operators closely resembling the quantum free-wave propagator. We investigate the divergence of quantum walk dynamics from the free-wave evolution and show that, in order for continuous-time quantum walks to display their characteristic propagation, the state space must be discrete. This behavior rules out many continuous quantum systems as possible candidates for implementing continuous-time quantum random walks.
Shotorban, Babak
2010-04-01
The dynamic least-squares kernel density (LSQKD) model [C. Pantano and B. Shotorban, Phys. Rev. E 76, 066705 (2007)] is used to solve the Fokker-Planck equations. In this model the probability density function (PDF) is approximated by a linear combination of basis functions with unknown parameters whose governing equations are determined by a global least-squares approximation of the PDF in the phase space. In this work basis functions are set to be Gaussian for which the mean, variance, and covariances are governed by a set of partial differential equations (PDEs) or ordinary differential equations (ODEs) depending on what phase-space variables are approximated by Gaussian functions. Three sample problems of univariate double-well potential, bivariate bistable neurodynamical system [G. Deco and D. Martí, Phys. Rev. E 75, 031913 (2007)], and bivariate Brownian particles in a nonuniform gas are studied. The LSQKD is verified for these problems as its results are compared against the results of the method of characteristics in nondiffusive cases and the stochastic particle method in diffusive cases. For the double-well potential problem it is observed that for low to moderate diffusivity the dynamic LSQKD well predicts the stationary PDF for which there is an exact solution. A similar observation is made for the bistable neurodynamical system. In both these problems least-squares approximation is made on all phase-space variables resulting in a set of ODEs with time as the independent variable for the Gaussian function parameters. In the problem of Brownian particles in a nonuniform gas, this approximation is made only for the particle velocity variable leading to a set of PDEs with time and particle position as independent variables. Solving these PDEs, a very good performance by LSQKD is observed for a wide range of diffusivities.
Responsive space: Concept analysis and theoretical framework
NASA Astrophysics Data System (ADS)
Saleh, Joseph H.; Dubos, Gregory F.
2009-08-01
Customers' needs are dynamic and evolve in response to unfolding environmental uncertainties. The ability of a company or an industry to address these changing customers' needs in a timely and cost-effective way is a measure of its responsiveness. In the space industry, a systemic discrepancy exists between the time constants associated with the change of customers' needs, and the response time of the industry in delivering on-orbit solutions to these needs. There are important penalties associated with such delays, and space responsiveness is recognized as a strategic imperative in commercial competitive and military environments. In this paper, we provide a critical assessment of the literature on responsive space and introduce a new multi-disciplinary framework for thinking about and addressing issues of space responsiveness. Our framework advocates three levels of responsiveness: a global industry-wide responsiveness, a local stakeholder responsiveness, and an interactive or inter-stakeholder responsiveness. We introduce and motivate the use of "responsiveness maps" for multiple stakeholders. We then identify "levers of responsiveness": technical spacecraft- and launch-centric, as well as "soft" levers (e.g., acquisition policies) for improving the responsiveness of the space industry. Finally, we propose a series of research questions to aggressively tackle problems associated with space responsiveness.
NASA Astrophysics Data System (ADS)
Field, F.; Goodbun, J.; Watson, V.
Architects have a role to play in interplanetary space that has barely yet been explored. The architectural community is largely unaware of this new territory, for which there is still no agreed method of practice. There is moreover a general confusion, in scientific and related fields, over what architects might actually do there today. Current extra-planetary designs generally fail to explore the dynamic and relational nature of space-time, and often reduce human habitation to a purely functional problem. This is compounded by a crisis over the representation (drawing) of space-time. The present work returns to first principles of architecture in order to realign them with current socio-economic and technological trends surrounding the space industry. What emerges is simultaneously the basis for an ecological space architecture, and the representational strategies necessary to draw it. We explore this approach through a work of design-based research that describes the construction of Ocean; a huge body of water formed by the collision of two asteroids at the Translunar Lagrange Point (L2), that would serve as a site for colonisation, and as a resource to fuel future missions. Ocean is an experimental model for extra-planetary space design and its representation, within the autonomous discipline of architecture.
A Linear-Elasticity Solver for Higher-Order Space-Time Mesh Deformation
NASA Technical Reports Server (NTRS)
Diosady, Laslo T.; Murman, Scott M.
2018-01-01
A linear-elasticity approach is presented for the generation of meshes appropriate for a higher-order space-time discontinuous finite-element method. The equations of linear-elasticity are discretized using a higher-order, spatially-continuous, finite-element method. Given an initial finite-element mesh, and a specified boundary displacement, we solve for the mesh displacements to obtain a higher-order curvilinear mesh. Alternatively, for moving-domain problems we use the linear-elasticity approach to solve for a temporally discontinuous mesh velocity on each time-slab and recover a continuous mesh deformation by integrating the velocity. The applicability of this methodology is presented for several benchmark test cases.
Wills, Thomas A.; Sargent, James D.; Gibbons, Frederick X.; Gerrard, Meg; Stoolmiller, Mike
2009-01-01
The authors tested a theoretical model of how exposure to alcohol cues in movies predicts level of alcohol use (ever use plus ever and recent binge drinking) and alcohol-related problems. A national sample of younger adolescents was interviewed by telephone with 4 repeated assessments spaced at 8-month intervals. A structural equation modeling analysis performed for ever-drinkers at Time 3 (N = 961) indicated that, controlling for a number of covariates, movie alcohol exposure at Time 1 was related to increases in peer alcohol use and adolescent alcohol use at Time 2. Movie exposure had indirect effects to alcohol use and problems at Times 3 and 4 through these pathways, with direct effects to problems from Time 1 rebelliousness and Time 2 movie exposure also found. Prospective risk-promoting effects were also found for alcohol expectancies, peer alcohol use, and availability of alcohol in the home; protective effects were found for mother’s responsiveness and for adolescent’s school performance and self-control. Theoretical and practical implications are discussed. PMID:19290687
Wills, Thomas A; Sargent, James D; Gibbons, Frederick X; Gerrard, Meg; Stoolmiller, Mike
2009-03-01
The authors tested a theoretical model of how exposure to alcohol cues in movies predicts level of alcohol use (ever use plus ever and recent binge drinking) and alcohol-related problems. A national sample of younger adolescents was interviewed by telephone with 4 repeated assessments spaced at 8-month intervals. A structural equation modeling analysis performed for ever-drinkers at Time 3 (N = 961) indicated that, controlling for a number of covariates, movie alcohol exposure at Time 1 was related to increases in peer alcohol use and adolescent alcohol use at Time 2. Movie exposure had indirect effects to alcohol use and problems at Times 3 and 4 through these pathways, with direct effects to problems from Time 1 rebelliousness and Time 2 movie exposure also found. Prospective risk-promoting effects were also found for alcohol expectancies, peer alcohol use, and availability of alcohol in the home; protective effects were found for mother's responsiveness and for adolescent's school performance and self-control. Theoretical and practical implications are discussed. (PsycINFO Database Record (c) 2009 APA, all rights reserved).
Chaos control in delayed phase space constructed by the Takens embedding theory
NASA Astrophysics Data System (ADS)
Hajiloo, R.; Salarieh, H.; Alasty, A.
2018-01-01
In this paper, the problem of chaos control in discrete-time chaotic systems with unknown governing equations and limited measurable states is investigated. Using the time-series of only one measurable state, an algorithm is proposed to stabilize unstable fixed points. The approach consists of three steps: first, using Takens embedding theory, a delayed phase space preserving the topological characteristics of the unknown system is reconstructed. Second, a dynamic model is identified by recursive least squares method to estimate the time-series data in the delayed phase space. Finally, based on the reconstructed model, an appropriate linear delayed feedback controller is obtained for stabilizing unstable fixed points of the system. Controller gains are computed using a systematic approach. The effectiveness of the proposed algorithm is examined by applying it to the generalized hyperchaotic Henon system, prey-predator population map, and the discrete-time Lorenz system.
Space-time wiring specificity supports direction selectivity in the retina
Zlateski, Aleksandar; Lee, Kisuk; Richardson, Mark; Turaga, Srinivas C.; Purcaro, Michael; Balkam, Matthew; Robinson, Amy; Behabadi, Bardia F.; Campos, Michael; Denk, Winfried; Seung, H. Sebastian
2014-01-01
How does the mammalian retina detect motion? This classic problem in visual neuroscience has remained unsolved for 50 years. In search of clues, we reconstructed Off-type starburst amacrine cells (SACs) and bipolar cells (BCs) in serial electron microscopic images with help from EyeWire, an online community of “citizen neuroscientists.” Based on quantitative analyses of contact area and branch depth in the retina, we found evidence that one BC type prefers to wire with a SAC dendrite near the SAC soma, while another BC type prefers to wire far from the soma. The near type is known to lag the far type in time of visual response. A mathematical model shows how such “space-time wiring specificity” could endow SAC dendrites with receptive fields that are oriented in space-time and therefore respond selectively to stimuli that move in the outward direction from the soma. PMID:24805243
Space-time wiring specificity supports direction selectivity in the retina.
Kim, Jinseop S; Greene, Matthew J; Zlateski, Aleksandar; Lee, Kisuk; Richardson, Mark; Turaga, Srinivas C; Purcaro, Michael; Balkam, Matthew; Robinson, Amy; Behabadi, Bardia F; Campos, Michael; Denk, Winfried; Seung, H Sebastian
2014-05-15
How does the mammalian retina detect motion? This classic problem in visual neuroscience has remained unsolved for 50 years. In search of clues, here we reconstruct Off-type starburst amacrine cells (SACs) and bipolar cells (BCs) in serial electron microscopic images with help from EyeWire, an online community of 'citizen neuroscientists'. On the basis of quantitative analyses of contact area and branch depth in the retina, we find evidence that one BC type prefers to wire with a SAC dendrite near the SAC soma, whereas another BC type prefers to wire far from the soma. The near type is known to lag the far type in time of visual response. A mathematical model shows how such 'space-time wiring specificity' could endow SAC dendrites with receptive fields that are oriented in space-time and therefore respond selectively to stimuli that move in the outward direction from the soma.
Optimization of Time-Dependent Particle Tracing Using Tetrahedral Decomposition
NASA Technical Reports Server (NTRS)
Kenwright, David; Lane, David
1995-01-01
An efficient algorithm is presented for computing particle paths, streak lines and time lines in time-dependent flows with moving curvilinear grids. The integration, velocity interpolation and step-size control are all performed in physical space which avoids the need to transform the velocity field into computational space. This leads to higher accuracy because there are no Jacobian matrix approximations or expensive matrix inversions. Integration accuracy is maintained using an adaptive step-size control scheme which is regulated by the path line curvature. The problem of cell-searching, point location and interpolation in physical space is simplified by decomposing hexahedral cells into tetrahedral cells. This enables the point location to be done analytically and substantially faster than with a Newton-Raphson iterative method. Results presented show this algorithm is up to six times faster than particle tracers which operate on hexahedral cells yet produces almost identical particle trajectories.
Spaces and Places for Disrupting Thinking about Inclusive Education in "Hard Times"
ERIC Educational Resources Information Center
Winter, Christine
2012-01-01
This paper sets out to read closely the National Curriculum Statutory Inclusion Statement in England (2007) alongside "Hard Times" to see if Dickens offers any insights into ethical responsibility and conceptualisations of inclusive education. I begin by presenting some of the meanings and associated problems of the term "inclusive…
A Dynamic Programming Approach for Base Station Sleeping in Cellular Networks
NASA Astrophysics Data System (ADS)
Gong, Jie; Zhou, Sheng; Niu, Zhisheng
The energy consumption of the information and communication technology (ICT) industry, which has become a serious problem, is mostly due to the network infrastructure rather than the mobile terminals. In this paper, we focus on reducing the energy consumption of base stations (BSs) by adjusting their working modes (active or sleep). Specifically, the objective is to minimize the energy consumption while satisfying quality of service (QoS, e.g., blocking probability) requirement and, at the same time, avoiding frequent mode switching to reduce signaling and delay overhead. The problem is modeled as a dynamic programming (DP) problem, which is NP-hard in general. Based on cooperation among neighboring BSs, a low-complexity algorithm is proposed to reduce the size of state space as well as that of action space. Simulations demonstrate that, with the proposed algorithm, the active BS pattern well meets the time variation and the non-uniform spatial distribution of system traffic. Moreover, the tradeoff between the energy saving from BS sleeping and the cost of switching is well balanced by the proposed scheme.
Coping strategies during and after spaceflight: Data from retired cosmonauts
NASA Astrophysics Data System (ADS)
Suedfeld, Peter; Brcic, Jelena; Johnson, Phyllis J.; Gushin, Vadim
2015-05-01
Coping is a dynamic physiological and psychological process in response to perceived environmental stress that functions to restore physiological homeostasis and reduce negative affect [1]. Thematic content analysis was employed for references to 13 well-established coping strategies in interviews with 20 retired long-duration male cosmonauts. As in previous research with other space samples [2,3] the retired cosmonauts mentioned Problem-Oriented strategies more frequently than Emotion-Oriented ones. In the present sample, Seeking Social Support, Planful Problem Solving and Endurance/Obedience/Effort were the top three most mentioned coping strategies. Cosmonauts who had spent more than a year in space, compared to those who had spent less than a year, mentioned using Planful Problem Solving more as they recalled their career and retirement. Examining changes over time, spaceflight had a positive effect on Accepting Responsibility. Endurance/Obedience/Effort steadily decreased over time, while we found an inverted-U pattern for Distancing and Self-Control. Additional results in relation to other astronaut samples and the relationship between coping and post-flight growth are discussed.
NASA'S Water Resources Element Within the Applied Sciences Program
NASA Technical Reports Server (NTRS)
Toll, David; Doorn, Bradley; Engman, Edwin
2011-01-01
The NASA Earth Systems Division has the primary responsibility for the Applied Science Program and the objective to accelerate the use of NASA science results in applications to help solve problems important to society and the economy. The primary goal of the NASA Applied Science Program is to improve future and current operational systems by infusing them with scientific knowledge of the Earth system gained through space-based observation, assimilation of new observations, and development and deployment of enabling technologies, systems, and capabilities. This paper discusses major problems facing water resources managers, including having timely and accurate data to drive their decision support tools. It then describes how NASA's science and space based satellites may be used to overcome this problem. Opportunities for the water resources community to participate in NASA's Water Resources Applications Program are described.
Field quantization and squeezed states generation in resonators with time-dependent parameters
NASA Technical Reports Server (NTRS)
Dodonov, V. V.; Klimov, A. B.; Nikonov, D. E.
1992-01-01
The problem of electromagnetic field quantization is usually considered in textbooks under the assumption that the field occupies some empty box. The case when a nonuniform time-dependent dielectric medium is confined in some space region with time-dependent boundaries is studied. The basis of the subsequent consideration is the system of Maxwell's equations in linear passive time-dependent dielectric and magnetic medium without sources.
GrammarViz 3.0: Interactive Discovery of Variable-Length Time Series Patterns
Senin, Pavel; Lin, Jessica; Wang, Xing; ...
2018-02-23
The problems of recurrent and anomalous pattern discovery in time series, e.g., motifs and discords, respectively, have received a lot of attention from researchers in the past decade. However, since the pattern search space is usually intractable, most existing detection algorithms require that the patterns have discriminative characteristics and have its length known in advance and provided as input, which is an unreasonable requirement for many real-world problems. In addition, patterns of similar structure, but of different lengths may co-exist in a time series. In order to address these issues, we have developed algorithms for variable-length time series pattern discoverymore » that are based on symbolic discretization and grammar inference—two techniques whose combination enables the structured reduction of the search space and discovery of the candidate patterns in linear time. In this work, we present GrammarViz 3.0—a software package that provides implementations of proposed algorithms and graphical user interface for interactive variable-length time series pattern discovery. The current version of the software provides an alternative grammar inference algorithm that improves the time series motif discovery workflow, and introduces an experimental procedure for automated discretization parameter selection that builds upon the minimum cardinality maximum cover principle and aids the time series recurrent and anomalous pattern discovery.« less
GrammarViz 3.0: Interactive Discovery of Variable-Length Time Series Patterns
DOE Office of Scientific and Technical Information (OSTI.GOV)
Senin, Pavel; Lin, Jessica; Wang, Xing
The problems of recurrent and anomalous pattern discovery in time series, e.g., motifs and discords, respectively, have received a lot of attention from researchers in the past decade. However, since the pattern search space is usually intractable, most existing detection algorithms require that the patterns have discriminative characteristics and have its length known in advance and provided as input, which is an unreasonable requirement for many real-world problems. In addition, patterns of similar structure, but of different lengths may co-exist in a time series. In order to address these issues, we have developed algorithms for variable-length time series pattern discoverymore » that are based on symbolic discretization and grammar inference—two techniques whose combination enables the structured reduction of the search space and discovery of the candidate patterns in linear time. In this work, we present GrammarViz 3.0—a software package that provides implementations of proposed algorithms and graphical user interface for interactive variable-length time series pattern discovery. The current version of the software provides an alternative grammar inference algorithm that improves the time series motif discovery workflow, and introduces an experimental procedure for automated discretization parameter selection that builds upon the minimum cardinality maximum cover principle and aids the time series recurrent and anomalous pattern discovery.« less
Pre-vector variational inequality
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lin, Lai-Jiu
1994-12-31
Let X be a Hausdorff topological vector space, (Y, D) be an ordered Hausdorff topological vector space ordered by convex cone D. Let L(X, Y) be the space of all bounded linear operator, E {improper_subset} X be a nonempty set, T : E {yields} L(X, Y), {eta} : E {times} E {yields} E be functions. For x, y {element_of} Y, we denote x {not_lt} y if y - x intD, where intD is the interior of D. We consider the following two problems: Find x {element_of} E such that < T(x), {eta}(y, x) > {not_lt} 0 for all y {element_of}more » E and find x {element_of} E, < T(x), {eta}(y, x) > {not_gt} 0 for all y {element_of} E and < T(x), {eta}(y, x) >{element_of} C{sub p}{sup w+} = {l_brace} {element_of} L(X, Y) {vert_bar}< l, {eta}(x, 0) >{not_lt} 0 for all x {element_of} E{r_brace} where < T(x), y > denotes linear operator T(x) at y, that is T(x), (y). We called Pre-VVIP the Pre-vector variational inequality problem and Pre-VCP complementary problem. If X = R{sup n}, Y = R, D = R{sub +} {eta}(y, x) = y - x, then our problem is the well-known variational inequality first studies by Hartman and Stampacchia. If Y = R, D = R{sub +}, {eta}(y, x) = y - x, our problem is the variational problem in infinite dimensional space. In this research, we impose different condition on T(x), {eta}, X, and < T(x), {eta}(y, x) > and investigate the existences theorem of these problems. As an application of one of our results, we establish the existence theorem of weak minimum of the problem. (P) V - min f(x) subject to x {element_of} E where f : X {yields} Y si a Frechet differentiable invex function.« less
Space debris tracking at San Fernando laser station
NASA Astrophysics Data System (ADS)
Catalán, M.; Quijano, M.; Pazos, A.; Martín Davila, J.; Cortina, L. M.
2016-12-01
For years to come space debris will be a major issue for society. It has a negative impact on active artificial satellites, having implications for future missions. Tracking space debris as accurately as possible is the first step towards controlling this problem, yet it presents a challenge for science. The main limitation is the relatively low accuracy of the methods used to date for tracking these objects. Clearly, improving the predicted orbit accuracy is crucial (avoiding unnecessary anti-collision maneuvers). A new field of research was recently instituted by our satellite laser ranging station: tracking decommissioned artificial satellites equipped with retroreflectors. To this end we work in conjunction with international space agencies which provide increasing attention to this problem. We thus proposed to share our time-schedule of use of the satellite laser ranging station for obtaining data that would make orbital element predictions far more accurate (meter accuracy), whilst maintaining our tracking routines for active satellites. This manuscript reports on the actions carried out so far.
Effective degrees of freedom of a random walk on a fractal.
Balankin, Alexander S
2015-12-01
We argue that a non-Markovian random walk on a fractal can be treated as a Markovian process in a fractional dimensional space with a suitable metric. This allows us to define the fractional dimensional space allied to the fractal as the ν-dimensional space F(ν) equipped with the metric induced by the fractal topology. The relation between the number of effective spatial degrees of freedom of walkers on the fractal (ν) and fractal dimensionalities is deduced. The intrinsic time of random walk in F(ν) is inferred. The Laplacian operator in F(ν) is constructed. This allows us to map physical problems on fractals into the corresponding problems in F(ν). In this way, essential features of physics on fractals are revealed. Particularly, subdiffusion on path-connected fractals is elucidated. The Coulomb potential of a point charge on a fractal embedded in the Euclidean space is derived. Intriguing attributes of some types of fractals are highlighted.
Modeling volatility using state space models.
Timmer, J; Weigend, A S
1997-08-01
In time series problems, noise can be divided into two categories: dynamic noise which drives the process, and observational noise which is added in the measurement process, but does not influence future values of the system. In this framework, we show that empirical volatilities (the squared relative returns of prices) exhibit a significant amount of observational noise. To model and predict their time evolution adequately, we estimate state space models that explicitly include observational noise. We obtain relaxation times for shocks in the logarithm of volatility ranging from three weeks (for foreign exchange) to three to five months (for stock indices). In most cases, a two-dimensional hidden state is required to yield residuals that are consistent with white noise. We compare these results with ordinary autoregressive models (without a hidden state) and find that autoregressive models underestimate the relaxation times by about two orders of magnitude since they do not distinguish between observational and dynamic noise. This new interpretation of the dynamics of volatility in terms of relaxators in a state space model carries over to stochastic volatility models and to GARCH models, and is useful for several problems in finance, including risk management and the pricing of derivative securities. Data sets used: Olsen & Associates high frequency DEM/USD foreign exchange rates (8 years). Nikkei 225 index (40 years). Dow Jones Industrial Average (25 years).
Risk-Sensitive Control of Pure Jump Process on Countable Space with Near Monotone Cost
DOE Office of Scientific and Technical Information (OSTI.GOV)
Suresh Kumar, K., E-mail: suresh@math.iitb.ac.in; Pal, Chandan, E-mail: cpal@math.iitb.ac.in
2013-12-15
In this article, we study risk-sensitive control problem with controlled continuous time pure jump process on a countable space as state dynamics. We prove multiplicative dynamic programming principle, elliptic and parabolic Harnack’s inequalities. Using the multiplicative dynamic programing principle and the Harnack’s inequalities, we prove the existence and a characterization of optimal risk-sensitive control under the near monotone condition.
National Space Transportation System (NSTS) technology needs
NASA Technical Reports Server (NTRS)
Winterhalter, David L.; Ulrich, Kimberly K.
1990-01-01
The National Space Transportation System (NSTS) is one of the Nation's most valuable resources, providing manned transportation to and from space in support of payloads and scientific research. The NSTS program is currently faced with the problem of hardware obsolescence, which could result in unacceptable schedule and cost impacts to the flight program. Obsolescence problems occur because certain components are no longer being manufactured or repair turnaround time is excessive. In order to achieve a long-term, reliable transportation system that can support manned access to space through 2010 and beyond, NASA must develop a strategic plan for a phased implementation of enhancements which will satisfy this long-term goal. The NSTS program has initiated the Assured Shuttle Availability (ASA) project with the following objectives: eliminate hardware obsolescence in critical areas, increase reliability and safety of the vehicle, decrease operational costs and turnaround time, and improve operational capability. The strategy for ASA will be to first meet the mandatory needs - keep the Shuttle flying. Non-mandatory changes that will improve operational capability and enhance performance will then be considered if funding is adequate. Upgrade packages should be developed to install within designated inspection periods, grouped in a systematic approach to reduce cost and schedule impacts, and allow the capability to provide a Block 2 Shuttle (Phase 3).
Landau problem with time dependent mass in time dependent electric and harmonic background fields
NASA Astrophysics Data System (ADS)
Lawson, Latévi M.; Avossevou, Gabriel Y. H.
2018-04-01
The spectrum of a Hamiltonian describing the dynamics of a Landau particle with time-dependent mass and frequency undergoing the influence of a uniform time-dependent electric field is obtained. The configuration space wave function of the model is expressed in terms of the generalised Laguerre polynomials. To diagonalize the time-dependent Hamiltonian, we employ the Lewis-Riesenfeld method of invariants. To this end, we introduce a unitary transformation in the framework of the algebraic formalism to construct the invariant operator of the system and then to obtain the exact solution of the Hamiltonian. We recover the solutions of the ordinary Landau problem in the absence of the electric and harmonic fields for a constant particle mass.
A direct method for nonlinear ill-posed problems
NASA Astrophysics Data System (ADS)
Lakhal, A.
2018-02-01
We propose a direct method for solving nonlinear ill-posed problems in Banach-spaces. The method is based on a stable inversion formula we explicitly compute by applying techniques for analytic functions. Furthermore, we investigate the convergence and stability of the method and prove that the derived noniterative algorithm is a regularization. The inversion formula provides a systematic sensitivity analysis. The approach is applicable to a wide range of nonlinear ill-posed problems. We test the algorithm on a nonlinear problem of travel-time inversion in seismic tomography. Numerical results illustrate the robustness and efficiency of the algorithm.
Efficient Fourier-based algorithms for time-periodic unsteady problems
NASA Astrophysics Data System (ADS)
Gopinath, Arathi Kamath
2007-12-01
This dissertation work proposes two algorithms for the simulation of time-periodic unsteady problems via the solution of Unsteady Reynolds-Averaged Navier-Stokes (URANS) equations. These algorithms use a Fourier representation in time and hence solve for the periodic state directly without resolving transients (which consume most of the resources in a time-accurate scheme). In contrast to conventional Fourier-based techniques which solve the governing equations in frequency space, the new algorithms perform all the calculations in the time domain, and hence require minimal modifications to an existing solver. The complete space-time solution is obtained by iterating in a fifth pseudo-time dimension. Various time-periodic problems such as helicopter rotors, wind turbines, turbomachinery and flapping-wings can be simulated using the Time Spectral method. The algorithm is first validated using pitching airfoil/wing test cases. The method is further extended to turbomachinery problems, and computational results verified by comparison with a time-accurate calculation. The technique can be very memory intensive for large problems, since the solution is computed (and hence stored) simultaneously at all time levels. Often, the blade counts of a turbomachine are rescaled such that a periodic fraction of the annulus can be solved. This approximation enables the solution to be obtained at a fraction of the cost of a full-scale time-accurate solution. For a viscous computation over a three-dimensional single-stage rescaled compressor, an order of magnitude savings is achieved. The second algorithm, the reduced-order Harmonic Balance method is applicable only to turbomachinery flows, and offers even larger computational savings than the Time Spectral method. It simulates the true geometry of the turbomachine using only one blade passage per blade row as the computational domain. In each blade row of the turbomachine, only the dominant frequencies are resolved, namely, combinations of neighbor's blade passing. An appropriate set of frequencies can be chosen by the analyst/designer based on a trade-off between accuracy and computational resources available. A cost comparison with a time-accurate computation for an Euler calculation on a two-dimensional multi-stage compressor obtained an order of magnitude savings, and a RANS calculation on a three-dimensional single-stage compressor achieved two orders of magnitude savings, with comparable accuracy.
Canonical Gravity, Non-Inertial Frames, Relativistic Metrology and Dark Matter
NASA Astrophysics Data System (ADS)
Lusanna, Luca
Clock synchronization leads to the definition of instantaneous 3-spaces (to be used as Cauchy surfaces) in non-inertial frames, the only ones allowed by the equivalence principle. ADM canonical tetrad gravity in asymptotically Minkowskian space-times can be described in this framework. This allows to find the York canonical basis in which the inertial (gauge) and tidal (physical) degrees of freedom of the gravitational field can be identified. A Post-Minkowskian linearization with respect to the asymptotic Minkowski metric (asymptotic background) allows to solve the Dirac constraints in non-harmonic 3-orthogonal gauges and to find non-harmonic TT gravitational waves. The inertial gauge variable York time (the trace of the extrinsic curvature of the 3-space) describes the general relativistic freedom in clock synchronization. After a digression on the gauge problem in general relativity and its connection with relativistic metrology, it is shown that dark matter, whose experimental signatures are the rotation curves and the mass of galaxies, may be described (at least partially) as an inertial relativistic effect (absent in Newtonian gravity) connected with the York time, namely with the non-Euclidean nature of 3-spaces as 3-sub-manifolds of space-time.
Evolution inclusions governed by the difference of two subdifferentials in reflexive Banach spaces
NASA Astrophysics Data System (ADS)
Akagi, Goro; Ôtani, Mitsuharu
The existence of strong solutions of Cauchy problem for the following evolution equation du(t)/dt+∂ϕ1(u(t))-∂ϕ2(u(t))∋f(t) is considered in a real reflexive Banach space V, where ∂ϕ1 and ∂ϕ2 are subdifferential operators from V into its dual V*. The study for this type of problems has been done by several authors in the Hilbert space setting. The scope of our study is extended to the V- V* setting. The main tool employed here is a certain approximation argument in a Hilbert space and for this purpose we need to assume that there exists a Hilbert space H such that V⊂H≡H*⊂V* with densely defined continuous injections. The applicability of our abstract framework will be exemplified in discussing the existence of solutions for the nonlinear heat equation: ut(x,t)-Δpu(x,t)-|u|u(x,t)=f(x,t), x∈Ω, t>0, u|=0, where Ω is a bounded domain in RN. In particular, the existence of local (in time) weak solution is shown under the subcritical growth condition q
Quantum mechanics of hyperbolic orbits in the Kepler problem
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rauh, Alexander; Parisi, Juergen
2011-04-15
The problem of deriving macroscopic properties from the Hamiltonian of the hydrogen atom is resumed by extending previous results in the literature, which predicted elliptic orbits, into the region of hyperbolic orbits. As a main tool, coherent states of the harmonic oscillator are used which are continued to imaginary frequencies. The Kustaanheimo-Stiefel (KS) map is applied to transform the original configuration space into the product space of four harmonic oscillators with a constraint. The relation derived between real time and oscillator (pseudo) time includes quantum corrections. In the limit ({h_bar}/2{pi}){yields}0, the time-dependent mean values of position and velocity describe themore » classical motion on a hyperbola and a circular hodograph, respectively. Moreover, the connection between pseudotime and real time comes out in analogy to Kepler's equation for elliptic orbits. The mean-square-root deviations of position and velocity components behave similarly in time to the corresponding ones of a spreading Gaussian wave packet in free space. To check the approximate treatment of the constraint, its contribution to the mean energy is determined with the result that it is negligible except for energy values close to the parabolic orbit with eccentricity equal to 1. It is inevitable to introduce a suitable scalar product in R{sup 4} which makes both the transformed Hamiltonian and the velocity operators Hermitian. An elementary necessary criterion is given for the energy interval where the constraint can be approximated by averaging.« less
Approximate matching of regular expressions.
Myers, E W; Miller, W
1989-01-01
Given a sequence A and regular expression R, the approximate regular expression matching problem is to find a sequence matching R whose optimal alignment with A is the highest scoring of all such sequences. This paper develops an algorithm to solve the problem in time O(MN), where M and N are the lengths of A and R. Thus, the time requirement is asymptotically no worse than for the simpler problem of aligning two fixed sequences. Our method is superior to an earlier algorithm by Wagner and Seiferas in several ways. First, it treats real-valued costs, in addition to integer costs, with no loss of asymptotic efficiency. Second, it requires only O(N) space to deliver just the score of the best alignment. Finally, its structure permits implementation techniques that make it extremely fast in practice. We extend the method to accommodate gap penalties, as required for typical applications in molecular biology, and further refine it to search for sub-strings of A that strongly align with a sequence in R, as required for typical data base searches. We also show how to deliver an optimal alignment between A and R in only O(N + log M) space using O(MN log M) time. Finally, an O(MN(M + N) + N2log N) time algorithm is presented for alignment scoring schemes where the cost of a gap is an arbitrary increasing function of its length.
NASA Astrophysics Data System (ADS)
Shakeri, Nadim; Jalili, Saeed; Ahmadi, Vahid; Rasoulzadeh Zali, Aref; Goliaei, Sama
2015-01-01
The problem of finding the Hamiltonian path in a graph, or deciding whether a graph has a Hamiltonian path or not, is an NP-complete problem. No exact solution has been found yet, to solve this problem using polynomial amount of time and space. In this paper, we propose a two dimensional (2-D) optical architecture based on optical electronic devices such as micro ring resonators, optical circulators and MEMS based mirror (MEMS-M) to solve the Hamiltonian Path Problem, for undirected graphs in linear time. It uses a heuristic algorithm and employs n+1 different wavelengths of a light ray, to check whether a Hamiltonian path exists or not on a graph with n vertices. Then if a Hamiltonian path exists, it reports the path. The device complexity of the proposed architecture is O(n2).
Indexing a sequence for mapping reads with a single mismatch.
Crochemore, Maxime; Langiu, Alessio; Rahman, M Sohel
2014-05-28
Mapping reads against a genome sequence is an interesting and useful problem in computational molecular biology and bioinformatics. In this paper, we focus on the problem of indexing a sequence for mapping reads with a single mismatch. We first focus on a simpler problem where the length of the pattern is given beforehand during the data structure construction. This version of the problem is interesting in its own right in the context of the next generation sequencing. In the sequel, we show how to solve the more general problem. In both cases, our algorithm can construct an efficient data structure in O(n log(1+ε) n) time and space and can answer subsequent queries in O(m log log n + K) time. Here, n is the length of the sequence, m is the length of the read, 0<ε<1 and is the optimal output size.
NASA Technical Reports Server (NTRS)
Biswas, Rupak
2018-01-01
Quantum computing promises an unprecedented ability to solve intractable problems by harnessing quantum mechanical effects such as tunneling, superposition, and entanglement. The Quantum Artificial Intelligence Laboratory (QuAIL) at NASA Ames Research Center is the space agency's primary facility for conducting research and development in quantum information sciences. QuAIL conducts fundamental research in quantum physics but also explores how best to exploit and apply this disruptive technology to enable NASA missions in aeronautics, Earth and space sciences, and space exploration. At the same time, machine learning has become a major focus in computer science and captured the imagination of the public as a panacea to myriad big data problems. In this talk, we will discuss how classical machine learning can take advantage of quantum computing to significantly improve its effectiveness. Although we illustrate this concept on a quantum annealer, other quantum platforms could be used as well. If explored fully and implemented efficiently, quantum machine learning could greatly accelerate a wide range of tasks leading to new technologies and discoveries that will significantly change the way we solve real-world problems.
Local Approximation and Hierarchical Methods for Stochastic Optimization
NASA Astrophysics Data System (ADS)
Cheng, Bolong
In this thesis, we present local and hierarchical approximation methods for two classes of stochastic optimization problems: optimal learning and Markov decision processes. For the optimal learning problem class, we introduce a locally linear model with radial basis function for estimating the posterior mean of the unknown objective function. The method uses a compact representation of the function which avoids storing the entire history, as is typically required by nonparametric methods. We derive a knowledge gradient policy with the locally parametric model, which maximizes the expected value of information. We show the policy is asymptotically optimal in theory, and experimental works suggests that the method can reliably find the optimal solution on a range of test functions. For the Markov decision processes problem class, we are motivated by an application where we want to co-optimize a battery for multiple revenue, in particular energy arbitrage and frequency regulation. The nature of this problem requires the battery to make charging and discharging decisions at different time scales while accounting for the stochastic information such as load demand, electricity prices, and regulation signals. Computing the exact optimal policy becomes intractable due to the large state space and the number of time steps. We propose two methods to circumvent the computation bottleneck. First, we propose a nested MDP model that structure the co-optimization problem into smaller sub-problems with reduced state space. This new model allows us to understand how the battery behaves down to the two-second dynamics (that of the frequency regulation market). Second, we introduce a low-rank value function approximation for backward dynamic programming. This new method only requires computing the exact value function for a small subset of the state space and approximate the entire value function via low-rank matrix completion. We test these methods on historical price data from the PJM Interconnect and show that it outperforms the baseline approach used in the industry.
An Insightful Problem Involving the Electromagnetic Radiation from a Pair of Dipoles
ERIC Educational Resources Information Center
Smith, Glenn S.
2010-01-01
The time-average power radiated by a pair of infinitesimal dipoles is examined as their spacing is varied. The results elucidate the effect of the interaction of the dipoles on their radiation. (Contains 4 figures.)
ERIC Educational Resources Information Center
Maxon, Hazel Carter; Negron, Jaime
1977-01-01
Two full-time university bookstores, with three satellites helping during rush period, serve the Howard students and faculty. Solutions to problems of space, acquiring used books, and communications with faculty members are discussed, and the successful retailing of black studies books is described. (LBH)
Ingebritsen, Steven E.; Gleeson, Tom
2017-01-01
Permeability is the dominant parameter in most hydrogeologic studies. There is abundant evidence for dynamic variations in permeability in time as well as space, and throughout the crust. Whether this dynamic behavior should be included in quantitative models depends on the problem at hand.
ORES - Objective Referenced Evaluation in Science.
ERIC Educational Resources Information Center
Shaw, Terry
Science process skills considered important in making decisions and solving problems include: observing, classifying, measuring, using numbers, using space/time relationships, communicating, predicting, inferring, manipulating variables, making operational definitions, forming hypotheses, interpreting data, and experimenting. This 60-item test,…
View of the SRB problems with Challenger after launch
NASA Technical Reports Server (NTRS)
1986-01-01
51-L investigation at time 73.200 seconds, Flash from region between orbiter and external tank (ET) liquid hydrogen (LH2) tank, USC-10. Kennedy Space Center alternative photo number is 108-KSC-386C-560/53.
View of the SRB problems with Challenger after launch
NASA Technical Reports Server (NTRS)
1986-01-01
51-L investigation at time 59.249 seconds, Well defined intense plume on side of right hand solid rocket booster -Z direction, MIGOR/USC-10. Kennedy Space Center alternative photo number is 108-KSC-386C-648/52.
Space and Time Partitioning with Hardware Support for Space Applications
NASA Astrophysics Data System (ADS)
Pinto, S.; Tavares, A.; Montenegro, S.
2016-08-01
Complex and critical systems like airplanes and spacecraft implement a very fast growing amount of functions. Typically, those systems were implemented with fully federated architectures, but the number and complexity of desired functions of todays systems led aerospace industry to follow another strategy. Integrated Modular Avionics (IMA) arose as an attractive approach for consolidation, by combining several applications into one single generic computing resource. Current approach goes towards higher integration provided by space and time partitioning (STP) of system virtualization. The problem is existent virtualization solutions are not ready to fully provide what the future of aerospace are demanding: performance, flexibility, safety, security while simultaneously containing Size, Weight, Power and Cost (SWaP-C).This work describes a real time hypervisor for space applications assisted by commercial off-the-shell (COTS) hardware. ARM TrustZone technology is exploited to implement a secure virtualization solution with low overhead and low memory footprint. This is demonstrated by running multiple guest partitions of RODOS operating system on a Xilinx Zynq platform.
On a model of electromagnetic field propagation in ferroelectric media
NASA Astrophysics Data System (ADS)
Picard, Rainer
2007-04-01
The Maxwell system in an anisotropic, inhomogeneous medium with non-linear memory effect produced by a Maxwell type system for the polarization is investigated under low regularity assumptions on data and domain. The particular form of memory in the system is motivated by a model for electromagnetic wave propagation in ferromagnetic materials suggested by Greenberg, MacCamy and Coffman [J.M. Greenberg, R.C. MacCamy, C.V. Coffman, On the long-time behavior of ferroelectric systems, Phys. D 134 (1999) 362-383]. To avoid unnecessary regularity requirements the problem is approached as a system of space-time operator equation in the framework of extrapolation spaces (Sobolev lattices), a theoretical framework developed in [R. Picard, Evolution equations as space-time operator equations, Math. Anal. Appl. 173 (2) (1993) 436-458; R. Picard, Evolution equations as operator equations in lattices of Hilbert spaces, Glasnik Mat. 35 (2000) 111-136]. A solution theory for a large class of ferromagnetic materials confined to an arbitrary open set (with suitably generalized boundary conditions) is obtained.
Artificial Neural Network Based Mission Planning Mechanism for Spacecraft
NASA Astrophysics Data System (ADS)
Li, Zhaoyu; Xu, Rui; Cui, Pingyuan; Zhu, Shengying
2018-04-01
The ability to plan and react fast in dynamic space environments is central to intelligent behavior of spacecraft. For space and robotic applications, many planners have been used. But it is difficult to encode the domain knowledge and directly use existing techniques such as heuristic to improve the performance of the application systems. Therefore, regarding planning as an advanced control problem, this paper first proposes an autonomous mission planning and action selection mechanism through a multiple layer perceptron neural network approach to select actions in planning process and improve efficiency. To prove the availability and effectiveness, we use autonomous mission planning problems of the spacecraft, which is a sophisticated system with complex subsystems and constraints as an example. Simulation results have shown that artificial neural networks (ANNs) are usable for planning problems. Compared with the existing planning method in EUROPA, the mechanism using ANNs is more efficient and can guarantee stable performance. Therefore, the mechanism proposed in this paper is more suitable for planning problems of spacecraft that require real time and stability.
On computing the global time-optimal motions of robotic manipulators in the presence of obstacles
NASA Technical Reports Server (NTRS)
Shiller, Zvi; Dubowsky, Steven
1991-01-01
A method for computing the time-optimal motions of robotic manipulators is presented that considers the nonlinear manipulator dynamics, actuator constraints, joint limits, and obstacles. The optimization problem is reduced to a search for the time-optimal path in the n-dimensional position space. A small set of near-optimal paths is first efficiently selected from a grid, using a branch and bound search and a series of lower bound estimates on the traveling time along a given path. These paths are further optimized with a local path optimization to yield the global optimal solution. Obstacles are considered by eliminating the collision points from the tessellated space and by adding a penalty function to the motion time in the local optimization. The computational efficiency of the method stems from the reduced dimensionality of the searched spaced and from combining the grid search with a local optimization. The method is demonstrated in several examples for two- and six-degree-of-freedom manipulators with obstacles.
NASA Astrophysics Data System (ADS)
Schleich, Wolfgang P.
2001-04-01
Quantum Optics in Phase Space provides a concise introduction to the rapidly moving field of quantum optics from the point of view of phase space. Modern in style and didactically skillful, Quantum Optics in Phase Space prepares students for their own research by presenting detailed derivations, many illustrations and a large set of workable problems at the end of each chapter. Often, the theoretical treatments are accompanied by the corresponding experiments. An exhaustive list of references provides a guide to the literature. Quantum Optics in Phase Space also serves advanced researchers as a comprehensive reference book. Starting with an extensive review of the experiments that define quantum optics and a brief summary of the foundations of quantum mechanics the author Wolfgang P. Schleich illustrates the properties of quantum states with the help of the Wigner phase space distribution function. His description of waves ala WKB connects semi-classical phase space with the Berry phase. These semi-classical techniques provide deeper insight into the timely topics of wave packet dynamics, fractional revivals and the Talbot effect. Whereas the first half of the book deals with mechanical oscillators such as ions in a trap or atoms in a standing wave the second half addresses problems where the quantization of the radiation field is of importance. Such topics extensively discussed include optical interferometry, the atom-field interaction, quantum state preparation and measurement, entanglement, decoherence, the one-atom maser and atom optics in quantized light fields. Quantum Optics in Phase Space presents the subject of quantum optics as transparently as possible. Giving wide-ranging references, it enables students to study and solve problems with modern scientific literature. The result is a remarkably concise yet comprehensive and accessible text- and reference book - an inspiring source of information and insight for students, teachers and researchers alike.
Evaluation of new collision-pair selection models in DSMC
NASA Astrophysics Data System (ADS)
Akhlaghi, Hassan; Roohi, Ehsan
2017-10-01
The current paper investigates new collision-pair selection procedures in a direct simulation Monte Carlo (DSMC) method. Collision partner selection based on the random procedure from nearest neighbor particles and deterministic selection of nearest neighbor particles have already been introduced as schemes that provide accurate results in a wide range of problems. In the current research, new collision-pair selections based on the time spacing and direction of the relative movement of particles are introduced and evaluated. Comparisons between the new and existing algorithms are made considering appropriate test cases including fluctuations in homogeneous gas, 2D equilibrium flow, and Fourier flow problem. Distribution functions for number of particles and collisions in cell, velocity components, and collisional parameters (collision separation, time spacing, relative velocity, and the angle between relative movements of particles) are investigated and compared with existing analytical relations for each model. The capability of each model in the prediction of the heat flux in the Fourier problem at different cell numbers, numbers of particles, and time steps is examined. For new and existing collision-pair selection schemes, the effect of an alternative formula for the number of collision-pair selections and avoiding repetitive collisions are investigated via the prediction of the Fourier heat flux. The simulation results demonstrate the advantages and weaknesses of each model in different test cases.
Engineering calculations for communications satellite systems planning
NASA Technical Reports Server (NTRS)
Reilly, C. H.; Levis, C. A.; Mount-Campbell, C.; Gonsalvez, D. J.; Wang, C. W.; Yamamura, Y.
1985-01-01
Computer-based techniques for optimizing communications-satellite orbit and frequency assignments are discussed. A gradient-search code was tested against a BSS scenario derived from the RARC-83 data. Improvement was obtained, but each iteration requires about 50 minutes of IBM-3081 CPU time. Gradient-search experiments on a small FSS test problem, consisting of a single service area served by 8 satellites, showed quickest convergence when the satellites were all initially placed near the center of the available orbital arc with moderate spacing. A transformation technique is proposed for investigating the surface topography of the objective function used in the gradient-search method. A new synthesis approach is based on transforming single-entry interference constraints into corresponding constraints on satellite spacings. These constraints are used with linear objective functions to formulate the co-channel orbital assignment task as a linear-programming (LP) problem or mixed integer programming (MIP) problem. Globally optimal solutions are always found with the MIP problems, but not necessarily with the LP problems. The MIP solutions can be used to evaluate the quality of the LP solutions. The initial results are very encouraging.
Real-Time Operation of the International Space Station
NASA Astrophysics Data System (ADS)
Suffredini, M. T.
2002-01-01
The International Space Station is on orbit and real-time operations are well underway. Along with the assembly challenges of building and operating the International Space Station , scientific activities are also underway. Flight control teams in three countries are working together as a team to plan, coordinate and command the systems on the International Space Station.Preparations are being made to add the additional International Partner elements including their operations teams and facilities. By October 2002, six Expedition crews will have lived on the International Space Station. Management of real-time operations has been key to these achievements. This includes the activities of ground teams in control centers around the world as well as the crew on orbit. Real-time planning is constantly challenged with balancing the requirements and setting the priorities for the assembly, maintenance, science and crew health functions on the International Space Station. It requires integrating the Shuttle, Soyuz and Progress requirements with the Station. It is also necessary to be able to respond in case of on-orbit anomalies and to set plans and commands in place to ensure the continues safe operation of the Station. Bringing together the International Partner operations teams has been challenging and intensely rewarding. Utilization of the assets of each partner has resulted in efficient solutions to problems. This paper will describe the management of the major real-time operations processes, significant achievements, and future challenges.
The scientific data acquisition system of the GAMMA-400 space project
NASA Astrophysics Data System (ADS)
Bobkov, S. G.; Serdin, O. V.; Gorbunov, M. S.; Arkhangelskiy, A. I.; Topchiev, N. P.
2016-02-01
The description of scientific data acquisition system (SDAS) designed by SRISA for the GAMMA-400 space project is presented. We consider the problem of different level electronics unification: the set of reliable fault-tolerant integrated circuits fabricated on Silicon-on-Insulator 0.25 mkm CMOS technology and the high-speed interfaces and reliable modules used in the space instruments. The characteristics of reliable fault-tolerant very large scale integration (VLSI) technology designed by SRISA for the developing of computation systems for space applications are considered. The scalable net structure of SDAS based on Serial RapidIO interface including real-time operating system BAGET is described too.
NASA Technical Reports Server (NTRS)
Mccray, Richard; Ostriker, Jeremiah P.; Acton, Loren W.; Bahcall, Neta A.; Bless, Robert C.; Brown, Robert A.; Burbidge, Geoffrey; Burke, Bernard F.; Clark, George W.; Cordova, France A.
1991-01-01
Recommendations are given regarding National Science Foundation (NSF) astronomy programs and the NASA Space Astrophysics program. The role of ground based astronomy is reviewed. The role of National Optical Astronomy Observatories (NOAO) in ground-based night-time astronomical research is discussed. An enhanced Explored Program, costs and management of small and moderate space programs, the role of astrophysics within NASA's space exploration initiative, suborbital and airborne astronomical research, the problems of the Hubble Space Telescope, and astronomy education are discussed. Also covered are policy issues related to the role of science advisory committees, international cooperation and competition, archiving and distribution of astronomical data, and multi-wavelength observations of variable sources.
Project Golden Gate: towards real-time Java in space missions
NASA Technical Reports Server (NTRS)
Dvorak, Daniel; Bollella, Greg; Canham, Tim; Carson, Vanessa; Champlin, Virgil; Giovannoni, Brian; Indictor, Mark; Meyer, Kenny; Murray, Alex; Reinholtz, Kirk
2004-01-01
This paper describes the problem domain and our experimentation with the first commercial implementation of the Real Time Specification for Java. The two main issues explored in this report are: (1) the effect of RTSJ's non-heap memory on the programming model, and (2) performance benchmarking of RTSJ/Linux relative to C++/VxWorks.
Decay of the compressible magneto-micropolar fluids
NASA Astrophysics Data System (ADS)
Zhang, Peixin
2018-02-01
This paper considers the large-time behavior of solutions to the Cauchy problem on the compressible magneto-micropolar fluid system under small perturbation in regular Sobolev space. Based on the time-weighted energy estimate, the asymptotic stability of the steady state with the strictly positive constant density, vanishing velocity, micro-rotational velocity, and magnetic field is established.
Random walk in degree space and the time-dependent Watts-Strogatz model
NASA Astrophysics Data System (ADS)
Casa Grande, H. L.; Cotacallapa, M.; Hase, M. O.
2017-01-01
In this work, we propose a scheme that provides an analytical estimate for the time-dependent degree distribution of some networks. This scheme maps the problem into a random walk in degree space, and then we choose the paths that are responsible for the dominant contributions. The method is illustrated on the dynamical versions of the Erdős-Rényi and Watts-Strogatz graphs, which were introduced as static models in the original formulation. We have succeeded in obtaining an analytical form for the dynamics Watts-Strogatz model, which is asymptotically exact for some regimes.
Random walk in degree space and the time-dependent Watts-Strogatz model.
Casa Grande, H L; Cotacallapa, M; Hase, M O
2017-01-01
In this work, we propose a scheme that provides an analytical estimate for the time-dependent degree distribution of some networks. This scheme maps the problem into a random walk in degree space, and then we choose the paths that are responsible for the dominant contributions. The method is illustrated on the dynamical versions of the Erdős-Rényi and Watts-Strogatz graphs, which were introduced as static models in the original formulation. We have succeeded in obtaining an analytical form for the dynamics Watts-Strogatz model, which is asymptotically exact for some regimes.
Applications of asynoptic space - Time Fourier transform methods to scanning satellite measurements
NASA Technical Reports Server (NTRS)
Lait, Leslie R.; Stanford, John L.
1988-01-01
A method proposed by Salby (1982) for computing the zonal space-time Fourier transform of asynoptically acquired satellite data is discussed. The method and its relationship to other techniques are briefly described, and possible problems in applying it to real data are outlined. Examples of results obtained using this technique are given which demonstrate its sensitivity to small-amplitude signals. A number of waves are found which have previously been observed as well as two not heretofore reported. A possible extension of the method which could increase temporal and longitudinal resolution is described.
Reference Frames in Relativistic Space-Time
NASA Astrophysics Data System (ADS)
Soffel, M.; Herold, H.; Ruder, H.; Schneider, M.
Three fundamental concepts of reference frames in relativistic space-time are confronted: 1. the gravitation compass, 2. the stellar compass and 3. the inertial compass. It is argued that under certain conditions asymptotically fixed (stellar) reference frames can be introduced with the same rigour as local Fermi frames, thereby eliminating one possible psychological reason why the importance of Fermi frames frequently has been overestimated in the past. As applications of these three concepts the authors discuss: 1. a relativistic definition of the geoid, 2. a relativistic astrometric problem and 3. the post-Newtonian theory of a laser gyroscope fixed to the Earth's surface.
Kaltenbacher, Barbara; Kaltenbacher, Manfred; Sim, Imbo
2013-01-01
We consider the second order wave equation in an unbounded domain and propose an advanced perfectly matched layer (PML) technique for its efficient and reliable simulation. In doing so, we concentrate on the time domain case and use the finite-element (FE) method for the space discretization. Our un-split-PML formulation requires four auxiliary variables within the PML region in three space dimensions. For a reduced version (rPML), we present a long time stability proof based on an energy analysis. The numerical case studies and an application example demonstrate the good performance and long time stability of our formulation for treating open domain problems. PMID:23888085
Classical space-times from the S-matrix
NASA Astrophysics Data System (ADS)
Neill, Duff; Rothstein, Ira Z.
2013-12-01
We show that classical space-times can be derived directly from the S-matrix for a theory of massive particles coupled to a massless spin two particle. As an explicit example we derive the Schwarzchild space-time as a series in GN. At no point of the derivation is any use made of the Einstein-Hilbert action or the Einstein equations. The intermediate steps involve only on-shell S-matrix elements which are generated via BCFW recursion relations and unitarity sewing techniques. The notion of a space-time metric is only introduced at the end of the calculation where it is extracted by matching the potential determined by the S-matrix to the geodesic motion of a test particle. Other static space-times such as Kerr follow in a similar manner. Furthermore, given that the procedure is action independent and depends only upon the choice of the representation of the little group, solutions to Yang-Mills (YM) theory can be generated in the same fashion. Moreover, the squaring relation between the YM and gravity three point functions shows that the seeds that generate solutions in the two theories are algebraically related. From a technical standpoint our methodology can also be utilized to calculate quantities relevant for the binary inspiral problem more efficiently then the more traditional Feynman diagram approach.
Li, Zhenyu; Wang, Bin; Liu, Hong
2016-08-30
Satellite capturing with free-floating space robots is still a challenging task due to the non-fixed base and unknown mass property issues. In this paper gyro and eye-in-hand camera data are adopted as an alternative choice for solving this problem. For this improved system, a new modeling approach that reduces the complexity of system control and identification is proposed. With the newly developed model, the space robot is equivalent to a ground-fixed manipulator system. Accordingly, a self-tuning control scheme is applied to handle such a control problem including unknown parameters. To determine the controller parameters, an estimator is designed based on the least-squares technique for identifying the unknown mass properties in real time. The proposed method is tested with a credible 3-dimensional ground verification experimental system, and the experimental results confirm the effectiveness of the proposed control scheme.
Li, Zhenyu; Wang, Bin; Liu, Hong
2016-01-01
Satellite capturing with free-floating space robots is still a challenging task due to the non-fixed base and unknown mass property issues. In this paper gyro and eye-in-hand camera data are adopted as an alternative choice for solving this problem. For this improved system, a new modeling approach that reduces the complexity of system control and identification is proposed. With the newly developed model, the space robot is equivalent to a ground-fixed manipulator system. Accordingly, a self-tuning control scheme is applied to handle such a control problem including unknown parameters. To determine the controller parameters, an estimator is designed based on the least-squares technique for identifying the unknown mass properties in real time. The proposed method is tested with a credible 3-dimensional ground verification experimental system, and the experimental results confirm the effectiveness of the proposed control scheme. PMID:27589748
NASA Technical Reports Server (NTRS)
Charles, John B.
2013-01-01
The technique of neutral buoyancy during water immersion was applied to a variety of questions pertaining to human performance factors in the early years of the space age. It was independently initiated by numerous aerospace contractors at nearly the same time, but specific applications depended on the problems that the developers were trying to solve. Those problems dealt primarily with human restraint and maneuverability and were often generic across extravehicular activity (EVA) and intravehicular activity (IVA) worksites. The same groups often also considered fractional gravity as well as weightless settings and experimented with ballasting to achieve lunar and Mars-equivalent loads as part of their on-going research and development. Dr. John Charles reviewed the association of those tasks with contemporary perceptions of the direction of NASA's future space exploration activities and with Air Force assessments of the military value of man in space.
RADECS Short Course Session I: The Space Radiation Environment
NASA Technical Reports Server (NTRS)
Xapsos, Michael; Bourdarie, Sebastien
2007-01-01
The presented slides and accompanying paper focus on radiation in the space environment. Since space exploration has begun it has become evident that the space environment is a highly aggressive medium. Beyond the natural protection provided by the Earth's atmosphere, various types of radiation can be encountered. Their characteristics (energy and nature), origins and distributions in space are extremely variable. This environment degrades electronic systems and on-board equipment in particular and creates radiobiological hazards during manned space flights. Based on several years of space exploration, a detailed analysis of the problems on satellites shows that the part due to the space environment is not negligible. It appears that the malfunctions are due to problems linked to the space environment, electronic problems, design problems, quality problems, other issues, and unexplained reasons. The space environment is largely responsible for about 20% of the anomalies occurring on satellites and a better knowledge of that environment could only increase the average lifetime of space vehicles. This naturally leads to a detailed study of the space environment and of the effects that it induces on space vehicles and astronauts. Sources of radiation in the space environment are discussed here and include the solar activity cycle, galactic cosmic rays, solar particle events, and Earth radiation belts. Future challenges for space radiation environment models are briefly addressed.
NASA Technical Reports Server (NTRS)
Chang, Sin-Chung; Wang, Xiao-Yen; Chow, Chuen-Yen
1994-01-01
A new numerical discretization method for solving conservation laws is being developed. This new approach differs substantially in both concept and methodology from the well-established methods, i.e., finite difference, finite volume, finite element, and spectral methods. It is motivated by several important physical/numerical considerations and designed to avoid several key limitations of the above traditional methods. As a result of the above considerations, a set of key principles for the design of numerical schemes was put forth in a previous report. These principles were used to construct several numerical schemes that model a 1-D time-dependent convection-diffusion equation. These schemes were then extended to solve the time-dependent Euler and Navier-Stokes equations of a perfect gas. It was shown that the above schemes compared favorably with the traditional schemes in simplicity, generality, and accuracy. In this report, the 2-D versions of the above schemes, except the Navier-Stokes solver, are constructed using the same set of design principles. Their constructions are simplified greatly by the use of a nontraditional space-time mesh. Its use results in the simplest stencil possible, i.e., a tetrahedron in a 3-D space-time with a vertex at the upper time level and other three at the lower time level. Because of the similarity in their design, each of the present 2-D solvers virtually shares with its 1-D counterpart the same fundamental characteristics. Moreover, it is shown that the present Euler solver is capable of generating highly accurate solutions for a famous 2-D shock reflection problem. Specifically, both the incident and the reflected shocks can be resolved by a single data point without the presence of numerical oscillations near the discontinuity.
Insight into the ten-penny problem: guiding search by constraints and maximization.
Öllinger, Michael; Fedor, Anna; Brodt, Svenja; Szathmáry, Eörs
2017-09-01
For a long time, insight problem solving has been either understood as nothing special or as a particular class of problem solving. The first view implicates the necessity to find efficient heuristics that restrict the search space, the second, the necessity to overcome self-imposed constraints. Recently, promising hybrid cognitive models attempt to merge both approaches. In this vein, we were interested in the interplay of constraints and heuristic search, when problem solvers were asked to solve a difficult multi-step problem, the ten-penny problem. In three experimental groups and one control group (N = 4 × 30) we aimed at revealing, what constraints drive problem difficulty in this problem, and how relaxing constraints, and providing an efficient search criterion facilitates the solution. We also investigated how the search behavior of successful problem solvers and non-solvers differ. We found that relaxing constraints was necessary but not sufficient to solve the problem. Without efficient heuristics that facilitate the restriction of the search space, and testing the progress of the problem solving process, the relaxation of constraints was not effective. Relaxing constraints and applying the search criterion are both necessary to effectively increase solution rates. We also found that successful solvers showed promising moves earlier and had a higher maximization and variation rate across solution attempts. We propose that this finding sheds light on how different strategies contribute to solving difficult problems. Finally, we speculate about the implications of our findings for insight problem solving.
USA: Economics, Politics, Ideology, No. 10, October 1977
1977-11-22
Stereotype Space exploration cannot begin until many complex scientific and technical problems have been solved. The very fact that a program of space...a time of cold war. It was precisely under these conditions that the rigid stereotype of American reaction to the crises it had encountered in the...the possibility of the improvement of their relations and were in a rush to advertise the end of the cold war, and they believed, with unwarranted
Blind source separation problem in GPS time series
NASA Astrophysics Data System (ADS)
Gualandi, A.; Serpelloni, E.; Belardinelli, M. E.
2016-04-01
A critical point in the analysis of ground displacement time series, as those recorded by space geodetic techniques, is the development of data-driven methods that allow the different sources of deformation to be discerned and characterized in the space and time domains. Multivariate statistic includes several approaches that can be considered as a part of data-driven methods. A widely used technique is the principal component analysis (PCA), which allows us to reduce the dimensionality of the data space while maintaining most of the variance of the dataset explained. However, PCA does not perform well in finding the solution to the so-called blind source separation (BSS) problem, i.e., in recovering and separating the original sources that generate the observed data. This is mainly due to the fact that PCA minimizes the misfit calculated using an L2 norm (χ 2), looking for a new Euclidean space where the projected data are uncorrelated. The independent component analysis (ICA) is a popular technique adopted to approach the BSS problem. However, the independence condition is not easy to impose, and it is often necessary to introduce some approximations. To work around this problem, we test the use of a modified variational Bayesian ICA (vbICA) method to recover the multiple sources of ground deformation even in the presence of missing data. The vbICA method models the probability density function (pdf) of each source signal using a mix of Gaussian distributions, allowing for more flexibility in the description of the pdf of the sources with respect to standard ICA, and giving a more reliable estimate of them. Here we present its application to synthetic global positioning system (GPS) position time series, generated by simulating deformation near an active fault, including inter-seismic, co-seismic, and post-seismic signals, plus seasonal signals and noise, and an additional time-dependent volcanic source. We evaluate the ability of the PCA and ICA decomposition techniques in explaining the data and in recovering the original (known) sources. Using the same number of components, we find that the vbICA method fits the data almost as well as a PCA method, since the χ 2 increase is less than 10 % the value calculated using a PCA decomposition. Unlike PCA, the vbICA algorithm is found to correctly separate the sources if the correlation of the dataset is low (<0.67) and the geodetic network is sufficiently dense (ten continuous GPS stations within a box of side equal to two times the locking depth of a fault where an earthquake of Mw >6 occurred). We also provide a cookbook for the use of the vbICA algorithm in analyses of position time series for tectonic and non-tectonic applications.
Research on Scheduling Algorithm for Multi-satellite and Point Target Task on Swinging Mode
NASA Astrophysics Data System (ADS)
Wang, M.; Dai, G.; Peng, L.; Song, Z.; Chen, G.
2012-12-01
Nowadays, using satellite in space to observe ground is an important and major method to obtain ground information. With the development of the scientific technology in the field of space, many fields such as military and economic and other areas have more and more requirement of space technology because of the benefits of the satellite's widespread, timeliness and unlimited of area and country. And at the same time, because of the wide use of all kinds of satellites, sensors, repeater satellites and ground receiving stations, ground control system are now facing great challenge. Therefore, how to make the best value of satellite resources so as to make full use of them becomes an important problem of ground control system. Satellite scheduling is to distribute the resource to all tasks without conflict to obtain the scheduling result so as to complete as many tasks as possible to meet user's requirement under considering the condition of the requirement of satellites, sensors and ground receiving stations. Considering the size of the task, we can divide tasks into point task and area task. This paper only considers point targets. In this paper, a description of satellite scheduling problem and a chief introduction of the theory of satellite scheduling are firstly made. We also analyze the restriction of resource and task in scheduling satellites. The input and output flow of scheduling process are also chiefly described in the paper. On the basis of these analyses, we put forward a scheduling model named as multi-variable optimization model for multi-satellite and point target task on swinging mode. In the multi-variable optimization model, the scheduling problem is transformed the parametric optimization problem. The parameter we wish to optimize is the swinging angle of every time-window. In the view of the efficiency and accuracy, some important problems relating the satellite scheduling such as the angle relation between satellites and ground targets, positive and negative swinging angle and the computation of time window are analyzed and discussed. And many strategies to improve the efficiency of this model are also put forward. In order to solve the model, we bring forward the conception of activity sequence map. By using the activity sequence map, the activity choice and the start time of the activity can be divided. We also bring forward three neighborhood operators to search the result space. The front movement remaining time and the back movement remaining time are used to analyze the feasibility to generate solution from neighborhood operators. Lastly, the algorithm to solve the problem and model is put forward based genetic algorithm. Population initialization, crossover operator, mutation operator, individual evaluation, collision decrease operator, select operator and collision elimination operator is designed in the paper. Finally, the scheduling result and the simulation for a practical example on 5 satellites and 100 point targets with swinging mode is given, and the scheduling performances are also analyzed while the swinging angle in 0, 5, 10, 15, 25. It can be shown by the result that the model and the algorithm are more effective than those ones without swinging mode.
An Autonomous Sensor Tasking Approach for Large Scale Space Object Cataloging
NASA Astrophysics Data System (ADS)
Linares, R.; Furfaro, R.
The field of Space Situational Awareness (SSA) has progressed over the last few decades with new sensors coming online, the development of new approaches for making observations, and new algorithms for processing them. Although there has been success in the development of new approaches, a missing piece is the translation of SSA goals to sensors and resource allocation; otherwise known as the Sensor Management Problem (SMP). This work solves the SMP using an artificial intelligence approach called Deep Reinforcement Learning (DRL). Stable methods for training DRL approaches based on neural networks exist, but most of these approaches are not suitable for high dimensional systems. The Asynchronous Advantage Actor-Critic (A3C) method is a recently developed and effective approach for high dimensional systems, and this work leverages these results and applies this approach to decision making in SSA. The decision space for the SSA problems can be high dimensional, even for tasking of a single telescope. Since the number of SOs in space is relatively high, each sensor will have a large number of possible actions at a given time. Therefore, efficient DRL approaches are required when solving the SMP for SSA. This work develops a A3C based method for DRL applied to SSA sensor tasking. One of the key benefits of DRL approaches is the ability to handle high dimensional data. For example DRL methods have been applied to image processing for the autonomous car application. For example, a 256x256 RGB image has 196608 parameters (256*256*3=196608) which is very high dimensional, and deep learning approaches routinely take images like this as inputs. Therefore, when applied to the whole catalog the DRL approach offers the ability to solve this high dimensional problem. This work has the potential to, for the first time, solve the non-myopic sensor tasking problem for the whole SO catalog (over 22,000 objects) providing a truly revolutionary result.
Scattering of focused ultrasonic beams by cavities in a solid half-space.
Rahni, Ehsan Kabiri; Hajzargarbashi, Talieh; Kundu, Tribikram
2012-08-01
The ultrasonic field generated by a point focused acoustic lens placed in a fluid medium adjacent to a solid half-space, containing one or more spherical cavities, is modeled. The semi-analytical distributed point source method (DPSM) is followed for the modeling. This technique properly takes into account the interaction effect between the cavities placed in the focused ultrasonic field, fluid-solid interface and the lens surface. The approximate analytical solution that is available in the literature for the single cavity geometry is very restrictive and cannot handle multiple cavity problems. Finite element solutions for such problems are also prohibitively time consuming at high frequencies. Solution of this problem is necessary to predict when two cavities placed in close proximity inside a solid can be distinguished by an acoustic lens placed outside the solid medium and when such distinction is not possible.
A minimum propellant solution to an orbit-to-orbit transfer using a low thrust propulsion system
NASA Technical Reports Server (NTRS)
Cobb, Shannon S.
1991-01-01
The Space Exploration Initiative is considering the use of low thrust (nuclear electric, solar electric) and intermediate thrust (nuclear thermal) propulsion systems for transfer to Mars and back. Due to the duration of such a mission, a low thrust minimum-fuel solution is of interest; a savings of fuel can be substantial if the propulsion system is allowed to be turned off and back on. This switching of the propulsion system helps distinguish the minimal-fuel problem from the well-known minimum-time problem. Optimal orbit transfers are also of interest to the development of a guidance system for orbital maneuvering vehicles which will be needed, for example, to deliver cargoes to the Space Station Freedom. The problem of optimizing trajectories for an orbit-to-orbit transfer with minimum-fuel expenditure using a low thrust propulsion system is addressed.
NASA Technical Reports Server (NTRS)
Madden, Michael G.; Wyrick, Roberta; O'Neill, Dale E.
2005-01-01
Space Shuttle Processing is a complicated and highly variable project. The planning and scheduling problem, categorized as a Resource Constrained - Stochastic Project Scheduling Problem (RC-SPSP), has a great deal of variability in the Orbiter Processing Facility (OPF) process flow from one flight to the next. Simulation Modeling is a useful tool in estimation of the makespan of the overall process. However, simulation requires a model to be developed, which itself is a labor and time consuming effort. With such a dynamic process, often the model would potentially be out of synchronization with the actual process, limiting the applicability of the simulation answers in solving the actual estimation problem. Integration of TEAMS model enabling software with our existing schedule program software is the basis of our solution. This paper explains the approach used to develop an auto-generated simulation model from planning and schedule efforts and available data.
NASA's Applied Sciences for Water Resources
NASA Technical Reports Server (NTRS)
Doorn, Bradley; Toll, David; Engman, Ted
2011-01-01
The Earth Systems Division within NASA has the primary responsibility for the Earth Science Applied Science Program and the objective to accelerate the use of NASA science results in applications to help solve problems important to society and the economy. The primary goal of the Earth Science Applied Science Program is to improve future and current operational systems by infusing them with scientific knowledge of the Earth system gained through space-based observation, assimilation of new observations, and development and deployment of enabling technologies, systems, and capabilities. This paper discusses one of the major problems facing water resources managers, that of having timely and accurate data to drive their decision support tools. It then describes how NASA?s science and space based satellites may be used to overcome this problem. Opportunities for the water resources community to participate in NASA?s Water Resources Applications Program are described.
Genomic big data hitting the storage bottleneck.
Papageorgiou, Louis; Eleni, Picasi; Raftopoulou, Sofia; Mantaiou, Meropi; Megalooikonomou, Vasileios; Vlachakis, Dimitrios
2018-01-01
During the last decades, there is a vast data explosion in bioinformatics. Big data centres are trying to face this data crisis, reaching high storage capacity levels. Although several scientific giants examine how to handle the enormous pile of information in their cupboards, the problem remains unsolved. On a daily basis, there is a massive quantity of permanent loss of extensive information due to infrastructure and storage space problems. The motivation for sequencing has fallen behind. Sometimes, the time that is spent to solve storage space problems is longer than the one dedicated to collect and analyse data. To bring sequencing to the foreground, scientists have to slide over such obstacles and find alternative ways to approach the issue of data volume. Scientific community experiences the data crisis era, where, out of the box solutions may ease the typical research workflow, until technological development meets the needs of Bioinformatics.
A heuristic constraint programmed planner for deep space exploration problems
NASA Astrophysics Data System (ADS)
Jiang, Xiao; Xu, Rui; Cui, Pingyuan
2017-10-01
In recent years, the increasing numbers of scientific payloads and growing constraints on the probe have made constraint processing technology a hotspot in the deep space planning field. In the procedure of planning, the ordering of variables and values plays a vital role. This paper we present two heuristic ordering methods for variables and values. On this basis a graphplan-like constraint-programmed planner is proposed. In the planner we convert the traditional constraint satisfaction problem to a time-tagged form with different levels. Inspired by the most constrained first principle in constraint satisfaction problem (CSP), the variable heuristic is designed by the number of unassigned variables in the constraint and the value heuristic is designed by the completion degree of the support set. The simulation experiments show that the planner proposed is effective and its performance is competitive with other kind of planners.
Efficient parallel algorithms for string editing and related problems
NASA Technical Reports Server (NTRS)
Apostolico, Alberto; Atallah, Mikhail J.; Larmore, Lawrence; Mcfaddin, H. S.
1988-01-01
The string editing problem for input strings x and y consists of transforming x into y by performing a series of weighted edit operations on x of overall minimum cost. An edit operation on x can be the deletion of a symbol from x, the insertion of a symbol in x or the substitution of a symbol x with another symbol. This problem has a well known O((absolute value of x)(absolute value of y)) time sequential solution (25). The efficient Program Requirements Analysis Methods (PRAM) parallel algorithms for the string editing problem are given. If m = ((absolute value of x),(absolute value of y)) and n = max((absolute value of x),(absolute value of y)), then the CREW bound is O (log m log n) time with O (mn/log m) processors. In all algorithms, space is O (mn).
A Maximal Element Theorem in FWC-Spaces and Its Applications
Hu, Qingwen; Miao, Yulin
2014-01-01
A maximal element theorem is proved in finite weakly convex spaces (FWC-spaces, in short) which have no linear, convex, and topological structure. Using the maximal element theorem, we develop new existence theorems of solutions to variational relation problem, generalized equilibrium problem, equilibrium problem with lower and upper bounds, and minimax problem in FWC-spaces. The results represented in this paper unify and extend some known results in the literature. PMID:24782672
NASA Technical Reports Server (NTRS)
Feiveson, Alan H.; Fiedler, James; Lee, Stuart M. M.; Westby, Christian M.; Stenger, Michael B.; Platts, Steven H.
2014-01-01
Orthostatic Intolerance (OI) is the propensity to develop symptoms of fainting during upright standing. OI is associated with changes in heart rate, blood pressure and other measures of cardiac function. Problem: NASA astronauts have shown increased susceptibility to OI on return from space missions. Current methods for counteracting OI in astronauts include fluid loading and the use of compression garments. Multivariate trajectory spread is greater as OI increases. Pairwise comparisons at the same time within subjects allows incorporation of pass/fail outcomes. Path length, convex hull area, and covariance matrix determinant do well as statistics to summarize this spread Missing data problems Time series analysis need many more time points per OTT session treatment of trend? how incorporate survival information?
Optimal bounds and extremal trajectories for time averages in nonlinear dynamical systems
NASA Astrophysics Data System (ADS)
Tobasco, Ian; Goluskin, David; Doering, Charles R.
2018-02-01
For any quantity of interest in a system governed by ordinary differential equations, it is natural to seek the largest (or smallest) long-time average among solution trajectories, as well as the extremal trajectories themselves. Upper bounds on time averages can be proved a priori using auxiliary functions, the optimal choice of which is a convex optimization problem. We prove that the problems of finding maximal trajectories and minimal auxiliary functions are strongly dual. Thus, auxiliary functions provide arbitrarily sharp upper bounds on time averages. Moreover, any nearly minimal auxiliary function provides phase space volumes in which all nearly maximal trajectories are guaranteed to lie. For polynomial equations, auxiliary functions can be constructed by semidefinite programming, which we illustrate using the Lorenz system.
NASA Astrophysics Data System (ADS)
Lü, Boqiang; Shi, Xiaoding; Zhong, Xin
2018-06-01
We are concerned with the Cauchy problem of the two-dimensional (2D) nonhomogeneous incompressible Navier–Stokes equations with vacuum as far-field density. It is proved that if the initial density decays not too slow at infinity, the 2D Cauchy problem of the density-dependent Navier–Stokes equations on the whole space admits a unique global strong solution. Note that the initial data can be arbitrarily large and the initial density can contain vacuum states and even have compact support. Furthermore, we also obtain the large time decay rates of the spatial gradients of the velocity and the pressure, which are the same as those of the homogeneous case.
[Taylor and Hill, Incorporated's JSC Cryo Chamber A
NASA Technical Reports Server (NTRS)
Morales, Rito
2008-01-01
NASA commissioned construction of an environmental simulation test chamber which was completed in 1964 at Johnson Space Center (JSC) in Houston, Texas. The facility, Chamber A, was invaluable for testing spacecraft and satellites before deployment to space. By testing spacecraft in an environment similar to the one they would be functioning in, potential problems could be addressed before launch. A new addition to NASA's observatory inventory is called the James Webb Space Telescope (JWST), after a former Administrator of NASA. The new telescope will have 7 times the mirror area of the Hubble, with a target destination approximately one million miles from earth. Scheduled for launch in 2013, the JWST will allow scientists the ability to see, for the first time, the first galaxies that formed in the early Universe. Pre-launch testing of JWST must be performed in environments that approximate its final target space environment as closely as possible.
Understanding climate: A strategy for climate modeling and predictability research, 1985-1995
NASA Technical Reports Server (NTRS)
Thiele, O. (Editor); Schiffer, R. A. (Editor)
1985-01-01
The emphasis of the NASA strategy for climate modeling and predictability research is on the utilization of space technology to understand the processes which control the Earth's climate system and it's sensitivity to natural and man-induced changes and to assess the possibilities for climate prediction on time scales of from about two weeks to several decades. Because the climate is a complex multi-phenomena system, which interacts on a wide range of space and time scales, the diversity of scientific problems addressed requires a hierarchy of models along with the application of modern empirical and statistical techniques which exploit the extensive current and potential future global data sets afforded by space observations. Observing system simulation experiments, exploiting these models and data, will also provide the foundation for the future climate space observing system, e.g., Earth observing system (EOS), 1985; Tropical Rainfall Measuring Mission (TRMM) North, et al. NASA, 1984.
Annihilation cross section of Kaluza Klien dark matter
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sharma, Rakesh, E-mail: rakesh-sharma-ujn@yahoo.co.in; Upadhyaya, G. K., E-mail: gopalujjain@yahoo.co.in; Sharma, S.
2015-07-31
The question as to how this universe came into being and as to how it has evolved to its present stage, is an old question. The answer to this question unfolds many secrets regarding fundamental particles and forces between them. Theodor Kaluza proposed the concept that the universe is composed of more than four space-time dimensions. In his work, electromagnetism is united with gravity. Various extra dimension formulations have been proposed to solve a variety of problems. Recently, the idea of more than four space time dimensions is applied to the search for particle identity of dark matter (DM). Signaturemore » of dark matter can be revealed by analysis of very high energy electrons which are coming from outer space. We investigate recent advancement in the field of dark matter search with reference to very high energy electrons from outer space [1-8].« less
Problem Space Matters: Evaluation of a German Enrichment Program for Gifted Children.
Welter, Marisete M; Jaarsveld, Saskia; Lachmann, Thomas
2018-01-01
We studied the development of cognitive abilities related to intelligence and creativity ( N = 48, 6-10 years old), using a longitudinal design (over one school year), in order to evaluate an Enrichment Program for gifted primary school children initiated by the government of the German federal state of Rhineland-Palatinate ( Entdeckertag Rheinland Pfalz , Germany; ET; Day of Discoverers). A group of German primary school children ( N = 24), identified earlier as intellectually gifted and selected to join the ET program was compared to a gender-, class- and IQ- matched group of control children that did not participate in this program. All participants performed the Standard Progressive Matrices (SPM) test, which measures intelligence in well-defined problem space; the Creative Reasoning Task (CRT), which measures intelligence in ill-defined problem space; and the test of creative thinking-drawing production (TCT-DP), which measures creativity, also in ill-defined problem space. Results revealed that problem space matters: the ET program is effective only for the improvement of intelligence operating in well-defined problem space. An effect was found for intelligence as measured by SPM only, but neither for intelligence operating in ill-defined problem space (CRT) nor for creativity (TCT-DP). This suggests that, depending on the type of problem spaces presented, different cognitive abilities are elicited in the same child. Therefore, enrichment programs for gifted, but also for children attending traditional schools, should provide opportunities to develop cognitive abilities related to intelligence, operating in both well- and ill-defined problem spaces, and to creativity in a parallel, using an interactive approach.
Problem Space Matters: Evaluation of a German Enrichment Program for Gifted Children
Welter, Marisete M.; Jaarsveld, Saskia; Lachmann, Thomas
2018-01-01
We studied the development of cognitive abilities related to intelligence and creativity (N = 48, 6–10 years old), using a longitudinal design (over one school year), in order to evaluate an Enrichment Program for gifted primary school children initiated by the government of the German federal state of Rhineland-Palatinate (Entdeckertag Rheinland Pfalz, Germany; ET; Day of Discoverers). A group of German primary school children (N = 24), identified earlier as intellectually gifted and selected to join the ET program was compared to a gender-, class- and IQ- matched group of control children that did not participate in this program. All participants performed the Standard Progressive Matrices (SPM) test, which measures intelligence in well-defined problem space; the Creative Reasoning Task (CRT), which measures intelligence in ill-defined problem space; and the test of creative thinking-drawing production (TCT-DP), which measures creativity, also in ill-defined problem space. Results revealed that problem space matters: the ET program is effective only for the improvement of intelligence operating in well-defined problem space. An effect was found for intelligence as measured by SPM only, but neither for intelligence operating in ill-defined problem space (CRT) nor for creativity (TCT-DP). This suggests that, depending on the type of problem spaces presented, different cognitive abilities are elicited in the same child. Therefore, enrichment programs for gifted, but also for children attending traditional schools, should provide opportunities to develop cognitive abilities related to intelligence, operating in both well- and ill-defined problem spaces, and to creativity in a parallel, using an interactive approach. PMID:29740367
Immune changes in humans concomitant with space flights of up to 10 days duration
NASA Technical Reports Server (NTRS)
Taylor, Gerald R.
1993-01-01
The time relation of various classes of in-flight human physiological changes is illustrated. Certain problems, such as neurovestibular, fluid, and electrolyte imbalances tend to occur early in a flight, followed by stabilization at some microgravity equilibrium level. Cardiovascular dysfunctions and erythrocyte mass losses appear to follow a similar pattern, although the significant changes occur later in flight. Bone and calcium changes and radiation effects are thought to progressively worsen with time, whereas the time course of immune change is yet to be fully understood. Significant immunologic changes in cosmonauts and astronauts during and after space flight have been documented as have microbiological changes. Thus, space flight can be expected to effect a blunting of the human cellular immune mechanism concomitant with a relative increase in potentially pathogenic microorganisms. This combination would seem to increase the probability of infectious disease events in flight.
Space-time interface-tracking with topology change (ST-TC)
NASA Astrophysics Data System (ADS)
Takizawa, Kenji; Tezduyar, Tayfun E.; Buscher, Austin; Asada, Shohei
2014-10-01
To address the computational challenges associated with contact between moving interfaces, such as those in cardiovascular fluid-structure interaction (FSI), parachute FSI, and flapping-wing aerodynamics, we introduce a space-time (ST) interface-tracking method that can deal with topology change (TC). In cardiovascular FSI, our primary target is heart valves. The method is a new version of the deforming-spatial-domain/stabilized space-time (DSD/SST) method, and we call it ST-TC. It includes a master-slave system that maintains the connectivity of the "parent" mesh when there is contact between the moving interfaces. It is an efficient, practical alternative to using unstructured ST meshes, but without giving up on the accurate representation of the interface or consistent representation of the interface motion. We explain the method with conceptual examples and present 2D test computations with models representative of the classes of problems we are targeting.
The problems inherent in teaching technical writing and report writing to native Americans
NASA Technical Reports Server (NTRS)
Zukowski/faust, J.
1981-01-01
Teaching technical writing to Native Americans contending with a second language and culture is addressed. Learning difficulties arising from differences between native and acquired language and cultural systems are examined. Compartmentalized teaching, which presents the ideals of technical writing in minimal units, and skills development are considered. Rhetorical problems treated include logic of arrangement, selection of support and scope of detail, and time and space. Specific problems selected include the concept of promptness, the contextualization of purpose, interpersonal relationships, wordiness, mixture of registers, and the problem of abstracting. Four inductive procedures for students having writing and perception problems are included. Four sample exercises and a bibliography of 13 references are also included.
Linear decomposition approach for a class of nonconvex programming problems.
Shen, Peiping; Wang, Chunfeng
2017-01-01
This paper presents a linear decomposition approach for a class of nonconvex programming problems by dividing the input space into polynomially many grids. It shows that under certain assumptions the original problem can be transformed and decomposed into a polynomial number of equivalent linear programming subproblems. Based on solving a series of liner programming subproblems corresponding to those grid points we can obtain the near-optimal solution of the original problem. Compared to existing results in the literature, the proposed algorithm does not require the assumptions of quasi-concavity and differentiability of the objective function, and it differs significantly giving an interesting approach to solving the problem with a reduced running time.
Time Crystal Platform: From Quasicrystal Structures in Time to Systems with Exotic Interactions.
Giergiel, Krzysztof; Miroszewski, Artur; Sacha, Krzysztof
2018-04-06
Time crystals are quantum many-body systems that, due to interactions between particles, are able to spontaneously self-organize their motion in a periodic way in time by analogy with the formation of crystalline structures in space in condensed matter physics. In solid state physics properties of space crystals are often investigated with the help of external potentials that are spatially periodic and reflect various crystalline structures. A similar approach can be applied for time crystals, as periodically driven systems constitute counterparts of spatially periodic systems, but in the time domain. Here we show that condensed matter problems ranging from single particles in potentials of quasicrystal structure to many-body systems with exotic long-range interactions can be realized in the time domain with an appropriate periodic driving. Moreover, it is possible to create molecules where atoms are bound together due to destructive interference if the atomic scattering length is modulated in time.
Time Crystal Platform: From Quasicrystal Structures in Time to Systems with Exotic Interactions
NASA Astrophysics Data System (ADS)
Giergiel, Krzysztof; Miroszewski, Artur; Sacha, Krzysztof
2018-04-01
Time crystals are quantum many-body systems that, due to interactions between particles, are able to spontaneously self-organize their motion in a periodic way in time by analogy with the formation of crystalline structures in space in condensed matter physics. In solid state physics properties of space crystals are often investigated with the help of external potentials that are spatially periodic and reflect various crystalline structures. A similar approach can be applied for time crystals, as periodically driven systems constitute counterparts of spatially periodic systems, but in the time domain. Here we show that condensed matter problems ranging from single particles in potentials of quasicrystal structure to many-body systems with exotic long-range interactions can be realized in the time domain with an appropriate periodic driving. Moreover, it is possible to create molecules where atoms are bound together due to destructive interference if the atomic scattering length is modulated in time.
Deep space communication - A one billion mile noisy channel
NASA Technical Reports Server (NTRS)
Smith, J. G.
1982-01-01
Deep space exploration is concerned with the study of natural phenomena in the solar system with the aid of measurements made at spacecraft on deep space missions. Deep space communication refers to communication between earth and spacecraft in deep space. The Deep Space Network is an earth-based facility employed for deep space communication. It includes a network of large tracking antennas located at various positions around the earth. The goals and achievements of deep space exploration over the past 20 years are discussed along with the broad functional requirements of deep space missions. Attention is given to the differences in space loss between communication satellites and deep space vehicles, effects of the long round-trip light time on spacecraft autonomy, requirements for the use of massive nuclear power plants on spacecraft at large distances from the sun, and the kinds of scientific return provided by a deep space mission. Problems concerning a deep space link of one billion miles are also explored.
Collision warning and avoidance considerations for the Space Shuttle and Space Station Freedom
NASA Technical Reports Server (NTRS)
Vilas, Faith; Collins, Michael F.; Kramer, Paul C.; Arndt, G. Dickey; Suddath, Jerry H.
1990-01-01
The increasing hazard of manmade debris in low earth orbit (LEO) has focused attention on the requirement for collision detection, warning and avoidance systems to be developed in order to protect manned (and unmanned) spacecraft. With the number of debris objects expected to be increasing with time, the impact hazard will also be increasing. The safety of the Space Shuttle and the Space Station Freedom from destructive or catastrophic collision resulting from the hypervelocity impact of a LEO object is of increasing concern to NASA. A number of approaches to this problem are in effect or under development. The collision avoidance procedures now in effect for the Shuttle are described, and detection and avoidance procedures presently being developed at the Johnson Space Center for the Space Station Freedom are discussed.
Using Virtual Simulations in the Design of 21st Century Space Science Environments
NASA Technical Reports Server (NTRS)
Hutchinson, Sonya L.; Alves, Jeffery R.
1996-01-01
Space Technology has been rapidly increasing in the past decade. This can be attributed to the future construction of the International Space Station (ISS). New innovations must constantly be engineered to make ISS the safest, quality, research facility in space. Since space science must often be gathered by crew members, more attention must be geared to the human's safety and comfort. Virtual simulations are now being used to design environments that crew members can live in for long periods of time without harmful effects to their bodies. This paper gives a few examples of the ergonomic design problems that arise on manned space flights, and design solutions that follow NASA's strategic commitment to customer satisfaction. The conclusions show that virtual simulations are a great asset to 21st century design.
NASA Astrophysics Data System (ADS)
Mezgebo, Biniyam; Nagib, Karim; Fernando, Namal; Kordi, Behzad; Sherif, Sherif
2018-02-01
Swept Source optical coherence tomography (SS-OCT) is an important imaging modality for both medical and industrial diagnostic applications. A cross-sectional SS-OCT image is obtained by applying an inverse discrete Fourier transform (DFT) to axial interferograms measured in the frequency domain (k-space). This inverse DFT is typically implemented as a fast Fourier transform (FFT) that requires the data samples to be equidistant in k-space. As the frequency of light produced by a typical wavelength-swept laser is nonlinear in time, the recorded interferogram samples will not be uniformly spaced in k-space. Many image reconstruction methods have been proposed to overcome this problem. Most such methods rely on oversampling the measured interferogram then use either hardware, e.g., Mach-Zhender interferometer as a frequency clock module, or software, e.g., interpolation in k-space, to obtain equally spaced samples that are suitable for the FFT. To overcome the problem of nonuniform sampling in k-space without any need for interferogram oversampling, an earlier method demonstrated the use of the nonuniform discrete Fourier transform (NDFT) for image reconstruction in SS-OCT. In this paper, we present a more accurate method for SS-OCT image reconstruction from nonuniform samples in k-space using a scaled nonuniform Fourier transform. The result is demonstrated using SS-OCT images of Axolotl salamander eggs.
Basic research for the geodynamics program
NASA Technical Reports Server (NTRS)
1991-01-01
The mathematical models of space very long base interferometry (VLBI) observables suitable for least squares covariance analysis were derived and estimatability problems inherent in the space VLBI system were explored, including a detailed rank defect analysis and sensitivity analysis. An important aim is to carry out a comparative analysis of the mathematical models of the ground-based VLBI and space VLBI observables in order to describe the background in detail. Computer programs were developed in order to check the relations, assess errors, and analyze sensitivity. In order to investigate the estimatability of different geodetic and geodynamic parameters from the space VLBI observables, the mathematical models for time delay and time delay rate observables of space VLBI were analytically derived along with the partial derivatives with respect to the parameters. Rank defect analysis was carried out both by analytical and numerical testing of linear dependencies between the columns of the normal matrix thus formed. Definite conclusions were formed about the rank defects in the system.
Dissipative structure and global existence in critical space for Timoshenko system of memory type
NASA Astrophysics Data System (ADS)
Mori, Naofumi
2018-08-01
In this paper, we consider the initial value problem for the Timoshenko system with a memory term in one dimensional whole space. In the first place, we consider the linearized system: applying the energy method in the Fourier space, we derive the pointwise estimate of the solution in the Fourier space, which first gives the optimal decay estimate of the solution. Next, we give a characterization of the dissipative structure of the system by using the spectral analysis, which confirms our pointwise estimate is optimal. In the second place, we consider the nonlinear system: we show that the global-in-time existence and uniqueness result could be proved in the minimal regularity assumption in the critical Sobolev space H2. In the proof we don't need any time-weighted norm as recent works; we use just an energy method, which is improved to overcome the difficulties caused by regularity-loss property of Timoshenko system.
Crevillén-García, D
2018-04-01
Time-consuming numerical simulators for solving groundwater flow and dissolution models of physico-chemical processes in deep aquifers normally require some of the model inputs to be defined in high-dimensional spaces in order to return realistic results. Sometimes, the outputs of interest are spatial fields leading to high-dimensional output spaces. Although Gaussian process emulation has been satisfactorily used for computing faithful and inexpensive approximations of complex simulators, these have been mostly applied to problems defined in low-dimensional input spaces. In this paper, we propose a method for simultaneously reducing the dimensionality of very high-dimensional input and output spaces in Gaussian process emulators for stochastic partial differential equation models while retaining the qualitative features of the original models. This allows us to build a surrogate model for the prediction of spatial fields in such time-consuming simulators. We apply the methodology to a model of convection and dissolution processes occurring during carbon capture and storage.
NASA Astrophysics Data System (ADS)
Park, Suhyung; Park, Jaeseok
2015-05-01
Accelerated dynamic MRI, which exploits spatiotemporal redundancies in k - t space and coil dimension, has been widely used to reduce the number of signal encoding and thus increase imaging efficiency with minimal loss of image quality. Nonetheless, particularly in cardiac MRI it still suffers from artifacts and amplified noise in the presence of time-drifting coil sensitivity due to relative motion between coil and subject (e.g. free breathing). Furthermore, a substantial number of additional calibrating signals is to be acquired to warrant accurate calibration of coil sensitivity. In this work, we propose a novel, accelerated dynamic cardiac MRI with sparse-Kalman-smoother self-calibration and reconstruction (k - t SPARKS), which is robust to time-varying coil sensitivity even with a small number of calibrating signals. The proposed k - t SPARKS incorporates Kalman-smoother self-calibration in k - t space and sparse signal recovery in x - f space into a single optimization problem, leading to iterative, joint estimation of time-varying convolution kernels and missing signals in k - t space. In the Kalman-smoother calibration, motion-induced uncertainties over the entire time frames were included in modeling state transition while a coil-dependent noise statistic in describing measurement process. The sparse signal recovery iteratively alternates with the self-calibration to tackle the ill-conditioning problem potentially resulting from insufficient calibrating signals. Simulations and experiments were performed using both the proposed and conventional methods for comparison, revealing that the proposed k - t SPARKS yields higher signal-to-error ratio and superior temporal fidelity in both breath-hold and free-breathing cardiac applications over all reduction factors.
Park, Suhyung; Park, Jaeseok
2015-05-07
Accelerated dynamic MRI, which exploits spatiotemporal redundancies in k - t space and coil dimension, has been widely used to reduce the number of signal encoding and thus increase imaging efficiency with minimal loss of image quality. Nonetheless, particularly in cardiac MRI it still suffers from artifacts and amplified noise in the presence of time-drifting coil sensitivity due to relative motion between coil and subject (e.g. free breathing). Furthermore, a substantial number of additional calibrating signals is to be acquired to warrant accurate calibration of coil sensitivity. In this work, we propose a novel, accelerated dynamic cardiac MRI with sparse-Kalman-smoother self-calibration and reconstruction (k - t SPARKS), which is robust to time-varying coil sensitivity even with a small number of calibrating signals. The proposed k - t SPARKS incorporates Kalman-smoother self-calibration in k - t space and sparse signal recovery in x - f space into a single optimization problem, leading to iterative, joint estimation of time-varying convolution kernels and missing signals in k - t space. In the Kalman-smoother calibration, motion-induced uncertainties over the entire time frames were included in modeling state transition while a coil-dependent noise statistic in describing measurement process. The sparse signal recovery iteratively alternates with the self-calibration to tackle the ill-conditioning problem potentially resulting from insufficient calibrating signals. Simulations and experiments were performed using both the proposed and conventional methods for comparison, revealing that the proposed k - t SPARKS yields higher signal-to-error ratio and superior temporal fidelity in both breath-hold and free-breathing cardiac applications over all reduction factors.
Breathing SPACE-a practical approach to the breathless patient.
Hopkinson, Nicholas S; Baxter, Noel
2017-01-30
Breathlessness is a common symptom that may have multiple causes in any one individual and causes that may change over time. Breathlessness campaigns encourage people to see their General Practitioner if they are unduly breathless. Members of the London Respiratory Network collaborated to develop a tool that would encourage a holistic approach to breathlessness, which was applicable both at the time of diagnosis and during ongoing management. This has led to the development of the aide memoire "Breathing SPACE", which encompasses five key themes-smoking, pulmonary disease, anxiety/psychosocial factors, cardiac disease, and exercise/fitness. A particular concern was to ensure that high-value interventions (smoking cessation and exercise interventions) are prioritised across the life-course and throughout the course of disease management. The approach is relevant both to well people and in those with an underling diagnosis or diagnoses. The inclusion of anxiety draws attention to the importance of mental health issues. Parity of esteem requires the physical health problems of people with mental illness to be addressed. The SPACE mnemonic also addresses the problem of underdiagnosis of heart disease in people with lung disease and vice versa, as well as the systematic undertreatment of these conditions where they do co-occur.
A New Time-Space Accurate Scheme for Hyperbolic Problems. 1; Quasi-Explicit Case
NASA Technical Reports Server (NTRS)
Sidilkover, David
1998-01-01
This paper presents a new discretization scheme for hyperbolic systems of conservations laws. It satisfies the TVD property and relies on the new high-resolution mechanism which is compatible with the genuinely multidimensional approach proposed recently. This work can be regarded as a first step towards extending the genuinely multidimensional approach to unsteady problems. Discontinuity capturing capabilities and accuracy of the scheme are verified by a set of numerical tests.
Man in space - A time for perspective. [crew performance on Space Shuttle-Spacelab program
NASA Technical Reports Server (NTRS)
Winter, D. L.
1975-01-01
Factors affecting crew performances in long-term space flights are examined with emphasis on the Space Shuttle-Spacelab program. Biomedical investigations carried out during four Skylab missions indicate that initially rapid changes in certain physiological parameters, notably in cardiovascular response and red-blood-cell levels, lead to an adapted condition. Calcium loss remains a potential problem. Space Shuttle environmental control and life-support systems are described together with technology facilitating performance of mission objectives in a weightless environment. It is concluded that crew requirements are within the physical and psychological capability of astronauts, but the extent to which nonastronaut personnel will be able to participate without extensive training and pre-conditioning remains to be determined.
A Real-Time Greedy-Index Dispatching Policy for using PEVs to Provide Frequency Regulation Service
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ke, Xinda; Wu, Di; Lu, Ning
This article presents a real-time greedy-index dispatching policy (GIDP) for using plug-in electric vehicles (PEVs) to provide frequency regulation services. A new service cost allocation mechanism is proposed to award PEVs based on the amount of service they provided, while considering compensations for delayed-charging and reduction of battery lifetime due to participation of the service. The GIDP transforms the optimal dispatch problem from a high-dimensional space into a one-dimensional space while preserving the solution optimality. When solving the transformed problem in real-time, the global optimality of the GIDP solution can be guaranteed by mathematically proved “indexability”. Because the GIDP indexmore » can be calculated upon the PEV’s arrival and used for the entire decision making process till its departure, the computational burden is minimized and the complexity of the aggregator dispatch process is significantly reduced. Finally, simulation results are used to evaluate the proposed GIDP, and to demonstrate the potential profitability from providing frequency regulation service by using PEVs.« less
A Real-Time Greedy-Index Dispatching Policy for using PEVs to Provide Frequency Regulation Service
Ke, Xinda; Wu, Di; Lu, Ning
2017-09-18
This article presents a real-time greedy-index dispatching policy (GIDP) for using plug-in electric vehicles (PEVs) to provide frequency regulation services. A new service cost allocation mechanism is proposed to award PEVs based on the amount of service they provided, while considering compensations for delayed-charging and reduction of battery lifetime due to participation of the service. The GIDP transforms the optimal dispatch problem from a high-dimensional space into a one-dimensional space while preserving the solution optimality. When solving the transformed problem in real-time, the global optimality of the GIDP solution can be guaranteed by mathematically proved “indexability”. Because the GIDP indexmore » can be calculated upon the PEV’s arrival and used for the entire decision making process till its departure, the computational burden is minimized and the complexity of the aggregator dispatch process is significantly reduced. Finally, simulation results are used to evaluate the proposed GIDP, and to demonstrate the potential profitability from providing frequency regulation service by using PEVs.« less
The disposal of nuclear waste in space
NASA Technical Reports Server (NTRS)
Burns, R. E.
1978-01-01
The important problem of disposal of nuclear waste in space is addressed. A prior study proposed carrying only actinide wastes to space, but the present study assumes that all actinides and all fission products are to be carried to space. It is shown that nuclear waste in the calcine (oxide) form can be packaged in a container designed to provide thermal control, radiation shielding, mechanical containment, and an abort reentry thermal protection system. This package can be transported to orbit via the Space Shuttle. A second Space Shuttle delivers an oxygen-hydrogen orbit transfer vehicle to a rendezvous compatible orbit and the mated OTV and waste package are sent to the preferred destination. Preferred locations are either a lunar crater or a solar orbit. Shuttle traffic densities (which vary in time) are given and the safety of space disposal of wastes discussed.
Triangular node for Transmission-Line Modeling (TLM) applied to bio-heat transfer.
Milan, Hugo F M; Gebremedhin, Kifle G
2016-12-01
Transmission-Line Modeling (TLM) is a numerical method used to solve complex and time-domain bio-heat transfer problems. In TLM, rectangles are used to discretize two-dimensional problems. The drawback in using rectangular shapes is that instead of refining only the domain of interest, a large additional domain will also be refined in the x and y axes, which results in increased computational time and memory space. In this paper, we developed a triangular node for TLM applied to bio-heat transfer that does not have the drawback associated with the rectangular nodes. The model includes heat source, blood perfusion (advection), boundary conditions and initial conditions. The boundary conditions could be adiabatic, temperature, heat flux, or convection. A matrix equation for TLM, which simplifies the solution of time-domain problems or solves steady-state problems, was also developed. The predicted results were compared against results obtained from the solution of a simplified two-dimensional problem, and they agreed within 1% for a mesh length of triangular faces of 59µm±9µm (mean±standard deviation) and a time step of 1ms. Copyright © 2016 Elsevier Ltd. All rights reserved.
Program Helps Decompose Complex Design Systems
NASA Technical Reports Server (NTRS)
Rogers, James L., Jr.; Hall, Laura E.
1995-01-01
DeMAID (Design Manager's Aid for Intelligent Decomposition) computer program is knowledge-based software system for ordering sequence of modules and identifying possible multilevel structure for design problems such as large platforms in outer space. Groups modular subsystems on basis of interactions among them. Saves considerable amount of money and time in total design process, particularly in new design problem in which order of modules has not been defined. Originally written for design problems, also applicable to problems containing modules (processes) that take inputs and generate outputs. Available in three machine versions: Macintosh written in Symantec's Think C 3.01, Sun, and SGI IRIS in C language.
Lame problem for a multilayer viscoelastic hollow ball with regard to inhomogeneous aging
NASA Astrophysics Data System (ADS)
Davtyan, Z. A.; Mirzoyan, S. Y.; Gasparyan, A. V.
2018-04-01
Determination of characteristics of the stress strain state of compound viscoelastic bodies is of both theoretical and practical interest. In the present paper, the Lamé problem is investigated for an uneven-aged multilayer viscoelastic hollow ball in the framework of N. Kh. Arutyunyan’s theory of creep of nonhomogeneously aging bodies [1, 2]. Solving this problem reduces to solving an inhomogeneous finite-difference equation of second order that contains operators with coordinates of time and space. The obtained formulas allow one to determine the required contact stresses and other mechanical characteristics of the problem related to uneven age of contacting balls.
The use of x-ray pulsar-based navigation method for interplanetary flight
NASA Astrophysics Data System (ADS)
Yang, Bo; Guo, Xingcan; Yang, Yong
2009-07-01
As interplanetary missions are increasingly complex, the existing unique mature interplanetary navigation method mainly based on radiometric tracking techniques of Deep Space Network can not meet the rising demands of autonomous real-time navigation. This paper studied the applications for interplanetary flights of a new navigation technology under rapid development-the X-ray pulsar-based navigation for spacecraft (XPNAV), and valued its performance with a computer simulation. The XPNAV is an excellent autonomous real-time navigation method, and can provide comprehensive navigation information, including position, velocity, attitude, attitude rate and time. In the paper the fundamental principles and time transformation of the XPNAV were analyzed, and then the Delta-correction XPNAV blending the vehicles' trajectory dynamics with the pulse time-of-arrival differences at nominal and estimated spacecraft locations within an Unscented Kalman Filter (UKF) was discussed with a background mission of Mars Pathfinder during the heliocentric transferring orbit. The XPNAV has an intractable problem of integer pulse phase cycle ambiguities similar to the GPS carrier phase navigation. This article innovatively proposed the non-ambiguity assumption approach based on an analysis of the search space array method to resolve pulse phase cycle ambiguities between the nominal position and estimated position of the spacecraft. The simulation results show that the search space array method are computationally intensive and require long processing time when the position errors are large, and the non-ambiguity assumption method can solve ambiguity problem quickly and reliably. It is deemed that autonomous real-time integrated navigation system of the XPNAV blending with DSN, celestial navigation, inertial navigation and so on will be the development direction of interplanetary flight navigation system in the future.
NASA Astrophysics Data System (ADS)
Ray, Nadja; Rupp, Andreas; Knabner, Peter
2016-04-01
Soil is arguably the most prominent example of a natural porous medium that is composed of a porous matrix and a pore space. Within this framework and in terms of soil's heterogeneity, we first consider transport and fluid flow at the pore scale. From there, we develop a mechanistic model and upscale it mathematically to transfer our model from the small scale to that of the mesoscale (laboratory scale). The mathematical framework of (periodic) homogenization (in principal) rigorously facilitates such processes by exactly computing the effective coefficients/parameters by means of the pore geometry and processes. In our model, various small-scale soil processes may be taken into account: molecular diffusion, convection, drift emerging from electric forces, and homogeneous reactions of chemical species in a solvent. Additionally, our model may consider heterogeneous reactions at the porous matrix, thus altering both the porosity and the matrix. Moreover, our model may additionally address biophysical processes, such as the growth of biofilms and how this affects the shape of the pore space. Both of the latter processes result in an intrinsically variable soil structure in space and time. Upscaling such models under the assumption of a locally periodic setting must be performed meticulously to preserve information regarding the complex coupling of processes in the evolving heterogeneous medium. Generally, a micro-macro model emerges that is then comprised of several levels of couplings: Macroscopic equations that describe the transport and fluid flow at the scale of the porous medium (mesoscale) include averaged time- and space-dependent coefficient functions. These functions may be explicitly computed by means of auxiliary cell problems (microscale). Finally, the pore space in which the cell problems are defined is time- and space dependent and its geometry inherits information from the transport equation's solutions. Numerical computations using mixed finite elements and potentially random initial data, e.g. that of porosity, complement our theoretical results. Our investigations contribute to the theoretical understanding of the link between soil formation and soil functions. This general framework may be applied to various problems in soil science for a range of scales, such as the formation and turnover of microaggregates or soil remediation.
Boundaries on Range-Range Constrained Admissible Regions for Optical Space Surveillance
NASA Astrophysics Data System (ADS)
Gaebler, J. A.; Axelrad, P.; Schumacher, P. W., Jr.
We propose a new type of admissible-region analysis for track initiation in multi-satellite problems when apparent angles measured at known stations are the only observable. The goal is to create an efficient and parallelizable algorithm for computing initial candidate orbits for a large number of new targets. It takes at least three angles-only observations to establish an orbit by traditional means. Thus one is faced with a problem that requires N-choose-3 sets of calculations to test every possible combination of the N observations. An alternative approach is to reduce the number of combinations by making hypotheses of the range to a target along the observed line-of-sight. If realistic bounds on the range are imposed, consistent with a given partition of the space of orbital elements, a pair of range possibilities can be evaluated via Lambert’s method to find candidate orbits for that that partition, which then requires Nchoose- 2 times M-choose-2 combinations, where M is the average number of range hypotheses per observation. The contribution of this work is a set of constraints that establish bounds on the range-range hypothesis region for a given element-space partition, thereby minimizing M. Two effective constraints were identified, which together, constrain the hypothesis region in range-range space to nearly that of the true admissible region based on an orbital partition. The first constraint is based on the geometry of the vacant orbital focus. The second constraint is based on time-of-flight and Lagrange’s form of Kepler’s equation. A complete and efficient parallelization of the problem is possible on this approach because the element partitions can be arbitrary and can be handled independently of each other.
Models of Small-Scale Patchiness
NASA Technical Reports Server (NTRS)
McGillicuddy, D. J.
2001-01-01
Patchiness is perhaps the most salient characteristic of plankton populations in the ocean. The scale of this heterogeneity spans many orders of magnitude in its spatial extent, ranging from planetary down to microscale. It has been argued that patchiness plays a fundamental role in the functioning of marine ecosystems, insofar as the mean conditions may not reflect the environment to which organisms are adapted. Understanding the nature of this patchiness is thus one of the major challenges of oceanographic ecology. The patchiness problem is fundamentally one of physical-biological-chemical interactions. This interconnection arises from three basic sources: (1) ocean currents continually redistribute dissolved and suspended constituents by advection; (2) space-time fluctuations in the flows themselves impact biological and chemical processes, and (3) organisms are capable of directed motion through the water. This tripartite linkage poses a difficult challenge to understanding oceanic ecosystems: differentiation between the three sources of variability requires accurate assessment of property distributions in space and time, in addition to detailed knowledge of organismal repertoires and the processes by which ambient conditions control the rates of biological and chemical reactions. Various methods of observing the ocean tend to lie parallel to the axes of the space/time domain in which these physical-biological-chemical interactions take place. Given that a purely observational approach to the patchiness problem is not tractable with finite resources, the coupling of models with observations offers an alternative which provides a context for synthesis of sparse data with articulations of fundamental principles assumed to govern functionality of the system. In a sense, models can be used to fill the gaps in the space/time domain, yielding a framework for exploring the controls on spatially and temporally intermittent processes. The following discussion highlights only a few of the multitude of models which have yielded insight into the dynamics of plankton patchiness. In addition, this particular collection of examples is intended to furnish some exposure to the diversity of modeling approaches which can be brought to bear on the problem. These approaches range from abstract theoretical models intended to elucidate specific processes, to complex numerical formulations which can be used to actually simulate observed distributions in detail.
ERIC Educational Resources Information Center
Schoel, Jim; Butler, Steve; Murray, Mark; Gass, Mike; Carrick, Moe
2001-01-01
Presents five group problem-solving initiatives for use in adventure and experiential settings, focusing on conflict resolution, corporate workplace issues, or adjustment to change. Includes target group, group size, time and space needs, activity level, overview, goals, props, instructions, and suggestions for framing and debriefing the…
USDA-ARS?s Scientific Manuscript database
Allowing feces left on transport coops to dry is an effective way to reduce numbers of viable Campylobacter left by positive flocks. The problem with this approach is that poultry processors do not have the time, space or resources to maintain several times the minimum number of transport cages that...
Chandrasekhar equations for infinite dimensional systems
NASA Technical Reports Server (NTRS)
Ito, K.; Powers, R.
1985-01-01
The existence of Chandrasekhar equations for linear time-invariant systems defined on Hilbert spaces is investigated. An important consequence is that the solution to the evolutional Riccati equation is strongly differentiable in time, and that a strong solution of the Riccati differential equation can be defined. A discussion of the linear-quadratic optimal-control problem for hereditary differential systems is also included.
Second-order numerical solution of time-dependent, first-order hyperbolic equations
NASA Technical Reports Server (NTRS)
Shah, Patricia L.; Hardin, Jay
1995-01-01
A finite difference scheme is developed to find an approximate solution of two similar hyperbolic equations, namely a first-order plane wave and spherical wave problem. Finite difference approximations are made for both the space and time derivatives. The result is a conditionally stable equation yielding an exact solution when the Courant number is set to one.
An Implicit Characteristic Based Method for Electromagnetics
NASA Technical Reports Server (NTRS)
Beggs, John H.; Briley, W. Roger
2001-01-01
An implicit characteristic-based approach for numerical solution of Maxwell's time-dependent curl equations in flux conservative form is introduced. This method combines a characteristic based finite difference spatial approximation with an implicit lower-upper approximate factorization (LU/AF) time integration scheme. This approach is advantageous for three-dimensional applications because the characteristic differencing enables a two-factor approximate factorization that retains its unconditional stability in three space dimensions, and it does not require solution of tridiagonal systems. Results are given both for a Fourier analysis of stability, damping and dispersion properties, and for one-dimensional model problems involving propagation and scattering for free space and dielectric materials using both uniform and nonuniform grids. The explicit Finite Difference Time Domain Method (FDTD) algorithm is used as a convenient reference algorithm for comparison. The one-dimensional results indicate that for low frequency problems on a highly resolved uniform or nonuniform grid, this LU/AF algorithm can produce accurate solutions at Courant numbers significantly greater than one, with a corresponding improvement in efficiency for simulating a given period of time. This approach appears promising for development of dispersion optimized LU/AF schemes for three dimensional applications.
NASA Astrophysics Data System (ADS)
Mohebbi, Akbar
2018-02-01
In this paper we propose two fast and accurate numerical methods for the solution of multidimensional space fractional Ginzburg-Landau equation (FGLE). In the presented methods, to avoid solving a nonlinear system of algebraic equations and to increase the accuracy and efficiency of method, we split the complex problem into simpler sub-problems using the split-step idea. For a homogeneous FGLE, we propose a method which has fourth-order of accuracy in time component and spectral accuracy in space variable and for nonhomogeneous one, we introduce another scheme based on the Crank-Nicolson approach which has second-order of accuracy in time variable. Due to using the Fourier spectral method for fractional Laplacian operator, the resulting schemes are fully diagonal and easy to code. Numerical results are reported in terms of accuracy, computational order and CPU time to demonstrate the accuracy and efficiency of the proposed methods and to compare the results with the analytical solutions. The results show that the present methods are accurate and require low CPU time. It is illustrated that the numerical results are in good agreement with the theoretical ones.
Sex Differences in Mathematics Attainment at GCE Ordinary Level
ERIC Educational Resources Information Center
Wood, Robert
1976-01-01
In a comparison of mathematical abilities of boys and girls, after allowing for school effects, boys are seen to excel on problems involving scale or measurement, probability, and space-time relationships. Possible explanations for the observed differences are made. (Author/AV)
Toward Scalable Trustworthy Computing Using the Human-Physiology-Immunity Metaphor
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hively, Lee M; Sheldon, Frederick T
The cybersecurity landscape consists of an ad hoc patchwork of solutions. Optimal cybersecurity is difficult for various reasons: complexity, immense data and processing requirements, resource-agnostic cloud computing, practical time-space-energy constraints, inherent flaws in 'Maginot Line' defenses, and the growing number and sophistication of cyberattacks. This article defines the high-priority problems and examines the potential solution space. In that space, achieving scalable trustworthy computing and communications is possible through real-time knowledge-based decisions about cyber trust. This vision is based on the human-physiology-immunity metaphor and the human brain's ability to extract knowledge from data and information. The article outlines future steps towardmore » scalable trustworthy systems requiring a long-term commitment to solve the well-known challenges.« less
Periodically driven ergodic and many-body localized quantum systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ponte, Pedro; Department of Physics and Astronomy, University of Waterloo, ON N2L 3G1; Chandran, Anushya
2015-02-15
We study dynamics of isolated quantum many-body systems whose Hamiltonian is switched between two different operators periodically in time. The eigenvalue problem of the associated Floquet operator maps onto an effective hopping problem. Using the effective model, we establish conditions on the spectral properties of the two Hamiltonians for the system to localize in energy space. We find that ergodic systems always delocalize in energy space and heat up to infinite temperature, for both local and global driving. In contrast, many-body localized systems with quenched disorder remain localized at finite energy. We support our conclusions by numerical simulations of disorderedmore » spin chains. We argue that our results hold for general driving protocols, and discuss their experimental implications.« less
On integrability of the Killing equation
NASA Astrophysics Data System (ADS)
Houri, Tsuyoshi; Tomoda, Kentaro; Yasui, Yukinori
2018-04-01
Killing tensor fields have been thought of as describing the hidden symmetry of space(-time) since they are in one-to-one correspondence with polynomial first integrals of geodesic equations. Since many problems in classical mechanics can be formulated as geodesic problems in curved space and spacetime, solving the defining equation for Killing tensor fields (the Killing equation) is a powerful way to integrate equations of motion. Thus it has been desirable to formulate the integrability conditions of the Killing equation, which serve to determine the number of linearly independent solutions and also to restrict the possible forms of solutions tightly. In this paper, we show the prolongation for the Killing equation in a manner that uses Young symmetrizers. Using the prolonged equations, we provide the integrability conditions explicitly.
STS-114: Discovery Impromptu Briefing
NASA Technical Reports Server (NTRS)
2005-01-01
Dr. Griffin, NASA Administrator, is accompanied by members of The U.S. House of Representatives in this STS-114 Discovery Impromptu briefing. The U.S. House of Representatives present include: Sherwood Boehlert, House Science Committee Chairman, Senator Hutchinson, Sheila Jackson, 18th Congressional District Texas, Al Green, 9th Congressional District, Representative Jim Davis, Florida, and Gene Green, 29th District, Texas. Griffin talks about the problem that occurred with the external fuel tank sensor of the Space Shuttle Discovery and the effort NASA is pursuing to track the problem, and identify the root cause. He answers questions from the news media about the next steps for the Space Shuttle Discovery, time frame for the launch, and activities for the astronauts for the next few days.
Generalized Minimum-Time Follow-up Approaches Applied to Tasking Electro-Optical Sensor Tasking
NASA Astrophysics Data System (ADS)
Murphy, T. S.; Holzinger, M. J.
This work proposes a methodology for tasking of sensors to search an area of state space for a particular object, group of objects, or class of objects. This work creates a general unified mathematical framework for analyzing reacquisition, search, scheduling, and custody operations. In particular, this work looks at searching for unknown space object(s) with prior knowledge in the form of a set, which can be defined via an uncorrelated track, region of state space, or a variety of other methods. The follow-up tasking can occur from a variable location and time, which often requires searching a large region of the sky. This work analyzes the area of a search region over time to inform a time optimal search method. Simulation work looks at analyzing search regions relative to a particular sensor, and testing a tasking algorithm to search through the region. The tasking algorithm is also validated on a reacquisition problem with a telescope system at Georgia Tech.
2007-12-14
KENNEDY SPACE CENTER, FLA. -- Space shuttle Atlantis' external tank and solid rocket boosters will be the subject of a tanking test with the Time Domain Reflectometry, or TDR, test equipment whose wiring is being calibrated Dec. 14. The shuttle's planned launches on Dec. 6 and Dec. 9 were postponed because of false readings from the part of the ECO system that monitors the liquid hydrogen section of the tank. The liftoff date from NASA's Kennedy Space Center, Florida, is now targeted for Jan. 10, depending on the resolution of the problem in the fuel sensor system. Photo credit: NASA/Kim Shiflett
NASA Technical Reports Server (NTRS)
1977-01-01
A brief description of recommended supporting research and technology items resulting from the space station analysis study is provided. Descriptions include the title; the status with respect to the state of the art; the justification; the technical plan including objectives and technical approach; resource requirements categorized by manpower, specialized facilities, and funding in 1977 dollars; and also the target schedule. The goal is to provide high confidence in the solutions for the various functional system development problems, and to do so within a time period compatible with the overall evolutionary space construction base schedule.
An Introduction to the History of Aerospace Medicine
NASA Technical Reports Server (NTRS)
Tarver, William J.
2007-01-01
This viewgraph presentation reviews the history of aviation and space travel from the precursors of aviation to the lighter than air aviation to the heavier than air to the space travel. The improvement of heavier than air travel, in all dimensions from length of time traveled, to altitude and speed of travel between the first to World War I is reviewed. An understanding of the medical problems associated with air travel was begun. The beginning of the position of flight surgeons is reviewed. The advancement of flight from air to space flight is shown using photos.
An overview of expert systems. [artificial intelligence
NASA Technical Reports Server (NTRS)
Gevarter, W. B.
1982-01-01
An expert system is defined and its basic structure is discussed. The knowledge base, the inference engine, and uses of expert systems are discussed. Architecture is considered, including choice of solution direction, reasoning in the presence of uncertainty, searching small and large search spaces, handling large search spaces by transforming them and by developing alternative or additional spaces, and dealing with time. Existing expert systems are reviewed. Tools for building such systems, construction, and knowledge acquisition and learning are discussed. Centers of research and funding sources are listed. The state-of-the-art, current problems, required research, and future trends are summarized.
NASA Astrophysics Data System (ADS)
Pin, Victor Gómez
In his book about the Categories (that is about the ultimate elements of classification and order), in the chapter concerning the quantity (IV, 20) Aristotle says that this concept recovers two kinds of modalities: the discrete quantity and the continuous quantity and he gives as examples the number for the first one; line, surface, solid, times and space for the second one. The main philosophical problem raised by this text is to determine which of the two modalities of the quantity has the ontological priority over the other (given two concepts A and B, we assume that A has ontological priority over B if every entity that possesses the quality B possesses necessarily the quality A). The problem is magnified by the fact that space, which in some part of Aristotle's Physics is mentioned not only as a category properly speaking but even as the main category whose power can be amazing, is in the evoked text of the Categories's Book reduced to expression of the continuum, and sharing this condition with time. In this matter the controversy is constant through the common history of Science and Philosophy.
Santitissadeekorn, N; Bollt, E M
2007-06-01
In this paper, we present an approach to approximate the Frobenius-Perron transfer operator from a sequence of time-ordered images, that is, a movie dataset. Unlike time-series data, successive images do not provide a direct access to a trajectory of a point in a phase space; more precisely, a pixel in an image plane. Therefore, we reconstruct the velocity field from image sequences based on the infinitesimal generator of the Frobenius-Perron operator. Moreover, we relate this problem to the well-known optical flow problem from the computer vision community and we validate the continuity equation derived from the infinitesimal operator as a constraint equation for the optical flow problem. Once the vector field and then a discrete transfer operator are found, then, in addition, we present a graph modularity method as a tool to discover basin structure in the phase space. Together with a tool to reconstruct a velocity field, this graph-based partition method provides us with a way to study transport behavior and other ergodic properties of measurable dynamical systems captured only through image sequences.
Earthquake Declustering via a Nearest-Neighbor Approach in Space-Time-Magnitude Domain
NASA Astrophysics Data System (ADS)
Zaliapin, I. V.; Ben-Zion, Y.
2016-12-01
We propose a new method for earthquake declustering based on nearest-neighbor analysis of earthquakes in space-time-magnitude domain. The nearest-neighbor approach was recently applied to a variety of seismological problems that validate the general utility of the technique and reveal the existence of several different robust types of earthquake clusters. Notably, it was demonstrated that clustering associated with the largest earthquakes is statistically different from that of small-to-medium events. In particular, the characteristic bimodality of the nearest-neighbor distances that helps separating clustered and background events is often violated after the largest earthquakes in their vicinity, which is dominated by triggered events. This prevents using a simple threshold between the two modes of the nearest-neighbor distance distribution for declustering. The current study resolves this problem hence extending the nearest-neighbor approach to the problem of earthquake declustering. The proposed technique is applied to seismicity of different areas in California (San Jacinto, Coso, Salton Sea, Parkfield, Ventura, Mojave, etc.), as well as to the global seismicity, to demonstrate its stability and efficiency in treating various clustering types. The results are compared with those of alternative declustering methods.
The joint effects of person and situation factors on stress in spaceflight.
Endler, Norman S
2004-07-01
Psychologists can play an important role in spaceflight and exploration. Indeed, their input at four specific stages in spaceflight can help to ensure successful missions. Psychologists have a role to play in: 1) Selection; 2) Training; 3) Real or Simulated Space Missions; and 4) Post-Flight Problems. During selection, psychologists can provide guidelines as to the appropriate coping styles for reacting to mission-related stressors. During training, psychologists can help astronauts to plan for, and cope with, problems (e.g., group conflict). Training in social sensitivity, or in specific coping styles (i.e., to be high in task-oriented coping), would be beneficial so that astronauts are able to deal with unforeseen problems. Task-oriented coping is related to control and efficiency, and reduces stress and anxiety. Emotion-oriented coping relates to anxiety for controllable situations, and task-oriented coping is most efficacious. A primary role for psychologists would be the investigation of what problems could arise from living with others in a limited space and for long periods of times. As such, investigations into group dynamics, physical and psychological stress caused by such an environment, and the lack of normal ways to deal with these problems should all be considered. It is also clear that the impact of spaceflight on an individual does not end with physical re-entry. On the contrary, the astronaut will also have to make a psychological post-flight "re-entry" readjustment to life on Earth. Psychologists have an important role to play during all four stages of the space program, especially in regards to person by situation interactions. That is, space is a novel environment for human beings, and we need to investigate how we can better improve the fit between astronauts and space situational stressors. Not only are person by situation interactions relevant for each of the four stages discussed above, but each of the stages interacts with one another bidirectionally and even multidirectionally.
A simple method to calculate first-passage time densities with arbitrary initial conditions
NASA Astrophysics Data System (ADS)
Nyberg, Markus; Ambjörnsson, Tobias; Lizana, Ludvig
2016-06-01
Numerous applications all the way from biology and physics to economics depend on the density of first crossings over a boundary. Motivated by the lack of general purpose analytical tools for computing first-passage time densities (FPTDs) for complex problems, we propose a new simple method based on the independent interval approximation (IIA). We generalise previous formulations of the IIA to include arbitrary initial conditions as well as to deal with discrete time and non-smooth continuous time processes. We derive a closed form expression for the FPTD in z and Laplace-transform space to a boundary in one dimension. Two classes of problems are analysed in detail: discrete time symmetric random walks (Markovian) and continuous time Gaussian stationary processes (Markovian and non-Markovian). Our results are in good agreement with Langevin dynamics simulations.
NASA Astrophysics Data System (ADS)
Plimak, L. I.; Fleischhauer, M.; Olsen, M. K.; Collett, M. J.
2003-01-01
We present an introduction to phase-space techniques (PST) based on a quantum-field-theoretical (QFT) approach. In addition to bridging the gap between PST and QFT, our approach results in a number of generalizations of the PST. First, for problems where the usual PST do not result in a genuine Fokker-Planck equation (even after phase-space doubling) and hence fail to produce a stochastic differential equation (SDE), we show how the system in question may be approximated via stochastic difference equations (SΔE). Second, we show that introducing sources into the SDE’s (or SΔE’s) generalizes them to a full quantum nonlinear stochastic response problem (thus generalizing Kubo’s linear reaction theory to a quantum nonlinear stochastic response theory). Third, we establish general relations linking quantum response properties of the system in question to averages of operator products ordered in a way different from time normal. This extends PST to a much wider assemblage of operator products than are usually considered in phase-space approaches. In all cases, our approach yields a very simple and straightforward way of deriving stochastic equations in phase space.
Consistent Adjoint Driven Importance Sampling using Space, Energy and Angle
DOE Office of Scientific and Technical Information (OSTI.GOV)
Peplow, Douglas E.; Mosher, Scott W; Evans, Thomas M
2012-08-01
For challenging radiation transport problems, hybrid methods combine the accuracy of Monte Carlo methods with the global information present in deterministic methods. One of the most successful hybrid methods is CADIS Consistent Adjoint Driven Importance Sampling. This method uses a deterministic adjoint solution to construct a biased source distribution and consistent weight windows to optimize a specific tally in a Monte Carlo calculation. The method has been implemented into transport codes using just the spatial and energy information from the deterministic adjoint and has been used in many applications to compute tallies with much higher figures-of-merit than analog calculations. CADISmore » also outperforms user-supplied importance values, which usually take long periods of user time to develop. This work extends CADIS to develop weight windows that are a function of the position, energy, and direction of the Monte Carlo particle. Two types of consistent source biasing are presented: one method that biases the source in space and energy while preserving the original directional distribution and one method that biases the source in space, energy, and direction. Seven simple example problems are presented which compare the use of the standard space/energy CADIS with the new space/energy/angle treatments.« less
Time Synchronization and Distribution Mechanisms for Space Networks
NASA Technical Reports Server (NTRS)
Woo, Simon S.; Gao, Jay L.; Clare, Loren P.; Mills, David L.
2011-01-01
This work discusses research on the problems of synchronizing and distributing time information between spacecraft based on the Network Time Protocol (NTP), where NTP is a standard time synchronization protocol widely used in the terrestrial network. The Proximity-1 Space Link Interleaved Time Synchronization (PITS) Protocol was designed and developed for synchronizing spacecraft that are in proximity where proximity is less than 100,000 km distant. A particular application is synchronization between a Mars orbiter and rover. Lunar scenarios as well as outer-planet deep space mother-ship-probe missions may also apply. Spacecraft with more accurate time information functions as a time-server, and the other spacecraft functions as a time-client. PITS can be easily integrated and adaptable to the CCSDS Proximity-1 Space Link Protocol with minor modifications. In particular, PITS can take advantage of the timestamping strategy that underlying link layer functionality provides for accurate time offset calculation. The PITS algorithm achieves time synchronization with eight consecutive space network time packet exchanges between two spacecraft. PITS can detect and avoid possible errors from receiving duplicate and out-of-order packets by comparing with the current state variables and timestamps. Further, PITS is able to detect error events and autonomously recover from unexpected events that can possibly occur during the time synchronization and distribution process. This capability achieves an additional level of protocol protection on top of CRC or Error Correction Codes. PITS is a lightweight and efficient protocol, eliminating the needs for explicit frame sequence number and long buffer storage. The PITS protocol is capable of providing time synchronization and distribution services for a more general domain where multiple entities need to achieve time synchronization using a single point-to-point link.
Highly Parallel Alternating Directions Algorithm for Time Dependent Problems
NASA Astrophysics Data System (ADS)
Ganzha, M.; Georgiev, K.; Lirkov, I.; Margenov, S.; Paprzycki, M.
2011-11-01
In our work, we consider the time dependent Stokes equation on a finite time interval and on a uniform rectangular mesh, written in terms of velocity and pressure. For this problem, a parallel algorithm based on a novel direction splitting approach is developed. Here, the pressure equation is derived from a perturbed form of the continuity equation, in which the incompressibility constraint is penalized in a negative norm induced by the direction splitting. The scheme used in the algorithm is composed of two parts: (i) velocity prediction, and (ii) pressure correction. This is a Crank-Nicolson-type two-stage time integration scheme for two and three dimensional parabolic problems in which the second-order derivative, with respect to each space variable, is treated implicitly while the other variable is made explicit at each time sub-step. In order to achieve a good parallel performance the solution of the Poison problem for the pressure correction is replaced by solving a sequence of one-dimensional second order elliptic boundary value problems in each spatial direction. The parallel code is implemented using the standard MPI functions and tested on two modern parallel computer systems. The performed numerical tests demonstrate good level of parallel efficiency and scalability of the studied direction-splitting-based algorithm.
Karaoulis, M.; Revil, A.; Werkema, D.D.; Minsley, B.J.; Woodruff, W.F.; Kemna, A.
2011-01-01
Induced polarization (more precisely the magnitude and phase of impedance of the subsurface) is measured using a network of electrodes located at the ground surface or in boreholes. This method yields important information related to the distribution of permeability and contaminants in the shallow subsurface. We propose a new time-lapse 3-D modelling and inversion algorithm to image the evolution of complex conductivity over time. We discretize the subsurface using hexahedron cells. Each cell is assigned a complex resistivity or conductivity value. Using the finite-element approach, we model the in-phase and out-of-phase (quadrature) electrical potentials on the 3-D grid, which are then transformed into apparent complex resistivity. Inhomogeneous Dirichlet boundary conditions are used at the boundary of the domain. The calculation of the Jacobian matrix is based on the principles of reciprocity. The goal of time-lapse inversion is to determine the change in the complex resistivity of each cell of the spatial grid as a function of time. Each model along the time axis is called a 'reference space model'. This approach can be simplified into an inverse problem looking for the optimum of several reference space models using the approximation that the material properties vary linearly in time between two subsequent reference models. Regularizations in both space domain and time domain reduce inversion artefacts and improve the stability of the inversion problem. In addition, the use of the time-lapse equations allows the simultaneous inversion of data obtained at different times in just one inversion step (4-D inversion). The advantages of this new inversion algorithm are demonstrated on synthetic time-lapse data resulting from the simulation of a salt tracer test in a heterogeneous random material described by an anisotropic semi-variogram. ?? 2011 The Authors Geophysical Journal International ?? 2011 RAS.
NASA Technical Reports Server (NTRS)
Wie, Bong; Liu, Qiang
1992-01-01
Both feedback and feedforward control approaches for uncertain dynamical systems (in particular, with uncertainty in structural mode frequency) are investigated. The control objective is to achieve a fast settling time (high performance) and robustness (insensitivity) to plant uncertainty. Preshaping of an ideal, time optimal control input using a tapped-delay filter is shown to provide a fast settling time with robust performance. A robust, non-minimum-phase feedback controller is synthesized with particular emphasis on its proper implementation for a non-zero set-point control problem. It is shown that a properly designed, feedback controller performs well, as compared with a time optimal open loop controller with special preshaping for performance robustness. Also included are two separate papers by the same authors on this subject.
The nonequilibrium quantum many-body problem as a paradigm for extreme data science
NASA Astrophysics Data System (ADS)
Freericks, J. K.; Nikolić, B. K.; Frieder, O.
2014-12-01
Generating big data pervades much of physics. But some problems, which we call extreme data problems, are too large to be treated within big data science. The nonequilibrium quantum many-body problem on a lattice is just such a problem, where the Hilbert space grows exponentially with system size and rapidly becomes too large to fit on any computer (and can be effectively thought of as an infinite-sized data set). Nevertheless, much progress has been made with computational methods on this problem, which serve as a paradigm for how one can approach and attack extreme data problems. In addition, viewing these physics problems from a computer-science perspective leads to new approaches that can be tried to solve more accurately and for longer times. We review a number of these different ideas here.
On the Ck-embedding of Lorentzian manifolds in Ricci-flat spaces
NASA Astrophysics Data System (ADS)
Avalos, R.; Dahia, F.; Romero, C.
2018-05-01
In this paper, we investigate the problem of non-analytic embeddings of Lorentzian manifolds in Ricci-flat semi-Riemannian spaces. In order to do this, we first review some relevant results in the area and then motivate both the mathematical and physical interests in this problem. We show that any n-dimensional compact Lorentzian manifold (Mn, g), with g in the Sobolev space Hs+3, s >n/2 , admits an isometric embedding in a (2n + 2)-dimensional Ricci-flat semi-Riemannian manifold. The sharpest result available for these types of embeddings, in the general setting, comes as a corollary of Greene's remarkable embedding theorems R. Greene [Mem. Am. Math. Soc. 97, 1 (1970)], which guarantee the embedding of a compact n-dimensional semi-Riemannian manifold into an n(n + 5)-dimensional semi-Euclidean space, thereby guaranteeing the embedding into a Ricci-flat space with the same dimension. The theorem presented here improves this corollary in n2 + 3n - 2 codimensions by replacing the Riemann-flat condition with the Ricci-flat one from the beginning. Finally, we will present a corollary of this theorem, which shows that a compact strip in an n-dimensional globally hyperbolic space-time can be embedded in a (2n + 2)-dimensional Ricci-flat semi-Riemannian manifold.
Graph pyramids as models of human problem solving
NASA Astrophysics Data System (ADS)
Pizlo, Zygmunt; Li, Zheng
2004-05-01
Prior theories have assumed that human problem solving involves estimating distances among states and performing search through the problem space. The role of mental representation in those theories was minimal. Results of our recent experiments suggest that humans are able to solve some difficult problems quickly and accurately. Specifically, in solving these problems humans do not seem to rely on distances or on search. It is quite clear that producing good solutions without performing search requires a very effective mental representation. In this paper we concentrate on studying the nature of this representation. Our theory takes the form of a graph pyramid. To verify the psychological plausibility of this theory we tested subjects in a Euclidean Traveling Salesman Problem in the presence of obstacles. The role of the number and size of obstacles was tested for problems with 6-50 cities. We analyzed the effect of experimental conditions on solution time per city and on solution error. The main result is that time per city is systematically affected only by the size of obstacles, but not by their number, or by the number of cities.
Jouriles, Ernest N; Vu, Nicole L; McDonald, Renee; Rosenfield, David
2014-12-01
This research examined whether children's threat and self-blame appraisals regarding interparental conflict and their beliefs about the justifiability of aggression predicted children's externalizing problems in families in which there had been recent severe intimate partner violence (IPV). Participants were 106 children (62 boys, 44 girls) aged 7 to 10 and their mothers. Families in which there had been recent severe IPV were recruited during their stay at a domestic violence shelter. Children completed measures of threat, self-blame, beliefs about the justifiability of aggression, and externalizing problems. Mothers completed a measure of children's externalizing problems. Measures were completed at 3 time points, spaced 6 months apart. In multilevel modeling analyses, threat appraisals and beliefs about the justifiability of aggression were positively associated with children's reports of externalizing problems concurrently, and self-blame appraisals were positively associated with mothers' reports of externalizing problems concurrently. In prospective analyses, beliefs about the justifiability of aggression at 1 time point were positively associated with children's reports of externalizing problems 6 months later. The results provide partial support for the modified cognitive-contextual framework.
Parallel computing method for simulating hydrological processesof large rivers under climate change
NASA Astrophysics Data System (ADS)
Wang, H.; Chen, Y.
2016-12-01
Climate change is one of the proverbial global environmental problems in the world.Climate change has altered the watershed hydrological processes in time and space distribution, especially in worldlarge rivers.Watershed hydrological process simulation based on physically based distributed hydrological model can could have better results compared with the lumped models.However, watershed hydrological process simulation includes large amount of calculations, especially in large rivers, thus needing huge computing resources that may not be steadily available for the researchers or at high expense, this seriously restricted the research and application. To solve this problem, the current parallel method are mostly parallel computing in space and time dimensions.They calculate the natural features orderly thatbased on distributed hydrological model by grid (unit, a basin) from upstream to downstream.This articleproposes ahigh-performancecomputing method of hydrological process simulation with high speedratio and parallel efficiency.It combinedthe runoff characteristics of time and space of distributed hydrological model withthe methods adopting distributed data storage, memory database, distributed computing, parallel computing based on computing power unit.The method has strong adaptability and extensibility,which means it canmake full use of the computing and storage resources under the condition of limited computing resources, and the computing efficiency can be improved linearly with the increase of computing resources .This method can satisfy the parallel computing requirements ofhydrological process simulation in small, medium and large rivers.
High-Frequency Subband Compressed Sensing MRI Using Quadruplet Sampling
Sung, Kyunghyun; Hargreaves, Brian A
2013-01-01
Purpose To presents and validates a new method that formalizes a direct link between k-space and wavelet domains to apply separate undersampling and reconstruction for high- and low-spatial-frequency k-space data. Theory and Methods High- and low-spatial-frequency regions are defined in k-space based on the separation of wavelet subbands, and the conventional compressed sensing (CS) problem is transformed into one of localized k-space estimation. To better exploit wavelet-domain sparsity, CS can be used for high-spatial-frequency regions while parallel imaging can be used for low-spatial-frequency regions. Fourier undersampling is also customized to better accommodate each reconstruction method: random undersampling for CS and regular undersampling for parallel imaging. Results Examples using the proposed method demonstrate successful reconstruction of both low-spatial-frequency content and fine structures in high-resolution 3D breast imaging with a net acceleration of 11 to 12. Conclusion The proposed method improves the reconstruction accuracy of high-spatial-frequency signal content and avoids incoherent artifacts in low-spatial-frequency regions. This new formulation also reduces the reconstruction time due to the smaller problem size. PMID:23280540
Single-mode fiber systems for deep space communication network
NASA Technical Reports Server (NTRS)
Lutes, G.
1982-01-01
The present investigation is concerned with the development of single-mode optical fiber distribution systems. It is pointed out that single-mode fibers represent potentially a superior medium for the distribution of frequency and timing reference signals and wideband (400 MHz) IF signals. In this connection, single-mode fibers have the potential to improve the capability and precision of NASA's Deep Space Network (DSN). Attention is given to problems related to precise time synchronization throughout the DSN, questions regarding the selection of a transmission medium, and the function of the distribution systems, taking into account specific improvements possible by an employment of single-mode fibers.
Canonical quantization of general relativity in discrete space-times.
Gambini, Rodolfo; Pullin, Jorge
2003-01-17
It has long been recognized that lattice gauge theory formulations, when applied to general relativity, conflict with the invariance of the theory under diffeomorphisms. We analyze discrete lattice general relativity and develop a canonical formalism that allows one to treat constrained theories in Lorentzian signature space-times. The presence of the lattice introduces a "dynamical gauge" fixing that makes the quantization of the theories conceptually clear, albeit computationally involved. The problem of a consistent algebra of constraints is automatically solved in our approach. The approach works successfully in other field theories as well, including topological theories. A simple cosmological application exhibits quantum elimination of the singularity at the big bang.
NASA Astrophysics Data System (ADS)
Perez, Alejandro
2015-04-01
In an approach to quantum gravity where space-time arises from coarse graining of fundamentally discrete structures, black hole formation and subsequent evaporation can be described by a unitary evolution without the problems encountered by the standard remnant scenario or the schemes where information is assumed to come out with the radiation during evaporation (firewalls and complementarity). The final state is purified by correlations with the fundamental pre-geometric structures (in the sense of Wheeler), which are available in such approaches, and, like defects in the underlying space-time weave, can carry zero energy.
International Polar Research and Space Weather
NASA Astrophysics Data System (ADS)
Lanzerotti, Louis J.
2009-02-01
The fiftieth anniversary of the International Geophysical Year (IGY), currently celebrated in the 2007-2009 International Polar Year (IPY), highlights space weather's heritage from polar research. The polar regions were still very much "terra incognito" 50 years ago. At the same time, communications technologies had significantly advanced since the time of the second IPY, in 1932-1933. Yet even before the second IPY, several directors of international meteorological services stated in a 1928 resolution that "increased knowledge [of the polar regions] will be of practical application to problems connected with terrestrial magnetism, marine and aerial navigation, wireless telegraphy and weather forecasting" (see http://scaa.usask.ca/gallery/northern/currie/en_polaryear.shtml).
Can a quantum state over time resemble a quantum state at a single time?
NASA Astrophysics Data System (ADS)
Horsman, Dominic; Heunen, Chris; Pusey, Matthew F.; Barrett, Jonathan; Spekkens, Robert W.
2017-09-01
The standard formalism of quantum theory treats space and time in fundamentally different ways. In particular, a composite system at a given time is represented by a joint state, but the formalism does not prescribe a joint state for a composite of systems at different times. If there were a way of defining such a joint state, this would potentially permit a more even-handed treatment of space and time, and would strengthen the existing analogy between quantum states and classical probability distributions. Under the assumption that the joint state over time is an operator on the tensor product of single-time Hilbert spaces, we analyse various proposals for such a joint state, including one due to Leifer and Spekkens, one due to Fitzsimons, Jones and Vedral, and another based on discrete Wigner functions. Finding various problems with each, we identify five criteria for a quantum joint state over time to satisfy if it is to play a role similar to the standard joint state for a composite system: that it is a Hermitian operator on the tensor product of the single-time Hilbert spaces; that it represents probabilistic mixing appropriately; that it has the appropriate classical limit; that it has the appropriate single-time marginals; that composing over multiple time steps is associative. We show that no construction satisfies all these requirements. If Hermiticity is dropped, then there is an essentially unique construction that satisfies the remaining four criteria.
Mean Field Approach to the Giant Wormhole Problem
NASA Astrophysics Data System (ADS)
Gamba, A.; Kolokolov, I.; Martellini, M.
We introduce a gaussian probability density for the space-time distribution of worm-holes, thus taking effectively into account wormhole interaction. Using a mean-field approximation for the free energy, we show that giant wormholes are probabilistically suppressed in a homogenous isotropic “large” universe.
Applications of dynamic scheduling technique to space related problems: Some case studies
NASA Astrophysics Data System (ADS)
Nakasuka, Shinichi; Ninomiya, Tetsujiro
1994-10-01
The paper discusses the applications of 'Dynamic Scheduling' technique, which has been invented for the scheduling of Flexible Manufacturing System, to two space related scheduling problems: operation scheduling of a future space transportation system, and resource allocation in a space system with limited resources such as space station or space shuttle.
Remote Operations of the Deep Space Network Radio Science Subsystem
NASA Astrophysics Data System (ADS)
Caetta, J.; Asmar, S.; Abbate, S.; Connally, M.; Goltz, G.
1998-04-01
The capability for scientists to remotely control systems located at the Deep Space Network facilities only recently has been incorporated in the design and implementation of new equipment. However, time lines for the implementation, distribution, and operational readiness of such systems can extend much farther into the future than the users can wait. The Radio Science Systems Group was faced with just that circumstance; new hardware was not scheduled to become operational for several years, but the increasing number of experiments and configurations for Cassini, Galileo, Mars missions, and other flight projects made that time frame impractical because of the associated increasing risk of not acquiring critical data. Therefore, a method of interfacing with the current radio science subsystem has been developed and used with a high degree of success, although with occasional problems due to this capability not having been originally designed into the system. This article discusses both the method and the problems involved in integrating this new (remote) method of control with a legacy system.
Application of artificial intelligence to search ground-state geometry of clusters
NASA Astrophysics Data System (ADS)
Lemes, Maurício Ruv; Marim, L. R.; dal Pino, A.
2002-08-01
We introduce a global optimization procedure, the neural-assisted genetic algorithm (NAGA). It combines the power of an artificial neural network (ANN) with the versatility of the genetic algorithm. This method is suitable to solve optimization problems that depend on some kind of heuristics to limit the search space. If a reasonable amount of data is available, the ANN can ``understand'' the problem and provide the genetic algorithm with a selected population of elements that will speed up the search for the optimum solution. We tested the method in a search for the ground-state geometry of silicon clusters. We trained the ANN with information about the geometry and energetics of small silicon clusters. Next, the ANN learned how to restrict the configurational space for larger silicon clusters. For Si10 and Si20, we noticed that the NAGA is at least three times faster than the ``pure'' genetic algorithm. As the size of the cluster increases, it is expected that the gain in terms of time will increase as well.
Gravitational wave astronomy: needle in a haystack.
Cornish, Neil J
2013-02-13
A worldwide array of highly sensitive ground-based interferometers stands poised to usher in a new era in astronomy with the first direct detection of gravitational waves. The data from these instruments will provide a unique perspective on extreme astrophysical objects, such as neutron stars and black holes, and will allow us to test Einstein's theory of gravity in the strong field, dynamical regime. To fully realize these goals, we need to solve some challenging problems in signal processing and inference, such as finding rare and weak signals that are buried in non-stationary and non-Gaussian instrument noise, dealing with high-dimensional model spaces, and locating what are often extremely tight concentrations of posterior mass within the prior volume. Gravitational wave detection using space-based detectors and pulsar timing arrays bring with them the additional challenge of having to isolate individual signals that overlap one another in both time and frequency. Promising solutions to these problems will be discussed, along with some of the challenges that remain.
Compact time- and space-integrating SAR processor: design and development status
NASA Astrophysics Data System (ADS)
Haney, Michael W.; Levy, James J.; Christensen, Marc P.; Michael, Robert R., Jr.; Mock, Michael M.
1994-06-01
Progress toward a flight demonstration of the acousto-optic time- and space- integrating real-time SAR image formation processor program is reported. The concept overcomes the size and power consumption limitations of electronic approaches by using compact, rugged, and low-power analog optical signal processing techniques for the most computationally taxing portions of the SAR imaging problem. Flexibility and performance are maintained by the use of digital electronics for the critical low-complexity filter generation and output image processing functions. The results reported include tests of a laboratory version of the concept, a description of the compact optical design that will be implemented, and an overview of the electronic interface and controller modules of the flight-test system.
NASA Astrophysics Data System (ADS)
Sumin, M. I.
2015-06-01
A parametric nonlinear programming problem in a metric space with an operator equality constraint in a Hilbert space is studied assuming that its lower semicontinuous value function at a chosen individual parameter value has certain subdifferentiability properties in the sense of nonlinear (nonsmooth) analysis. Such subdifferentiability can be understood as the existence of a proximal subgradient or a Fréchet subdifferential. In other words, an individual problem has a corresponding generalized Kuhn-Tucker vector. Under this assumption, a stable sequential Kuhn-Tucker theorem in nondifferential iterative form is proved and discussed in terms of minimizing sequences on the basis of the dual regularization method. This theorem provides necessary and sufficient conditions for the stable construction of a minimizing approximate solution in the sense of Warga in the considered problem, whose initial data can be approximately specified. A substantial difference of the proved theorem from its classical same-named analogue is that the former takes into account the possible instability of the problem in the case of perturbed initial data and, as a consequence, allows for the inherited instability of classical optimality conditions. This theorem can be treated as a regularized generalization of the classical Uzawa algorithm to nonlinear programming problems. Finally, the theorem is applied to the "simplest" nonlinear optimal control problem, namely, to a time-optimal control problem.
Near-Optimal Guidance Method for Maximizing the Reachable Domain of Gliding Aircraft
NASA Astrophysics Data System (ADS)
Tsuchiya, Takeshi
This paper proposes a guidance method for gliding aircraft by using onboard computers to calculate a near-optimal trajectory in real-time, and thereby expanding the reachable domain. The results are applicable to advanced aircraft and future space transportation systems that require high safety. The calculation load of the optimal control problem that is used to maximize the reachable domain is too large for current computers to calculate in real-time. Thus the optimal control problem is divided into two problems: a gliding distance maximization problem in which the aircraft motion is limited to a vertical plane, and an optimal turning flight problem in a horizontal direction. First, the former problem is solved using a shooting method. It can be solved easily because its scale is smaller than that of the original problem, and because some of the features of the optimal solution are obtained in the first part of this paper. Next, in the latter problem, the optimal bank angle is computed from the solution of the former; this is an analytical computation, rather than an iterative computation. Finally, the reachable domain obtained from the proposed near-optimal guidance method is compared with that obtained from the original optimal control problem.
Sleep Does Not Promote Solving Classical Insight Problems and Magic Tricks
Schönauer, Monika; Brodt, Svenja; Pöhlchen, Dorothee; Breßmer, Anja; Danek, Amory H.; Gais, Steffen
2018-01-01
During creative problem solving, initial solution attempts often fail because of self-imposed constraints that prevent us from thinking out of the box. In order to solve a problem successfully, the problem representation has to be restructured by combining elements of available knowledge in novel and creative ways. It has been suggested that sleep supports the reorganization of memory representations, ultimately aiding problem solving. In this study, we systematically tested the effect of sleep and time on problem solving, using classical insight tasks and magic tricks. Solving these tasks explicitly requires a restructuring of the problem representation and may be accompanied by a subjective feeling of insight. In two sessions, 77 participants had to solve classical insight problems and magic tricks. The two sessions either occurred consecutively or were spaced 3 h apart, with the time in between spent either sleeping or awake. We found that sleep affected neither general solution rates nor the number of solutions accompanied by sudden subjective insight. Our study thus adds to accumulating evidence that sleep does not provide an environment that facilitates the qualitative restructuring of memory representations and enables problem solving. PMID:29535620
The medical implications of space tourism.
Tarzwell, R
2000-06-01
Commercial space travel may soon be a reality. If so, microgravity, high acceleration, and radiation exposure, all known hazards, will be accessible to the general public. Therefore, space tourism has medical implications. Even though the first flights will feature space exposure times of only a few minutes, the potential may someday exist for exposure times long enough to warrant careful consideration of the potential hazards to the space-faring public. The effects of acceleration and microgravity exposure are well known on the corps of astronauts and cosmonauts. The effects of space radiation are partially known on astronauts, but much remains to be discovered. However, there are problems using astronaut data to make inferences about the general public. Astronauts are not necessarily representative of the general public, since they are highly fit, highly screened individuals. Astronaut data can tell us very little about the potential hazards of microgravity in pediatric, obstetric and geriatric populations, all of whom are potential space tourists. Key issues in standard setting will be determining acceptable limits of pre-existing disease and inferring medical standards from mission profiles. It will not be a trivial task drafting minimal medical standards for commercial space travel. It will require the collaboration of space medicine physicians, making the best guesses possible, based on limited amounts of data, with limited applicability. A helpful departure point may be the USAF Class 3 medical standard, applicable to NASA payload specialists. It is time to begin preliminary discussions toward defining those standards. acceleration, aerospace medicine, medical standards, microgravity, radiation, space, space tourism, environmental hazards, environmental medicine.
Jahanian, Hesamoddin; Soltanian-Zadeh, Hamid; Hossein-Zadeh, Gholam-Ali
2005-09-01
To present novel feature spaces, based on multiscale decompositions obtained by scalar wavelet and multiwavelet transforms, to remedy problems associated with high dimension of functional magnetic resonance imaging (fMRI) time series (when they are used directly in clustering algorithms) and their poor signal-to-noise ratio (SNR) that limits accurate classification of fMRI time series according to their activation contents. Using randomization, the proposed method finds wavelet/multiwavelet coefficients that represent the activation content of fMRI time series and combines them to define new feature spaces. Using simulated and experimental fMRI data sets, the proposed feature spaces are compared to the cross-correlation (CC) feature space and their performances are evaluated. In these studies, the false positive detection rate is controlled using randomization. To compare different methods, several points of the receiver operating characteristics (ROC) curves, using simulated data, are estimated and compared. The proposed features suppress the effects of confounding signals and improve activation detection sensitivity. Experimental results show improved sensitivity and robustness of the proposed method compared to the conventional CC analysis. More accurate and sensitive activation detection can be achieved using the proposed feature spaces compared to CC feature space. Multiwavelet features show superior detection sensitivity compared to the scalar wavelet features. (c) 2005 Wiley-Liss, Inc.
Aerospace vehicle design, spacecraft section. Volume 2
NASA Technical Reports Server (NTRS)
1988-01-01
The next major step in the evolution of the space program is the exploration of the planet Mars. In preparation for this, much research is needed on the problem of surveying the planet surface. An aircraft appears to be a viable solution because it can carry men and equipment large distances in a short period of time as compared with ground transportation. The problems and design of an aircraft which would be able to survey the planet Mars are examined.
On dual and three space problems for the compact approximation property
NASA Astrophysics Data System (ADS)
Choi, Changsun; Kim, Ju Myung
2006-11-01
We introduce the properties W*D and BW*D for the dual space of a Banach space. And then solve the dual problem for the compact approximation property (CAP): if X* has the CAP and the W*D, then X has the CAP. Also, we solve the three space problem for the CAP: for example, if M is a closed subspace of a Banach space such that M[perpendicular] is complemented in X* and X* has the W*D, then X has the CAP whenever X/M has the CAP and M has the bounded CAP. Corresponding problems for the bounded compact approximation property are also addressed.
Some solutions of the general three body problem in form space
NASA Astrophysics Data System (ADS)
Titov, Vladimir
2018-05-01
Some solutions of three body problem with equal masses are first considered in form space. The solutions in usual euclidean space may be restored from these form space solutions. If constant energy h < 0, the trajectories are located inside of Hill's surface. Without loss of generality due to scale symmetry we can set h = -1. Such surface has a simple form in form space. Solutions of isosceles and rectilinear three body problems lie within Hill's curve; periodic solutions of free fall three body problem start in one point of this curve, and finish in another. The solutions are illustrated by number of figures.
NASA Astrophysics Data System (ADS)
Qi, Bin; Guo, Linli; Zhang, Zhixian
2016-07-01
Space life science and life support engineering are prominent problems in manned deep space exploration mission. Some typical problems are discussed in this paper, including long-term life support problem, physiological effect and defense of varying extraterrestrial environment. The causes of these problems are developed for these problems. To solve these problems, research on space life science and space medical-engineering should be conducted. In the aspect of space life science, the study of space gravity biology should focus on character of physiological effect in long term zero gravity, co-regulation of physiological systems, impact on stem cells in space, etc. The study of space radiation biology should focus on target effect and non-target effect of radiation, carcinogenicity of radiation, spread of radiation damage in life system, etc. The study of basic biology of space life support system should focus on theoretical basis and simulating mode of constructing the life support system, filtration and combination of species, regulation and optimization method of life support system, etc. In the aspect of space medical-engineering, the study of bio-regenerative life support technology should focus on plants cultivation technology, animal-protein production technology, waste treatment technology, etc. The study of varying gravity defense technology should focus on biological and medical measures to defend varying gravity effect, generation and evaluation of artificial gravity, etc. The study of extraterrestrial environment defense technology should focus on risk evaluation of radiation, monitoring and defending of radiation, compound prevention and removal technology of dust, etc. At last, a case of manned lunar base is analyzed, in which the effective schemes of life support system, defense of varying gravity, defense of extraterrestrial environment are advanced respectively. The points in this paper can be used as references for intensive study on key technologies.
Electricity Usage Scheduling in Smart Building Environments Using Smart Devices
Lee, Eunji; Bahn, Hyokyung
2013-01-01
With the recent advances in smart grid technologies as well as the increasing dissemination of smart meters, the electricity usage of every moment can be detected in modern smart building environments. Thus, the utility company adopts different price of electricity at each time slot considering the peak time. This paper presents a new electricity usage scheduling algorithm for smart buildings that adopts real-time pricing of electricity. The proposed algorithm detects the change of electricity prices by making use of a smart device and changes the power mode of each electric device dynamically. Specifically, we formulate the electricity usage scheduling problem as a real-time task scheduling problem and show that it is a complex search problem that has an exponential time complexity. An efficient heuristic based on genetic algorithms is performed on a smart device to cut down the huge searching space and find a reasonable schedule within a feasible time budget. Experimental results with various building conditions show that the proposed algorithm reduces the electricity charge of a smart building by 25.6% on average and up to 33.4%. PMID:24453860
Electricity usage scheduling in smart building environments using smart devices.
Lee, Eunji; Bahn, Hyokyung
2013-01-01
With the recent advances in smart grid technologies as well as the increasing dissemination of smart meters, the electricity usage of every moment can be detected in modern smart building environments. Thus, the utility company adopts different price of electricity at each time slot considering the peak time. This paper presents a new electricity usage scheduling algorithm for smart buildings that adopts real-time pricing of electricity. The proposed algorithm detects the change of electricity prices by making use of a smart device and changes the power mode of each electric device dynamically. Specifically, we formulate the electricity usage scheduling problem as a real-time task scheduling problem and show that it is a complex search problem that has an exponential time complexity. An efficient heuristic based on genetic algorithms is performed on a smart device to cut down the huge searching space and find a reasonable schedule within a feasible time budget. Experimental results with various building conditions show that the proposed algorithm reduces the electricity charge of a smart building by 25.6% on average and up to 33.4%.
A solution quality assessment method for swarm intelligence optimization algorithms.
Zhang, Zhaojun; Wang, Gai-Ge; Zou, Kuansheng; Zhang, Jianhua
2014-01-01
Nowadays, swarm intelligence optimization has become an important optimization tool and wildly used in many fields of application. In contrast to many successful applications, the theoretical foundation is rather weak. Therefore, there are still many problems to be solved. One problem is how to quantify the performance of algorithm in finite time, that is, how to evaluate the solution quality got by algorithm for practical problems. It greatly limits the application in practical problems. A solution quality assessment method for intelligent optimization is proposed in this paper. It is an experimental analysis method based on the analysis of search space and characteristic of algorithm itself. Instead of "value performance," the "ordinal performance" is used as evaluation criteria in this method. The feasible solutions were clustered according to distance to divide solution samples into several parts. Then, solution space and "good enough" set can be decomposed based on the clustering results. Last, using relative knowledge of statistics, the evaluation result can be got. To validate the proposed method, some intelligent algorithms such as ant colony optimization (ACO), particle swarm optimization (PSO), and artificial fish swarm algorithm (AFS) were taken to solve traveling salesman problem. Computational results indicate the feasibility of proposed method.
NASA Technical Reports Server (NTRS)
Hagstrom, Thomas; Hariharan, S. I.; Maccamy, R. C.
1993-01-01
We consider the solution of scattering problems for the wave equation using approximate boundary conditions at artificial boundaries. These conditions are explicitly viewed as approximations to an exact boundary condition satisfied by the solution on the unbounded domain. We study the short and long term behavior of the error. It is provided that, in two space dimensions, no local in time, constant coefficient boundary operator can lead to accurate results uniformly in time for the class of problems we consider. A variable coefficient operator is developed which attains better accuracy (uniformly in time) than is possible with constant coefficient approximations. The theory is illustrated by numerical examples. We also analyze the proposed boundary conditions using energy methods, leading to asymptotically correct error bounds.
Discontinuous Galerkin method for coupled problems of compressible flow and elastic structures
NASA Astrophysics Data System (ADS)
Kosík, A.; Feistauer, M.; Hadrava, M.; Horáček, J.
2013-10-01
This paper is concerned with the numerical simulation of the interaction of 2D compressible viscous flow and an elastic structure. We consider the model of dynamical linear elasticity. Each individual problem is discretized in space by the discontinuous Galerkin method (DGM). For the time discretization we can use either the BDF (backward difference formula) method or also the DGM. The time dependence of the domain occupied by the fluid is given by the deformation of the elastic structure adjacent to the flow domain. It is treated with the aid of the Arbitrary Lagrangian-Eulerian (ALE) method. The fluid-structure interaction, given by transient conditions, is realized by an iterative process. The developed method is applied to the simulation of the biomechanical problem containing the onset of the voice production.
Time: The Biggest Pattern in Natural History Research
NASA Astrophysics Data System (ADS)
Gontier, Nathalie
2016-10-01
We distinguish between four cosmological transitions in the history of Western intellectual thought, and focus on how these cosmologies differentially define matter, space and time. We demonstrate that how time is conceptualized significantly impacts a cosmology's notion on causality, and hone in on how time is conceptualized differentially in modern physics and evolutionary biology. The former conflates time with space into a single space-time continuum and focuses instead on the movement of matter, while the evolutionary sciences have a tradition to understand time as a given when they cartography how organisms change across generations over or in time, thereby proving the phenomenon of evolution. The gap becomes more fundamental when we take into account that phenomena studied by chrono-biologists demonstrate that numerous organisms, including humans, have evolved a "sense" of time. And micro-evolutionary/genetic, meso-evolutionary/developmental and macro-evolutionary phenomena including speciation and extinction not only occur by different evolutionary modes and at different rates, they are also timely phenomena that follow different periodicities. This article focusses on delineating the problem by finding its historical roots. We conclude that though time might be an obsolete concept for the physical sciences, it is crucial for the evolutionary sciences where evolution is defined as the change that biological individuals undergo in/over or through time.
1974-07-18
so on. We say OK, we will work with Honda . We made an agreement and got the engine. They don’t have the problem solved at allfor us, but they have me...with our prechamber Honda CVCC program. hhile there are more design problems to be solved with the PROCO engine, and more time may be necessary to...But Mr. Sugiura was quoted by the agency ofcials as explaining Honda still has problems with fuel economy and driving performances and that it Is Impos
NASA Astrophysics Data System (ADS)
Polprasert, Jirawadee; Ongsakul, Weerakorn; Dieu, Vo Ngoc
2011-06-01
This paper proposes a self-organizing hierarchical particle swarm optimization (SPSO) with time-varying acceleration coefficients (TVAC) for solving economic dispatch (ED) problem with non-smooth functions including multiple fuel options (MFO) and valve-point loading effects (VPLE). The proposed SPSO with TVAC is the new approach optimizer and good performance for solving ED problems. It can handle the premature convergence of the problem by re-initialization of velocity whenever particles are stagnated in the search space. To properly control both local and global explorations of the swarm during the optimization process, the performance of TVAC is included. The proposed method is tested in different ED problems with non-smooth cost functions and the obtained results are compared to those from many other methods in the literature. The results have revealed that the proposed SPSO with TVAC is effective in finding higher quality solutions for non-smooth ED problems than many other methods.
NASA Astrophysics Data System (ADS)
Bezruchko, Konstantin; Davidov, Albert
2009-01-01
In the given article scientific and technical complex for modeling, researching and testing of rocket-space vehicles' power installations which was created in Power Source Laboratory of National Aerospace University "KhAI" is described. This scientific and technical complex gives the opportunity to replace the full-sized tests on model tests and to reduce financial and temporary inputs at modeling, researching and testing of rocket-space vehicles' power installations. Using the given complex it is possible to solve the problems of designing and researching of rocket-space vehicles' power installations efficiently, and also to provide experimental researches of physical processes and tests of solar and chemical batteries of rocket-space complexes and space vehicles. Scientific and technical complex also allows providing accelerated tests, diagnostics, life-time control and restoring of chemical accumulators for rocket-space vehicles' power supply systems.
Parallel-In-Time For Moving Meshes
DOE Office of Scientific and Technical Information (OSTI.GOV)
Falgout, R. D.; Manteuffel, T. A.; Southworth, B.
2016-02-04
With steadily growing computational resources available, scientists must develop e ective ways to utilize the increased resources. High performance, highly parallel software has be- come a standard. However until recent years parallelism has focused primarily on the spatial domain. When solving a space-time partial di erential equation (PDE), this leads to a sequential bottleneck in the temporal dimension, particularly when taking a large number of time steps. The XBraid parallel-in-time library was developed as a practical way to add temporal parallelism to existing se- quential codes with only minor modi cations. In this work, a rezoning-type moving mesh is appliedmore » to a di usion problem and formulated in a parallel-in-time framework. Tests and scaling studies are run using XBraid and demonstrate excellent results for the simple model problem considered herein.« less
Brito, Irene; Mena, Filipe C
2017-08-01
We prove that, for a given spherically symmetric fluid distribution with tangential pressure on an initial space-like hypersurface with a time-like boundary, there exists a unique, local in time solution to the Einstein equations in a neighbourhood of the boundary. As an application, we consider a particular elastic fluid interior matched to a vacuum exterior.
NASA Technical Reports Server (NTRS)
Tempelman, W. H.
1973-01-01
The navigation and control of the space shuttle during atmospheric entry are discussed. A functional flow diagram presenting the basic approach to the deorbit targeting problem is presented. The major inputs to be considered are: (1) vehicle state vector, (2) landing site location, (3) entry interface parameters, (4) earliest desired time of landing, and (5) maximum cross range. Mathematical models of the navigational procedures based on controlled thrust times are developed.
Chandrasekhar equations for infinite dimensional systems
NASA Technical Reports Server (NTRS)
Ito, K.; Powers, R. K.
1985-01-01
Chandrasekhar equations are derived for linear time invariant systems defined on Hilbert spaces using a functional analytic technique. An important consequence of this is that the solution to the evolutional Riccati equation is strongly differentiable in time and one can define a strong solution of the Riccati differential equation. A detailed discussion on the linear quadratic optimal control problem for hereditary differential systems is also included.
Tracking change over time: River flooding
,
2014-01-01
The objective of the Tracking Change Over Time lesson plan is to get students excited about studying the changing Earth. Intended for students in grades 5-8, the lesson plan is flexible and may be used as a student self-guided tutorial or as a teacher-led class lesson. Enhance students' learning of geography, map reading, earth science, and problem solving by seeing landscape changes from space.
NASA Astrophysics Data System (ADS)
Gualandi, Adriano; Serpelloni, Enrico; Elina Belardinelli, Maria; Bonafede, Maurizio; Pezzo, Giuseppe; Tolomei, Cristiano
2015-04-01
A critical point in the analysis of ground displacement time series, as those measured by modern space geodetic techniques (primarly continuous GPS/GNSS and InSAR) is the development of data driven methods that allow to discern and characterize the different sources that generate the observed displacements. A widely used multivariate statistical technique is the Principal Component Analysis (PCA), which allows to reduce the dimensionality of the data space maintaining most of the variance of the dataset explained. It reproduces the original data using a limited number of Principal Components, but it also shows some deficiencies, since PCA does not perform well in finding the solution to the so-called Blind Source Separation (BSS) problem. The recovering and separation of the different sources that generate the observed ground deformation is a fundamental task in order to provide a physical meaning to the possible different sources. PCA fails in the BSS problem since it looks for a new Euclidean space where the projected data are uncorrelated. Usually, the uncorrelation condition is not strong enough and it has been proven that the BSS problem can be tackled imposing on the components to be independent. The Independent Component Analysis (ICA) is, in fact, another popular technique adopted to approach this problem, and it can be used in all those fields where PCA is also applied. An ICA approach enables us to explain the displacement time series imposing a fewer number of constraints on the model, and to reveal anomalies in the data such as transient deformation signals. However, the independence condition is not easy to impose, and it is often necessary to introduce some approximations. To work around this problem, we use a variational bayesian ICA (vbICA) method, which models the probability density function (pdf) of each source signal using a mix of Gaussian distributions. This technique allows for more flexibility in the description of the pdf of the sources, giving a more reliable estimate of them. Here we introduce the vbICA technique and present its application on synthetic data that simulate a GPS network recording ground deformation in a tectonically active region, with synthetic time-series containing interseismic, coseismic, and postseismic deformation, plus seasonal deformation, and white and coloured noise. We study the ability of the algorithm to recover the original (known) sources of deformation, and then apply it to a real scenario: the Emilia seismic sequence (2012, northern Italy), which is an example of seismic sequence occurred in a slowly converging tectonic setting, characterized by several local to regional anthropogenic or natural sources of deformation, mainly subsidence due to fluid withdrawal and sediments compaction. We apply both PCA and vbICA to displacement time-series recorded by continuous GPS and InSAR (Pezzo et al., EGU2015-8950).
Boundary particle method for Laplace transformed time fractional diffusion equations
NASA Astrophysics Data System (ADS)
Fu, Zhuo-Jia; Chen, Wen; Yang, Hai-Tian
2013-02-01
This paper develops a novel boundary meshless approach, Laplace transformed boundary particle method (LTBPM), for numerical modeling of time fractional diffusion equations. It implements Laplace transform technique to obtain the corresponding time-independent inhomogeneous equation in Laplace space and then employs a truly boundary-only meshless boundary particle method (BPM) to solve this Laplace-transformed problem. Unlike the other boundary discretization methods, the BPM does not require any inner nodes, since the recursive composite multiple reciprocity technique (RC-MRM) is used to convert the inhomogeneous problem into the higher-order homogeneous problem. Finally, the Stehfest numerical inverse Laplace transform (NILT) is implemented to retrieve the numerical solutions of time fractional diffusion equations from the corresponding BPM solutions. In comparison with finite difference discretization, the LTBPM introduces Laplace transform and Stehfest NILT algorithm to deal with time fractional derivative term, which evades costly convolution integral calculation in time fractional derivation approximation and avoids the effect of time step on numerical accuracy and stability. Consequently, it can effectively simulate long time-history fractional diffusion systems. Error analysis and numerical experiments demonstrate that the present LTBPM is highly accurate and computationally efficient for 2D and 3D time fractional diffusion equations.
Technology That's Ready and Able to Inspect Those Cables
NASA Technical Reports Server (NTRS)
2005-01-01
Attempting to locate a malfunctioning wire in a complex bundle of wires or in a cable that is concealed behind a wall is as difficult as trying to find a needle in a haystack. The result of such an effort can also be costly, time-consuming, and frustrating, whether it is the tedious process of examining cable connections for the Space Shuttle or troubleshooting a cable television hookup. Furthermore, other maintenance restrictions can compound the effort required to locate and repair a particular wiring problem. For example, on the Space Shuttle, once a repair is completed, all systems that have a wire passing through any of the connectors that were disconnected during troubleshooting are affected and, therefore, must undergo retesting, an arduous task that is completely unrelated to the original problem. In an effort to streamline wire inspection and maintenance, two contractors supporting NASA's Kennedy Space Center invented the Standing Wave Reflectometer (SWR) in 1999. In doing so, they leveraged technology that was first developed to detect problems that could lead to aircraft accidents, such as the one that resulted in the catastrophic failure of TWA flight 800 in 1996. The SWR performs a non-intrusive inspection that verifies the condition of electrical power and signal-distribution systems inside the Space Shuttle orbiters. Such testing reduces processing delays and ensures safe operation of these systems.
An efficient genetic algorithm for maximum coverage deployment in wireless sensor networks.
Yoon, Yourim; Kim, Yong-Hyuk
2013-10-01
Sensor networks have a lot of applications such as battlefield surveillance, environmental monitoring, and industrial diagnostics. Coverage is one of the most important performance metrics for sensor networks since it reflects how well a sensor field is monitored. In this paper, we introduce the maximum coverage deployment problem in wireless sensor networks and analyze the properties of the problem and its solution space. Random deployment is the simplest way to deploy sensor nodes but may cause unbalanced deployment and therefore, we need a more intelligent way for sensor deployment. We found that the phenotype space of the problem is a quotient space of the genotype space in a mathematical view. Based on this property, we propose an efficient genetic algorithm using a novel normalization method. A Monte Carlo method is adopted to design an efficient evaluation function, and its computation time is decreased without loss of solution quality using a method that starts from a small number of random samples and gradually increases the number for subsequent generations. The proposed genetic algorithms could be further improved by combining with a well-designed local search. The performance of the proposed genetic algorithm is shown by a comparative experimental study. When compared with random deployment and existing methods, our genetic algorithm was not only about twice faster, but also showed significant performance improvement in quality.
NASA Astrophysics Data System (ADS)
Novikova, Nataliya; Gusev, Oleg; Sugimoto, Manabu; Deshevaya, Elena; Levinskikh, Margarita; Sychev, Vladimir; Okuda, Takashi; Orlov, Oleg; Alekseev, Victor; Poddubko, Svetlana; Polikarpov, Nikolay
The planetary quarantine is one of the key problems of deep space exploration. Risks of the possible transfer of biological objects across interplanetary space should be necessarily assessed during space exploration. The risks associated with a possible transfer of biological objects and primarily microorganisms in interplanetary space is a priority for space studies We can assume, that on the exterior side of both unmanned and manned space stations there can be millions of microbial cells, many of which are in spore forms, the stability of which towards the unfavorable factors is extremely high. However, direct evidence to support this assumption, obtained only in recent years. “Biorisk” is an apparatus designed for conduction of space experiments focused on long-term exposition of latent stages of different forms of organism on the outer side of Russian Segment of International Space Station was developed and used in SSC RF - Institute for Biomedical Problems RAS. The purpose of this experiment is to determine the principle capability of preservation of life capacity in test-cultures of microorganisms during long-term exposure (comparable with the term of interplanetary flight) in space. The first experiment was performed using spores of bacteria (Bacillus) and fungi (Penicillium, Aspergillus and Cladosporium) housed in 3 boxes that were exposed to outer space for 7, 12 or 18 months. It was for the first time demonstrated that bacterial and fungal spores could survive an exposure to outer space during the time period comparable with the duration of a return mission to Mars. Moreover, the microbial strains proved viable and highly active. The second experiment was expanded by flying, in addition to the above spores, dormant forms of higher plants, insects, lower crustaceans and vertebrates. The 31-month experiment showed that, in spite of harsher than in the first study temperatures, some specimens remained viable and capable of further multiplication. In summary, our experiments provided evidence that not only bacterial and fungal spores but also dormant forms of organisms that reached higher levels of evolutionary development had the capability to survive a long-term exposure to outer space. This observation suggests that they can be transferred on outer walls of space platforms during interplanetary missions. Our findings are of scientific interest as well as of importance for the development of planetary quarantine concepts related to future space flight.
MGA trajectory planning with an ACO-inspired algorithm
NASA Astrophysics Data System (ADS)
Ceriotti, Matteo; Vasile, Massimiliano
2010-11-01
Given a set of celestial bodies, the problem of finding an optimal sequence of swing-bys, deep space manoeuvres (DSM) and transfer arcs connecting the elements of the set is combinatorial in nature. The number of possible paths grows exponentially with the number of celestial bodies. Therefore, the design of an optimal multiple gravity assist (MGA) trajectory is a NP-hard mixed combinatorial-continuous problem. Its automated solution would greatly improve the design of future space missions, allowing the assessment of a large number of alternative mission options in a short time. This work proposes to formulate the complete automated design of a multiple gravity assist trajectory as an autonomous planning and scheduling problem. The resulting scheduled plan will provide the optimal planetary sequence and a good estimation of the set of associated optimal trajectories. The trajectory model consists of a sequence of celestial bodies connected by two-dimensional transfer arcs containing one DSM. For each transfer arc, the position of the planet and the spacecraft, at the time of arrival, are matched by varying the pericentre of the preceding swing-by, or the magnitude of the launch excess velocity, for the first arc. For each departure date, this model generates a full tree of possible transfers from the departure to the destination planet. Each leaf of the tree represents a planetary encounter and a possible way to reach that planet. An algorithm inspired by ant colony optimization (ACO) is devised to explore the space of possible plans. The ants explore the tree from departure to destination adding one node at the time: every time an ant is at a node, a probability function is used to select a feasible direction. This approach to automatic trajectory planning is applied to the design of optimal transfers to Saturn and among the Galilean moons of Jupiter. Solutions are compared to those found through more traditional genetic-algorithm techniques.
NASA Technical Reports Server (NTRS)
Markopoulos, N.; Calise, A. J.
1993-01-01
The class of all piecewise time-continuous controllers tracking a given hypersurface in the state space of a dynamical system can be split by the present transformation technique into two disjoint classes; while the first of these contains all controllers which track the hypersurface in finite time, the second contains all controllers that track the hypersurface asymptotically. On this basis, a reformulation is presented for optimal control problems involving state-variable inequality constraints. If the state constraint is regarded as 'soft', there may exist controllers which are asymptotic, two-sided, and able to yield the optimal value of the performance index.
History of satellite missions and measurements of the Earth Radiation Budget (1957-1984)
NASA Technical Reports Server (NTRS)
House, F. B.; Gruber, A.; Hunt, G. E.; Mecherikunnel, A. T.
1986-01-01
The history of satellite missions and their measurements of the earth radiation budget from the beginning of the space age until the present time are reviewed. The survey emphasizes the early struggle to develop instrument systems to monitor reflected shortwave and emitted long-wave exitances from the earth, and the problems associated with the interpretation of these observations from space. In some instances, valuable data sets were developed from satellite measurements whose instruments were not specifically designed for earth radiation budget observations.
Li, Lixin; Losser, Travis; Yorke, Charles; Piltner, Reinhard
2014-09-03
Epidemiological studies have identified associations between mortality and changes in concentration of particulate matter. These studies have highlighted the public concerns about health effects of particulate air pollution. Modeling fine particulate matter PM2.5 exposure risk and monitoring day-to-day changes in PM2.5 concentration is a critical step for understanding the pollution problem and embarking on the necessary remedy. This research designs, implements and compares two inverse distance weighting (IDW)-based spatiotemporal interpolation methods, in order to assess the trend of daily PM2.5 concentration for the contiguous United States over the year of 2009, at both the census block group level and county level. Traditionally, when handling spatiotemporal interpolation, researchers tend to treat space and time separately and reduce the spatiotemporal interpolation problems to a sequence of snapshots of spatial interpolations. In this paper, PM2.5 data interpolation is conducted in the continuous space-time domain by integrating space and time simultaneously, using the so-called extension approach. Time values are calculated with the help of a factor under the assumption that spatial and temporal dimensions are equally important when interpolating a continuous changing phenomenon in the space-time domain. Various IDW-based spatiotemporal interpolation methods with different parameter configurations are evaluated by cross-validation. In addition, this study explores computational issues (computer processing speed) faced during implementation of spatiotemporal interpolation for huge data sets. Parallel programming techniques and an advanced data structure, named k-d tree, are adapted in this paper to address the computational challenges. Significant computational improvement has been achieved. Finally, a web-based spatiotemporal IDW-based interpolation application is designed and implemented where users can visualize and animate spatiotemporal interpolation results.
Li, Lixin; Losser, Travis; Yorke, Charles; Piltner, Reinhard
2014-01-01
Epidemiological studies have identified associations between mortality and changes in concentration of particulate matter. These studies have highlighted the public concerns about health effects of particulate air pollution. Modeling fine particulate matter PM2.5 exposure risk and monitoring day-to-day changes in PM2.5 concentration is a critical step for understanding the pollution problem and embarking on the necessary remedy. This research designs, implements and compares two inverse distance weighting (IDW)-based spatiotemporal interpolation methods, in order to assess the trend of daily PM2.5 concentration for the contiguous United States over the year of 2009, at both the census block group level and county level. Traditionally, when handling spatiotemporal interpolation, researchers tend to treat space and time separately and reduce the spatiotemporal interpolation problems to a sequence of snapshots of spatial interpolations. In this paper, PM2.5 data interpolation is conducted in the continuous space-time domain by integrating space and time simultaneously, using the so-called extension approach. Time values are calculated with the help of a factor under the assumption that spatial and temporal dimensions are equally important when interpolating a continuous changing phenomenon in the space-time domain. Various IDW-based spatiotemporal interpolation methods with different parameter configurations are evaluated by cross-validation. In addition, this study explores computational issues (computer processing speed) faced during implementation of spatiotemporal interpolation for huge data sets. Parallel programming techniques and an advanced data structure, named k-d tree, are adapted in this paper to address the computational challenges. Significant computational improvement has been achieved. Finally, a web-based spatiotemporal IDW-based interpolation application is designed and implemented where users can visualize and animate spatiotemporal interpolation results. PMID:25192146
A global approach to kinematic path planning to robots with holonomic and nonholonomic constraints
NASA Technical Reports Server (NTRS)
Divelbiss, Adam; Seereeram, Sanjeev; Wen, John T.
1993-01-01
Robots in applications may be subject to holonomic or nonholonomic constraints. Examples of holonomic constraints include a manipulator constrained through the contact with the environment, e.g., inserting a part, turning a crank, etc., and multiple manipulators constrained through a common payload. Examples of nonholonomic constraints include no-slip constraints on mobile robot wheels, local normal rotation constraints for soft finger and rolling contacts in grasping, and conservation of angular momentum of in-orbit space robots. The above examples all involve equality constraints; in applications, there are usually additional inequality constraints such as robot joint limits, self collision and environment collision avoidance constraints, steering angle constraints in mobile robots, etc. The problem of finding a kinematically feasible path that satisfies a given set of holonomic and nonholonomic constraints, of both equality and inequality types is addressed. The path planning problem is first posed as a finite time nonlinear control problem. This problem is subsequently transformed to a static root finding problem in an augmented space which can then be iteratively solved. The algorithm has shown promising results in planning feasible paths for redundant arms satisfying Cartesian path following and goal endpoint specifications, and mobile vehicles with multiple trailers. In contrast to local approaches, this algorithm is less prone to problems such as singularities and local minima.
ERIC Educational Resources Information Center
Leon, Warren
1992-01-01
Discusses themes and ideas for social studies teachers to use in teaching about energy. Addresses energy transitions, the rate of technology change, energy and economic growth, and comparing and developing new energy sources. Concludes that students need a historical and social context for understanding the environmental and economic problems that…
View of the SRB problems with Challenger after launch
NASA Technical Reports Server (NTRS)
1986-01-01
51-L investigation at time 66.174 seconds, Bright spot on right hand solid rocket booster (SRB) in plume in-Z direction start of bright spots on +Z side, MIGOR/USC-10. Kennedy Space Center alternative photo number is 108-KSC-386C-648/319.
Impulsive Control for Continuous-Time Markov Decision Processes: A Linear Programming Approach
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dufour, F., E-mail: dufour@math.u-bordeaux1.fr; Piunovskiy, A. B., E-mail: piunov@liv.ac.uk
2016-08-15
In this paper, we investigate an optimization problem for continuous-time Markov decision processes with both impulsive and continuous controls. We consider the so-called constrained problem where the objective of the controller is to minimize a total expected discounted optimality criterion associated with a cost rate function while keeping other performance criteria of the same form, but associated with different cost rate functions, below some given bounds. Our model allows multiple impulses at the same time moment. The main objective of this work is to study the associated linear program defined on a space of measures including the occupation measures ofmore » the controlled process and to provide sufficient conditions to ensure the existence of an optimal control.« less
Field patterns without blow up
NASA Astrophysics Data System (ADS)
Mattei, Ornella; Milton, Graeme W.
2017-09-01
Field patterns, first proposed by the authors in Milton and Mattei (2017 Proc. R. Soc. A 473 20160819), are a new type of wave propagating along orderly patterns of characteristic lines which arise in specific space-time microstructures whose geometry in one spatial dimension plus time is somehow commensurate with the slope of the characteristic lines. In particular, in Milton and Mattei (2017 Proc. R. Soc. A 473 20160819) the authors propose two examples of space-time geometries in which field patterns occur: they are two-phase microstructures in which rectangular space-time inclusions of one material are embedded in another material. After a sufficiently long interval of time, field patterns have local periodicity both in time and space. This allows one to focus only on solving the problem on the discrete network on which a field pattern lives and to define a suitable transfer matrix that, given the solution at a certain time, provides the solution after one time period. For the aforementioned microstructures, many of the eigenvalues of this { P }{ T }-symmetric transfer matrix have unit norm and hence the corresponding eigenvectors correspond to propagating modes. However, there are also modes that blow up exponentially with time coupled with modes that decrease exponentially with time. The question arises as to whether there are space-time microstructures such that the transfer matrix only has eigenvalues on the unit circle, so that there are no growing modes (modes that blow-up)? The answer is found here, where we see that certain space-time checkerboards have the property that all the modes are propagating modes, within a certain range of the material parameters. Interestingly, when there is no blow-up, the waves generated by an instantaneous disturbance at a point look like shocks with a wake of oscillatory waves, whose amplitude, very remarkably, does not tend to zero away from the wave front.
Bayesian state space models for dynamic genetic network construction across multiple tissues.
Liang, Yulan; Kelemen, Arpad
2016-08-01
Construction of gene-gene interaction networks and potential pathways is a challenging and important problem in genomic research for complex diseases while estimating the dynamic changes of the temporal correlations and non-stationarity are the keys in this process. In this paper, we develop dynamic state space models with hierarchical Bayesian settings to tackle this challenge for inferring the dynamic profiles and genetic networks associated with disease treatments. We treat both the stochastic transition matrix and the observation matrix time-variant and include temporal correlation structures in the covariance matrix estimations in the multivariate Bayesian state space models. The unevenly spaced short time courses with unseen time points are treated as hidden state variables. Hierarchical Bayesian approaches with various prior and hyper-prior models with Monte Carlo Markov Chain and Gibbs sampling algorithms are used to estimate the model parameters and the hidden state variables. We apply the proposed Hierarchical Bayesian state space models to multiple tissues (liver, skeletal muscle, and kidney) Affymetrix time course data sets following corticosteroid (CS) drug administration. Both simulation and real data analysis results show that the genomic changes over time and gene-gene interaction in response to CS treatment can be well captured by the proposed models. The proposed dynamic Hierarchical Bayesian state space modeling approaches could be expanded and applied to other large scale genomic data, such as next generation sequence (NGS) combined with real time and time varying electronic health record (EHR) for more comprehensive and robust systematic and network based analysis in order to transform big biomedical data into predictions and diagnostics for precision medicine and personalized healthcare with better decision making and patient outcomes.
A Plasma Drag Hypervelocity Particle Accelerator (HYPER)
NASA Technical Reports Server (NTRS)
Best, Steve R.; Rose, M. Frank
1998-01-01
Current debris models are able to predict the growth of the space debris problem and suggest that spacecraft must employ armor or bumper shields for some orbital altitudes now and that the problem will become worse as a function of time. The practical upper limit to the velocity distribution is on the order of 40 km/s and is associated with the natural environment. The velocity distribution of the man-made component peaks at 9-10 km/s with maximum velocity in the 14-16 km/s range. Experience in space has verified that the "high probability of impact" particles are in the microgram to milligram range. These particles can have very significant effects on coatings, insulators, and thin metallic layers. The surface of thick materials becomes pitted and the local debris component is enhanced by ejecta from the impact events. In this paper, the HYPER facility is described which produces a reasonable simulation of the man-made space debris spectrum in a controlled environment. The facility capability is discussed in terms of drive geometry, energetics, velocity distribution, diagnostics, and projectile/debris loading. The facility has been used to study impact phenomena on Space Station Freedom's solar array structure, the calibration of space debris collectors, other solar array materials, potential structural materials for use in space, electrical breakdown in the space environment, and as a means of clarifying or duplicating the impact phenomena on surfaces which have been exposed in space.
Grand Challenge Problems in Real-Time Mission Control Systems for NASA's 21st Century Missions
NASA Technical Reports Server (NTRS)
Pfarr, Barbara B.; Donohue, John T.; Hughes, Peter M.
1999-01-01
Space missions of the 21st Century will be characterized by constellations of distributed spacecraft, miniaturized sensors and satellites, increased levels of automation, intelligent onboard processing, and mission autonomy. Programmatically, these missions will be noted for dramatically decreased budgets and mission development lifecycles. Current progress towards flexible, scaleable, low-cost, reusable mission control systems must accelerate given the current mission deployment schedule, and new technology will need to be infused to achieve desired levels of autonomy and processing capability. This paper will discuss current and future missions being managed at NASA's Goddard Space Flight Center in Greenbelt, MD. It will describe the current state of mission control systems and the problems they need to overcome to support the missions of the 21st Century.
Graph theory approach to the eigenvalue problem of large space structures
NASA Technical Reports Server (NTRS)
Reddy, A. S. S. R.; Bainum, P. M.
1981-01-01
Graph theory is used to obtain numerical solutions to eigenvalue problems of large space structures (LSS) characterized by a state vector of large dimensions. The LSS are considered as large, flexible systems requiring both orientation and surface shape control. Graphic interpretation of the determinant of a matrix is employed to reduce a higher dimensional matrix into combinations of smaller dimensional sub-matrices. The reduction is implemented by means of a Boolean equivalent of the original matrices formulated to obtain smaller dimensional equivalents of the original numerical matrix. Computation time becomes less and more accurate solutions are possible. An example is provided in the form of a free-free square plate. Linearized system equations and numerical values of a stiffness matrix are presented, featuring a state vector with 16 components.
NASA Technical Reports Server (NTRS)
Wang, Lui; Valenzuela-Rendon, Manuel
1993-01-01
The Space Station Freedom will require the supply of items in a regular fashion. A schedule for the delivery of these items is not easy to design due to the large span of time involved and the possibility of cancellations and changes in shuttle flights. This paper presents the basic concepts of a genetic algorithm model, and also presents the results of an effort to apply genetic algorithms to the design of propellant resupply schedules. As part of this effort, a simple simulator and an encoding by which a genetic algorithm can find near optimal schedules have been developed. Additionally, this paper proposes ways in which robust schedules, i.e., schedules that can tolerate small changes, can be found using genetic algorithms.
Controlling Real-Time Processes On The Space Station With Expert Systems
NASA Astrophysics Data System (ADS)
Leinweber, David; Perry, John
1987-02-01
Many aspects of space station operations involve continuous control of real-time processes. These processes include electrical power system monitoring, propulsion system health and maintenance, environmental and life support systems, space suit checkout, on-board manufacturing, and servicing of attached vehicles such as satellites, shuttles, orbital maneuvering vehicles, orbital transfer vehicles and remote teleoperators. Traditionally, monitoring of these critical real-time processes has been done by trained human experts monitoring telemetry data. However, the long duration of space station missions and the high cost of crew time in space creates a powerful economic incentive for the development of highly autonomous knowledge-based expert control procedures for these space stations. In addition to controlling the normal operations of these processes, the expert systems must also be able to quickly respond to anomalous events, determine their cause and initiate corrective actions in a safe and timely manner. This must be accomplished without excessive diversion of system resources from ongoing control activities and any events beyond the scope of the expert control and diagnosis functions must be recognized and brought to the attention of human operators. Real-time sensor based expert systems (as opposed to off-line, consulting or planning systems receiving data via the keyboard) pose particular problems associated with sensor failures, sensor degradation and data consistency, which must be explicitly handled in an efficient manner. A set of these systems must also be able to work together in a cooperative manner. This paper describes the requirements for real-time expert systems in space station control, and presents prototype implementations of space station expert control procedures in PICON (process intelligent control). PICON is a real-time expert system shell which operates in parallel with distributed data acquisition systems. It incorporates a specialized inference engine with a specialized scheduling portion specifically designed to match the allocation of system resources with the operational requirements of real-time control systems. Innovative knowledge engineering techniques used in PICON to facilitate the development of real-time sensor-based expert systems which use the special features of the inference engine are illustrated in the prototype examples.
A geometric viewpoint on generalized hydrodynamics
NASA Astrophysics Data System (ADS)
Doyon, Benjamin; Spohn, Herbert; Yoshimura, Takato
2018-01-01
Generalized hydrodynamics (GHD) is a large-scale theory for the dynamics of many-body integrable systems. It consists of an infinite set of conservation laws for quasi-particles traveling with effective ("dressed") velocities that depend on the local state. We show that these equations can be recast into a geometric dynamical problem. They are conservation equations with state-independent quasi-particle velocities, in a space equipped with a family of metrics, parametrized by the quasi-particles' type and speed, that depend on the local state. In the classical hard rod or soliton gas picture, these metrics measure the free length of space as perceived by quasi-particles; in the quantum picture, they weigh space with the density of states available to them. Using this geometric construction, we find a general solution to the initial value problem of GHD, in terms of a set of integral equations where time appears explicitly. These integral equations are solvable by iteration and provide an extremely efficient solution algorithm for GHD.
Effects of line-of-sight velocity on spaced-antenna measurements, part 3.5A
NASA Technical Reports Server (NTRS)
Royrvik, O.
1984-01-01
Horizontal wind velocities in the upper atmosphere, particularly the mesosphere, have been measured using a multitude of different techniques. Most techniques are based on stated or unstated assumptions about the wind field that may or may not be true. Some problems with the spaced antenna drifts (SAD) technique that usually appear to be overlooked are investigated. These problems are not unique to the SAD technique; very similar considerations apply to measurement of horizontal wind using multiple-beam Doppler radars as well. Simply stated, the SAD technique relies on scattering from multiple scatterers within an antenna beam of fairly large beam width. The combination of signals with random phase gives rise to an interference pattern on the ground. This pattern will drift across the ground with a velocity twice that of the ionospheric irregularities from which the radar signals are scattered. By using spaced receivers and measuring time delays of the signal fading in different antennas, it is possible to estimate the horizontal drift velocities.
NASA Astrophysics Data System (ADS)
Mitra, Debasis; Boutchko, Rostyslav; Ray, Judhajeet; Nilsen-Hamilton, Marit
2015-03-01
In this work we present a time-lapsed confocal microscopy image analysis technique for an automated gene expression study of multiple single living cells. Fluorescence Resonance Energy Transfer (FRET) is a technology by which molecule-to-molecule interactions are visualized. We analyzed a dynamic series of ~102 images obtained using confocal microscopy of fluorescence in yeast cells containing RNA reporters that give a FRET signal when the gene promoter is activated. For each time frame, separate images are available for three spectral channels and the integrated intensity snapshot of the system. A large number of time-lapsed frames must be analyzed to identify each cell individually across time and space, as it is moving in and out of the focal plane of the microscope. This makes it a difficult image processing problem. We have proposed an algorithm here, based on scale-space technique, which solves the problem satisfactorily. The algorithm has multiple directions for even further improvement. The ability to rapidly measure changes in gene expression simultaneously in many cells in a population will open the opportunity for real-time studies of the heterogeneity of genetic response in a living cell population and the interactions between cells that occur in a mixed population, such as the ones found in the organs and tissues of multicellular organisms.