Science.gov

Sample records for updating recursive xml

  1. XML in Libraries.

    ERIC Educational Resources Information Center

    Tennant, Roy, Ed.

    This book presents examples of how libraries are using XML (eXtensible Markup Language) to solve problems, expand services, and improve systems. Part I contains papers on using XML in library catalog records: "Updating MARC Records with XMLMARC" (Kevin S. Clarke, Stanford University) and "Searching and Retrieving XML Records via the Web" (Theo van…

  2. XML in Libraries.

    ERIC Educational Resources Information Center

    Tennant, Roy, Ed.

    This book presents examples of how libraries are using XML (eXtensible Markup Language) to solve problems, expand services, and improve systems. Part I contains papers on using XML in library catalog records: "Updating MARC Records with XMLMARC" (Kevin S. Clarke, Stanford University) and "Searching and Retrieving XML Records via the Web" (Theo van

  3. XML Files

    MedlinePlus

    ... nlm.nih.gov/medlineplus/xml.html MedlinePlus XML Files To use the sharing features on this page, ... information on all English and Spanish topic groups. Files generated on May 31, 2016 MedlinePlus Health Topic ...

  4. SU-E-T-327: The Update of a XML Composing Tool for TrueBeam Developer Mode

    SciTech Connect

    Yan, Y; Mao, W; Jiang, S

    2014-06-01

    Purpose: To introduce a major upgrade of a novel XML beam composing tool to scientists and engineers who strive to translate certain capabilities of TrueBeam Developer Mode to future clinical benefits of radiation therapy. Methods: TrueBeam Developer Mode provides the users with a test bed for unconventional plans utilizing certain unique features not accessible at the clinical mode. To access the full set of capabilities, a XML beam definition file accommodating all parameters including kV/MV imaging triggers in the plan can be locally loaded at this mode, however it is difficult and laborious to compose one in a text editor. In this study, a stand-along interactive XML beam composing application, TrueBeam TeachMod, was developed on Windows platforms to assist users in making their unique plans in a WYSWYG manner. A conventional plan can be imported in a DICOM RT object as the start of the beam editing process in which trajectories of all axes of a TrueBeam machine can be modified to the intended values at any control point. TeachMod also includes libraries of predefined imaging and treatment procedures to further expedite the process. Results: The TeachMod application is a major of the TeachMod module within DICOManTX. It fully supports TrueBeam 2.0. Trajectories of all axes including all MLC leaves can be graphically rendered and edited as needed. The time for XML beam composing has been reduced to a negligible amount regardless the complexity of the plan. A good understanding of XML language and TrueBeam schema is not required though preferred. Conclusion: Creating XML beams manually in a text editor will be a lengthy error-prone process for sophisticated plans. A XML beam composing tool is highly desirable for R and D activities. It will bridge the gap between scopes of TrueBeam capabilities and their clinical application potentials.

  5. Recursive SAR imaging

    NASA Astrophysics Data System (ADS)

    Moses, Randolph L.; Ash, Joshua N.

    2008-04-01

    We investigate a recursive procedure for synthetic aperture imaging. We consider a concept in which a SAR system persistently interrogates a scene, for example as it flies along or around that scene. In traditional SAR imaging, the radar measurements are processed in blocks, by partitioning the data into a set of non-overlapping or overlapping azimuth angles, then processing each block. We consider a recursive update approach, in which the SAR image is continually updated, as a linear combination of a small number of previous images and a term containing the current radar measurement. We investigate the crossrange sidelobes realized by such an imaging approach. We show that a first-order autoregression of the image gives crossrange sidelobes similar to a rectangular azimuth window, while a third-order autoregression gives sidelobes comparable to those obtained from widely-used windows in block-processing image formation. The computational and memory requirements of the recursive imaging approach are modest - on the order of M N2 where M is the recursion order (typically <= 3) and N2 is the image size. We compare images obtained from the recursive and block processing techniques, both for a synthetic scene and for X-band SAR measurements from the Gotcha data set.

  6. Probing recursion.

    PubMed

    Lobina, David J

    2014-11-01

    The experimental probing of recursion in human performance is faced with non-trivial problems. Here, I analyse three case studies from the literature and argue that they tell us little about the underlying mental processes at play within each of these domains: (a) the question of whether experimental participants employ recursive rules in parsing artificial strings of nonsense syllables; (b) the role of self-embedded structures in reasoning and general cognition; and (c) the reputed connection between structural features of a given object and the corresponding, recursive rules needed to represent/generate it. I then outline what a recursive process would actually look like and how one could go about probing its presence in human behaviour, concluding, however, that recursive processes in performance are very unlikely, at least as far as fast, mandatory, and automatic modular processes are concerned. PMID:24817314

  7. XML under the Hood.

    ERIC Educational Resources Information Center

    Scharf, David

    2002-01-01

    Discusses XML (extensible markup language), particularly as it relates to libraries. Topics include organizing information; cataloging; metadata; similarities to HTML; organizations dealing with XML; making XML useful; a history of XML; the semantic Web; related technologies; XML at the Library of Congress; and its role in improving the

  8. XML under the Hood.

    ERIC Educational Resources Information Center

    Scharf, David

    2002-01-01

    Discusses XML (extensible markup language), particularly as it relates to libraries. Topics include organizing information; cataloging; metadata; similarities to HTML; organizations dealing with XML; making XML useful; a history of XML; the semantic Web; related technologies; XML at the Library of Congress; and its role in improving the…

  9. Recursion Mathematics.

    ERIC Educational Resources Information Center

    Olson, Alton T.

    1989-01-01

    Discusses the use of the recursive method to permutations of n objects and a problem making c cents in change using pennies and nickels when order is important. Presents a LOGO program for the examples. (YP)

  10. On recursion.

    PubMed

    Watumull, Jeffrey; Hauser, Marc D; Roberts, Ian G; Hornstein, Norbert

    2014-01-01

    It is a truism that conceptual understanding of a hypothesis is required for its empirical investigation. However, the concept of recursion as articulated in the context of linguistic analysis has been perennially confused. Nowhere has this been more evident than in attempts to critique and extend Hauseretal's. (2002) articulation. These authors put forward the hypothesis that what is uniquely human and unique to the faculty of language-the faculty of language in the narrow sense (FLN)-is a recursive system that generates and maps syntactic objects to conceptual-intentional and sensory-motor systems. This thesis was based on the standard mathematical definition of recursion as understood by Gödel and Turing, and yet has commonly been interpreted in other ways, most notably and incorrectly as a thesis about the capacity for syntactic embedding. As we explain, the recursiveness of a function is defined independent of such output, whether infinite or finite, embedded or unembedded-existent or non-existent. And to the extent that embedding is a sufficient, though not necessary, diagnostic of recursion, it has not been established that the apparent restriction on embedding in some languages is of any theoretical import. Misunderstanding of these facts has generated research that is often irrelevant to the FLN thesis as well as to other theories of language competence that focus on its generative power of expression. This essay is an attempt to bring conceptual clarity to such discussions as well as to future empirical investigations by explaining three criterial properties of recursion: computability (i.e., rules in intension rather than lists in extension); definition by induction (i.e., rules strongly generative of structure); and mathematical induction (i.e., rules for the principled-and potentially unbounded-expansion of strongly generated structure). By these necessary and sufficient criteria, the grammars of all natural languages are recursive. PMID:24409164

  11. On recursion

    PubMed Central

    Watumull, Jeffrey; Hauser, Marc D.; Roberts, Ian G.; Hornstein, Norbert

    2014-01-01

    It is a truism that conceptual understanding of a hypothesis is required for its empirical investigation. However, the concept of recursion as articulated in the context of linguistic analysis has been perennially confused. Nowhere has this been more evident than in attempts to critique and extend Hauseretal's. (2002) articulation. These authors put forward the hypothesis that what is uniquely human and unique to the faculty of language—the faculty of language in the narrow sense (FLN)—is a recursive system that generates and maps syntactic objects to conceptual-intentional and sensory-motor systems. This thesis was based on the standard mathematical definition of recursion as understood by Gödel and Turing, and yet has commonly been interpreted in other ways, most notably and incorrectly as a thesis about the capacity for syntactic embedding. As we explain, the recursiveness of a function is defined independent of such output, whether infinite or finite, embedded or unembedded—existent or non-existent. And to the extent that embedding is a sufficient, though not necessary, diagnostic of recursion, it has not been established that the apparent restriction on embedding in some languages is of any theoretical import. Misunderstanding of these facts has generated research that is often irrelevant to the FLN thesis as well as to other theories of language competence that focus on its generative power of expression. This essay is an attempt to bring conceptual clarity to such discussions as well as to future empirical investigations by explaining three criterial properties of recursion: computability (i.e., rules in intension rather than lists in extension); definition by induction (i.e., rules strongly generative of structure); and mathematical induction (i.e., rules for the principled—and potentially unbounded—expansion of strongly generated structure). By these necessary and sufficient criteria, the grammars of all natural languages are recursive. PMID:24409164

  12. ScotlandsPlaces XML: Bespoke XML or XML Mapping?

    ERIC Educational Resources Information Center

    Beamer, Ashley; Gillick, Mark

    2010-01-01

    Purpose: The purpose of this paper is to investigate web services (in the form of parameterised URLs), specifically in the context of the ScotlandsPlaces project. This involves cross-domain querying, data retrieval and display via the development of a bespoke XML standard rather than existing XML formats and mapping between them.…

  13. Semantically Interoperable XML Data

    PubMed Central

    Vergara-Niedermayr, Cristobal; Wang, Fusheng; Pan, Tony; Kurc, Tahsin; Saltz, Joel

    2013-01-01

    XML is ubiquitously used as an information exchange platform for web-based applications in healthcare, life sciences, and many other domains. Proliferating XML data are now managed through latest native XML database technologies. XML data sources conforming to common XML schemas could be shared and integrated with syntactic interoperability. Semantic interoperability can be achieved through semantic annotations of data models using common data elements linked to concepts from ontologies. In this paper, we present a framework and software system to support the development of semantic interoperable XML based data sources that can be shared through a Grid infrastructure. We also present our work on supporting semantic validated XML data through semantic annotations for XML Schema, semantic validation and semantic authoring of XML data. We demonstrate the use of the system for a biomedical database of medical image annotations and markups. PMID:25298789

  14. XML Based Course Websites.

    ERIC Educational Resources Information Center

    Wollowski, Michael

    XML, the extensible markup language, is a quickly evolving technology that presents a viable alternative to courseware products and promises to ease the burden of Web authors, who edit their course pages directly. XML uses tags to label kinds of contents, rather than format information. The use of XML enables faculty to focus on providing…

  15. XML-BASED REPRESENTATION

    SciTech Connect

    R. KELSEY

    2001-02-01

    For focused applications with limited user and use application communities, XML can be the right choice for representation. It is easy to use, maintain, and extend and enjoys wide support in commercial and research sectors. When the knowledge and information to be represented is object-based and use of that knowledge and information is a high priority, then XML-based representation should be considered. This paper discusses some of the issues involved in using XML-based representation and presents an example application that successfully uses an XML-based representation.

  16. XML: A Publisher's Perspective.

    ERIC Educational Resources Information Center

    Andrews, Timothy M.

    1999-01-01

    Explains eXtensible Markup Language (XML) and describes how Dow Jones Interactive is using it to improve the news-gathering and dissemination process through intranets and the World Wide Web. Discusses benefits of using XML, the relationship to HyperText Markup Language (HTML), lack of available software tools and industry support, and future…

  17. Adaptable Iterative and Recursive Kalman Filter Schemes

    NASA Technical Reports Server (NTRS)

    Zanetti, Renato

    2014-01-01

    Nonlinear filters are often very computationally expensive and usually not suitable for real-time applications. Real-time navigation algorithms are typically based on linear estimators, such as the extended Kalman filter (EKF) and, to a much lesser extent, the unscented Kalman filter. The Iterated Kalman filter (IKF) and the Recursive Update Filter (RUF) are two algorithms that reduce the consequences of the linearization assumption of the EKF by performing N updates for each new measurement, where N is the number of recursions, a tuning parameter. This paper introduces an adaptable RUF algorithm to calculate N on the go, a similar technique can be used for the IKF as well.

  18. Recursion, Language, and Starlings

    ERIC Educational Resources Information Center

    Corballis, Michael C.

    2007-01-01

    It has been claimed that recursion is one of the properties that distinguishes human language from any other form of animal communication. Contrary to this claim, a recent study purports to demonstrate center-embedded recursion in starlings. I show that the performance of the birds in this study can be explained by a counting strategy, without any

  19. Recursion, Language, and Starlings

    ERIC Educational Resources Information Center

    Corballis, Michael C.

    2007-01-01

    It has been claimed that recursion is one of the properties that distinguishes human language from any other form of animal communication. Contrary to this claim, a recent study purports to demonstrate center-embedded recursion in starlings. I show that the performance of the birds in this study can be explained by a counting strategy, without any…

  20. Chapter 56: Introduction to XML

    NASA Astrophysics Data System (ADS)

    Greene, G.; Krughoff, K. S.

    This is an introduction to the eXtensible Markup Language (XML). XML is a key technology component in the VO development of data models and standard formats for services, e.g. VOTable data formats for cone and SIA (Simple Image Access) services, VOResource schemas for VO Registries, and Space Time Coordinate (STC) schema. We present here the basic anatomy of XML documents, discuss the advantages and uses of XML and provide simple examples for how to describe your own data types. We also discuss reading and writing XML including the commercially available APIs and tools included in the NVO Summer School library.

  1. Introducing streaming XML (SXML)

    NASA Astrophysics Data System (ADS)

    Rogge, Boris; Van de Walle, Rik; Lemahieu, Ignace L.; Philips, Wilfried R.

    2000-10-01

    In the near future broadband networks will become available to large groups of people. The amount of bandwidth available to these users in the future will be much more than it is now. The availability of bandwidth will give birth to a number of new applications. Application developers will need a framework that enable them to utilize the possibilities of these new networks. In this article we present a document type that will allow the addition of (meta-) information to data streams and the synchronization of a different data streams. It is called SXML (Streaming XML) and is based on the eXtensible Markup Language (XML). The SXML grammar is defined in a document type definition (SXML-DTD). The content of an SXML document can be processed real time or can be retrieved from disk. XML is being used in a complete new manner and in a totally different environment in order to easily describe the structure of the stream. Finally, a preliminary implementation has been developed and is being tested.

  2. Recursive Deadbeat Controller Design

    NASA Technical Reports Server (NTRS)

    Juang, Jer-Nan; Phan, Minh Q.

    1997-01-01

    This paper presents a recursive algorithm for a deadbeat predictive controller design. The method combines together the concepts of system identification and deadbeat controller designs. It starts with the multi-step output prediction equation and derives the control force in terms of past input and output time histories. The formulation thus derived satisfies simultaneously system identification and deadbeat controller design requirements. As soon as the coefficient matrices are identified satisfying the output prediction equation, no further work is required to compute the deadbeat control gain matrices. The method can be implemented recursively just as any typical recursive system identification techniques.

  3. Solving nonlinear recursions

    SciTech Connect

    Rabinovich, S.; Berkolaiko, G.; Havlin, S.

    1996-11-01

    A general method to map a polynomial recursion on a matrix linear one is suggested. The solution of the recursion is represented as a product of a matrix multiplied by the vector of initial values. This matrix is product of {ital transfer} matrices whose elements depend only on the polynomial and not on the initial conditions. The method is valid for systems of polynomial recursions and for polynomial recursions of arbitrary order. The only restriction on these recurrent relations is that the highest-order term can be written in explicit form as a function of the lower-order terms (existence of a normal form). A continuous analog of this method is described as well. {copyright} {ital 1996 American Institute of Physics.}

  4. Querying XML Data with SPARQL

    NASA Astrophysics Data System (ADS)

    Bikakis, Nikos; Gioldasis, Nektarios; Tsinaraki, Chrisa; Christodoulakis, Stavros

    SPARQL is today the standard access language for Semantic Web data. In the recent years XML databases have also acquired industrial importance due to the widespread applicability of XML in the Web. In this paper we present a framework that bridges the heterogeneity gap and creates an interoperable environment where SPARQL queries are used to access XML databases. Our approach assumes that fairly generic mappings between ontology constructs and XML Schema constructs have been automatically derived or manually specified. The mappings are used to automatically translate SPARQL queries to semantically equivalent XQuery queries which are used to access the XML databases. We present the algorithms and the implementation of SPARQL2XQuery framework, which is used for answering SPARQL queries over XML databases.

  5. Distinctive signatures of recursion

    PubMed Central

    Martins, Maurício Dias

    2012-01-01

    Although recursion has been hypothesized to be a necessary capacity for the evolution of language, the multiplicity of definitions being used has undermined the broader interpretation of empirical results. I propose that only a definition focused on representational abilities allows the prediction of specific behavioural traits that enable us to distinguish recursion from non-recursive iteration and from hierarchical embedding: only subjects able to represent recursion, i.e. to represent different hierarchical dependencies (related by parenthood) with the same set of rules, are able to generalize and produce new levels of embedding beyond those specified a priori (in the algorithm or in the input). The ability to use such representations may be advantageous in several domains: action sequencing, problem-solving, spatial navigation, social navigation and for the emergence of conventionalized communication systems. The ability to represent contiguous hierarchical levels with the same rules may lead subjects to expect unknown levels and constituents to behave similarly, and this prior knowledge may bias learning positively. Finally, a new paradigm to test for recursion is presented. Preliminary results suggest that the ability to represent recursion in the spatial domain recruits both visual and verbal resources. Implications regarding language evolution are discussed. PMID:22688640

  6. Distinctive signatures of recursion.

    PubMed

    Martins, Maurício Dias

    2012-07-19

    Although recursion has been hypothesized to be a necessary capacity for the evolution of language, the multiplicity of definitions being used has undermined the broader interpretation of empirical results. I propose that only a definition focused on representational abilities allows the prediction of specific behavioural traits that enable us to distinguish recursion from non-recursive iteration and from hierarchical embedding: only subjects able to represent recursion, i.e. to represent different hierarchical dependencies (related by parenthood) with the same set of rules, are able to generalize and produce new levels of embedding beyond those specified a priori (in the algorithm or in the input). The ability to use such representations may be advantageous in several domains: action sequencing, problem-solving, spatial navigation, social navigation and for the emergence of conventionalized communication systems. The ability to represent contiguous hierarchical levels with the same rules may lead subjects to expect unknown levels and constituents to behave similarly, and this prior knowledge may bias learning positively. Finally, a new paradigm to test for recursion is presented. Preliminary results suggest that the ability to represent recursion in the spatial domain recruits both visual and verbal resources. Implications regarding language evolution are discussed. PMID:22688640

  7. "The Wonder Years" of XML.

    ERIC Educational Resources Information Center

    Gazan, Rich

    2000-01-01

    Surveys the current state of Extensible Markup Language (XML), a metalanguage for creating structured documents that describe their own content, and its implications for information professionals. Predicts that XML will become the common language underlying Web, word processing, and database formats. Also discusses Extensible Stylesheet Language…

  8. How Does XML Help Libraries?

    ERIC Educational Resources Information Center

    Banerjee, Kyle

    2002-01-01

    Discusses XML, how it has transformed the way information is managed and delivered, and its impact on libraries. Topics include how XML differs from other markup languages; the document object model (DOM); style sheets; practical applications for archival materials, interlibrary loans, digital collections, and MARC data; and future possibilities.…

  9. Language and Recursion

    NASA Astrophysics Data System (ADS)

    Lowenthal, Francis

    2010-11-01

    This paper examines whether the recursive structure imbedded in some exercises used in the Non Verbal Communication Device (NVCD) approach is actually the factor that enables this approach to favor language acquisition and reacquisition in the case of children with cerebral lesions. For that a definition of the principle of recursion as it is used by logicians is presented. The two opposing approaches to the problem of language development are explained. For many authors such as Chomsky [1] the faculty of language is innate. This is known as the Standard Theory; the other researchers in this field, e.g. Bates and Elman [2], claim that language is entirely constructed by the young child: they thus speak of Language Acquisition. It is also shown that in both cases, a version of the principle of recursion is relevant for human language. The NVCD approach is defined and the results obtained in the domain of language while using this approach are presented: young subjects using this approach acquire a richer language structure or re-acquire such a structure in the case of cerebral lesions. Finally it is shown that exercises used in this framework imply the manipulation of recursive structures leading to regular grammars. It is thus hypothesized that language development could be favored using recursive structures with the young child. It could also be the case that the NVCD like exercises used with children lead to the elaboration of a regular language, as defined by Chomsky [3], which could be sufficient for language development but would not require full recursion. This double claim could reconcile Chomsky's approach with psychological observations made by adherents of the Language Acquisition approach, if it is confirmed by researches combining the use of NVCDs, psychometric methods and the use of Neural Networks. This paper thus suggests that a research group oriented towards this problematic should be organized.

  10. Publishing Scientific Articles in XML.

    NASA Astrophysics Data System (ADS)

    Shaya, E.; Borne, K.; Thomas, B.; Cheung, C. Y.

    2001-12-01

    Most publication houses are using SGML for electronic mark up of pages intended for hardcopy. Since XML is a major subset of SGML with W3C backing and greater database compatibility, many publication houses are naturally considering switching to or including XML. Now, if authors were also to switch to XML for their manuscripts, it would greatly reduce the work load at the publication houses and reduce the number of errors that are introduced in the translation process. XML is also a logical progression for authors since it is rapidly becoming incorporated into editors such as Word Perfect, Notepad, Emacs, etc. There is an XML standard for equation markup, MathML, and equation editors exist for it. It is easy to put these manuscripts onto the Web; all one needs is to link to a standard cascade style sheet (CSS2). Leveraging our experience with encapsulating scientific data in XML the ADC (Astronomical Data Center) staff are working out details of a scientific XML article format called "AXML" (Article XML Markup Language). We foresee using AXML eventually as an end to end solution for data from experiment/observation through analysis to publication. With fewer transformations needed on article text, equations, and tables, less human intervention will be required and fewer human errors will be introduced, for example, proofing of XML documents by publication houses could someday be unnecessary or (at least) vastly more efficient. In this poster we discuss examine several important aspects of this technology, give the technical details of AXML (including a DTD) and give examples which show the power of AXML.

  11. Recursive heuristic classification

    NASA Technical Reports Server (NTRS)

    Wilkins, David C.

    1994-01-01

    The author will describe a new problem-solving approach called recursive heuristic classification, whereby a subproblem of heuristic classification is itself formulated and solved by heuristic classification. This allows the construction of more knowledge-intensive classification programs in a way that yields a clean organization. Further, standard knowledge acquisition and learning techniques for heuristic classification can be used to create, refine, and maintain the knowledge base associated with the recursively called classification expert system. The method of recursive heuristic classification was used in the Minerva blackboard shell for heuristic classification. Minerva recursively calls itself every problem-solving cycle to solve the important blackboard scheduler task, which involves assigning a desirability rating to alternative problem-solving actions. Knowing these ratings is critical to the use of an expert system as a component of a critiquing or apprenticeship tutoring system. One innovation of this research is a method called dynamic heuristic classification, which allows selection among dynamically generated classification categories instead of requiring them to be prenumerated.

  12. Compression of Probabilistic XML Documents

    NASA Astrophysics Data System (ADS)

    Veldman, Irma; de Keijzer, Ander; van Keulen, Maurice

    Database techniques to store, query and manipulate data that contains uncertainty receives increasing research interest. Such UDBMSs can be classified according to their underlying data model: relational, XML, or RDF. We focus on uncertain XML DBMS with as representative example the Probabilistic XML model (PXML) of [10,9]. The size of a PXML document is obviously a factor in performance. There are PXML-specific techniques to reduce the size, such as a push down mechanism, that produces equivalent but more compact PXML documents. It can only be applied, however, where possibilities are dependent. For normal XML documents there also exist several techniques for compressing a document. Since Probabilistic XML is (a special form of) normal XML, it might benefit from these methods even more. In this paper, we show that existing compression mechanisms can be combined with PXML-specific compression techniques. We also show that best compression rates are obtained with a combination of PXML-specific technique with a rather simple generic DAG-compression technique.

  13. Tracking of Multiple Moving Sources Using Recursive EM Algorithm

    NASA Astrophysics Data System (ADS)

    Chung, Pei-Jung; Böhme, Johann F.; Hero, Alfred O.

    2005-12-01

    We deal with recursive direction-of-arrival (DOA) estimation of multiple moving sources. Based on the recursive EM algorithm, we develop two recursive procedures to estimate the time-varying DOA parameter for narrowband signals. The first procedure requires no prior knowledge about the source movement. The second procedure assumes that the motion of moving sources is described by a linear polynomial model. The proposed recursion updates the polynomial coefficients when a new data arrives. The suggested approaches have two major advantages: simple implementation and easy extension to wideband signals. Numerical experiments show that both procedures provide excellent results in a slowly changing environment. When the DOA parameter changes fast or two source directions cross with each other, the procedure designed for a linear polynomial model has a better performance than the general procedure. Compared to the beamforming technique based on the same parameterization, our approach is computationally favorable and has a wider range of applications.

  14. Recursive Objects--An Object Oriented Presentation of Recursion

    ERIC Educational Resources Information Center

    Sher, David B.

    2004-01-01

    Generally, when recursion is introduced to students the concept is illustrated with a toy (Towers of Hanoi) and some abstract mathematical functions (factorial, power, Fibonacci). These illustrate recursion in the same sense that counting to 10 can be used to illustrate a for loop. These are all good illustrations, but do not represent serious…

  15. XWeB: The XML Warehouse Benchmark

    NASA Astrophysics Data System (ADS)

    Mahboubi, Hadj; Darmont, Jérôme

    With the emergence of XML as a standard for representing business data, new decision support applications are being developed. These XML data warehouses aim at supporting On-Line Analytical Processing (OLAP) operations that manipulate irregular XML data. To ensure feasibility of these new tools, important performance issues must be addressed. Performance is customarily assessed with the help of benchmarks. However, decision support benchmarks do not currently support XML features. In this paper, we introduce the XML Warehouse Benchmark (XWeB), which aims at filling this gap. XWeB derives from the relational decision support benchmark TPC-H. It is mainly composed of a test data warehouse that is based on a unified reference model for XML warehouses and that features XML-specific structures, and its associate XQuery decision support workload. XWeB's usage is illustrated by experiments on several XML database management systems.

  16. Recursive Feature Extraction in Graphs

    Energy Science and Technology Software Center (ESTSC)

    2014-08-14

    ReFeX extracts recursive topological features from graph data. The input is a graph as a csv file and the output is a csv file containing feature values for each node in the graph. The features are based on topological counts in the neighborhoods of each nodes, as well as recursive summaries of neighbors' features.

  17. Setting the Standard: XML on Campus.

    ERIC Educational Resources Information Center

    Rawlins, Mike

    2001-01-01

    Explains what XML (Extensible Markup Language) is; where to find it in a few years (everywhere from Web pages, to database management systems, to common campus applications); issues that will make XML somewhat of an experimental strategy in the near term; and the importance of decision-makers being abreast of XML trends in standards, tools

  18. XML and the Future of Digital Libraries.

    ERIC Educational Resources Information Center

    Sperberg-McQueen, C. M.

    1998-01-01

    XML is a newly released subset of SGML that is intended to extend the benefits of that standard to the World Wide Web. Differences between SGML and XML are outlined, potential benefits are discussed, and answers are provided for some frequently asked questions regarding XML. (AEF)

  19. Setting the Standard: XML on Campus.

    ERIC Educational Resources Information Center

    Rawlins, Mike

    2001-01-01

    Explains what XML (Extensible Markup Language) is; where to find it in a few years (everywhere from Web pages, to database management systems, to common campus applications); issues that will make XML somewhat of an experimental strategy in the near term; and the importance of decision-makers being abreast of XML trends in standards, tools…

  20. Recursivity in Lingua Cosmica

    NASA Astrophysics Data System (ADS)

    Ollongren, Alexander

    2011-02-01

    In a sequence of papers on the topic of message construction for interstellar communication by means of a cosmic language, the present author has discussed various significant requirements such a lingua should satisfy. The author's Lingua Cosmica is a (meta) system for annotating contents of possibly large-scale messages for ETI. LINCOS, based on formal constructive logic, was primarily designed for dealing with logic contents of messages but is also applicable for denoting structural properties of more general abstractions embedded in such messages. The present paper explains ways and means for achieving this for a special case: recursive entities. As usual two stages are involved: first the domain of discourse is enriched with suitable representations of the entities concerned, after which properties over them can be dealt with within the system itself. As a representative example the case of Russian dolls (Matrjoshka's) is discussed in some detail and relations with linguistic structures in natural languages are briefly exploited.

  1. Tomcat, Oracle & XML Web Archive

    Energy Science and Technology Software Center (ESTSC)

    2008-01-01

    The TOX (Tomcat Oracle & XML) web archive is a foundation for development of HTTP-based applications using Tomcat (or some other servlet container) and an Oracle RDBMS. Use of TOX requires coding primarily in PL/SQL, JavaScript, and XSLT, but also in HTML, CSS and potentially Java. Coded in Java and PL/SQL itself, TOX provides the foundation for more complex applications to be built.

  2. XML Translator for Interface Descriptions

    NASA Technical Reports Server (NTRS)

    Boroson, Elizabeth R.

    2009-01-01

    A computer program defines an XML schema for specifying the interface to a generic FPGA from the perspective of software that will interact with the device. This XML interface description is then translated into header files for C, Verilog, and VHDL. User interface definition input is checked via both the provided XML schema and the translator module to ensure consistency and accuracy. Currently, programming used on both sides of an interface is inconsistent. This makes it hard to find and fix errors. By using a common schema, both sides are forced to use the same structure by using the same framework and toolset. This makes for easy identification of problems, which leads to the ability to formulate a solution. The toolset contains constants that allow a programmer to use each register, and to access each field in the register. Once programming is complete, the translator is run as part of the make process, which ensures that whenever an interface is changed, all of the code that uses the header files describing it is recompiled.

  3. Recursive adaptive frame integration limited

    NASA Astrophysics Data System (ADS)

    Rafailov, Michael K.

    2006-05-01

    Recursive Frame Integration Limited was proposed as a way to improve frame integration performance and mitigate issues related to high data rate needed for conventional frame integration. The technique applies two thresholds - one tuned for optimum probability of detection, the other to manage required false alarm rate - and allows a non-linear integration process that, along with Signal-to-Noise Ratio (SNR) gain, provides system designers more capability where cost, weight, or power considerations limit system data rate, processing, or memory capability. However, Recursive Frame Integration Limited may have performance issues when single frame SNR is really low. Recursive Adaptive Frame Integration Limited is proposed as a means to improve limited integration performance with really low single frame SNR. It combines the benefits of nonlinear recursive limited frame integration and adaptive thresholds with a kind of conventional frame integration.

  4. Recursion relations from soft theorems

    NASA Astrophysics Data System (ADS)

    Luo, Hui; Wen, Congkao

    2016-03-01

    We establish a set of new on-shell recursion relations for amplitudes satisfying soft theorems. The recursion relations can apply to those amplitudes whose additional physical inputs from soft theorems are enough to overcome the bad large- z behaviour. This work is a generalization of the recursion relations recently obtained by Cheung et al. for amplitudes in scalar effective field theories with enhanced vanishing soft behaviours, which can be regarded as a special case of those with non-vanishing soft limits. We apply the recursion relations to tree-level amplitudes in various theories, including amplitudes in the Akulov-Volkov theory and amplitudes containing dilatons of spontaneously-broken conformal symmetry.

  5. Hopf algebras and topological recursion

    NASA Astrophysics Data System (ADS)

    Esteves, João N.

    2015-11-01

    We consider a model for topological recursion based on the Hopf algebra of planar binary trees defined by Loday and Ronco (1998 Adv. Math. 139 293-309 We show that extending this Hopf algebra by identifying pairs of nearest neighbor leaves, and thus producing graphs with loops, we obtain the full recursion formula discovered by Eynard and Orantin (2007 Commun. Number Theory Phys. 1 347-452).

  6. XML technology planning database : lessons learned

    NASA Technical Reports Server (NTRS)

    Some, Raphael R.; Neff, Jon M.

    2005-01-01

    A hierarchical Extensible Markup Language(XML) database called XCALIBR (XML Analysis LIBRary) has been developed by Millennium Program to assist in technology investment (ROI) analysis and technology Language Capability the New return on portfolio optimization. The database contains mission requirements and technology capabilities, which are related by use of an XML dictionary. The XML dictionary codifies a standardized taxonomy for space missions, systems, subsystems and technologies. In addition to being used for ROI analysis, the database is being examined for use in project planning, tracking and documentation. During the past year, the database has moved from development into alpha testing. This paper describes the lessons learned during construction and testing of the prototype database and the motivation for moving from an XML taxonomy to a standard XML-based ontology.

  7. Speed up of XML parsers with PHP language implementation

    NASA Astrophysics Data System (ADS)

    Georgiev, Bozhidar; Georgieva, Adriana

    2012-11-01

    In this paper, authors introduce PHP5's XML implementation and show how to read, parse, and write a short and uncomplicated XML file using Simple XML in a PHP environment. The possibilities for mutual work of PHP5 language and XML standard are described. The details of parsing process with Simple XML are also cleared. A practical project PHP-XML-MySQL presents the advantages of XML implementation in PHP modules. This approach allows comparatively simple search of XML hierarchical data by means of PHP software tools. The proposed project includes database, which can be extended with new data and new XML parsing functions.

  8. XML Schema Representation of DICOM Structured Reporting

    PubMed Central

    Lee, K. P.; Hu, Jingkun

    2003-01-01

    Objective: The Digital Imaging and Communications in Medicine (DICOM) Structured Reporting (SR) standard improves the expressiveness, precision, and comparability of documentation about diagnostic images and waveforms. It supports the interchange of clinical reports in which critical features shown by images and waveforms can be denoted unambiguously by the observer, indexed, and retrieved selectively by subsequent reviewers. It is essential to provide access to clinical reports across the health care enterprise by using technologies that facilitate information exchange and processing by computers as well as provide support for robust and semantically rich standards, such as DICOM. This is supported by the current trend in the healthcare industry towards the use of Extensible Markup Language (XML) technologies for storage and exchange of medical information. The objective of the work reported here is to develop XML Schema for representing DICOM SR as XML documents. Design: We briefly describe the document type definition (DTD) for XML and its limitations, followed by XML Schema (the intended replacement for DTD) and its features. A framework for generating XML Schema for representing DICOM SR in XML is presented next. Measurements: None applicable. Results: A schema instance based on an SR example in the DICOM specification was created and validated against the schema. The schema is being used extensively in producing reports on Philips Medical Systems ultrasound equipment. Conclusion: With the framework described it is feasible to generate XML Schema using the existing DICOM SR specification. It can also be applied to generate XML Schemas for other DICOM information objects. PMID:12595410

  9. Recursive calculation of Hansen coefficients

    NASA Astrophysics Data System (ADS)

    Branham, Richard L., Jr.

    1990-06-01

    Hansen coefficients are used in expansions of the elliptic motion. Three methods for calculating the coefficients are studied: Tisserand's method, the Von Zeipel-Andoyer (VZA) method with explicit representation of the polynomials required to compute the Hansen coefficients, and the VZA method with the values of the polynomials calculated recursively. The VZA method with explicit polynomials is by far the most rapid, but the tabulation of the polynomials only extends to 12th order in powers of the eccentricity, and unless one has access to the polynomials in machine-readable form their entry is laborious and error-prone. The recursive calculation of the VZA polynomials, needed to compute the Hansen coefficients, while slower, is faster than the calculation of the Hansen coefficients by Tisserand's method, up to 10th order in the eccentricity and is still relatively efficient for higher orders. The main advantages of the recursive calculation are the simplicity of the program and one's being able to extend the expansions to any order of eccentricity with ease. Because FORTRAN does not implement recursive procedures, this paper used C for all of the calculations. The most important conclusion is recursion's genuine usefulness in scientific computing.

  10. Recursive implementations of temporal filters for image motion computation.

    PubMed

    Clifford, C W; Langley, K

    2000-05-01

    Efficient algorithms for image motion computation are important for computer vision applications and the modelling of biological vision systems. Intensity-based image motion computation proceeds in two stages: the convolution of linear spatiotemporal filter kernels with the image sequence, followed by the non-linear combination of the filter outputs. If the spatiotemporal extent of the filter kernels is large, then the convolution stage can be very intensive computationally. One effective means of reducing the storage required and computation involved in implementing the temporal convolutions is the introduction of recursive filtering. Non-recursive methods require the number of frames of the image sequence stored at any given time to be equal to the temporal extent of the slowest temporal filter. In contrast, recursive methods encode recent stimulus history implicitly in the values of a small number of variables updated through a series of feedback equations. Recursive filtering reduces the number of values stored in memory during convolution and the number of mathematical operations involved in computing the filters' outputs. This paper extends previous recursive implementations of gradient- and correlation-based motion analysis algorithms [Fleet DJ, Langley K (1995) IEEE PAMI 17: 61-67; Clifford CWG, Ibbotson MR, Langley K (1997) Vis Neurosci 14: 741-749], describing a recursive implementation of causal band-pass temporal filters suitable for use in energy- and phase-based algorithms for image motion computation. It is shown that the filters' temporal frequency tuning curves fit psychophysical estimates of the temporal properties of human visual filters. PMID:10836584

  11. Compressing Aviation Data in XML Format

    NASA Technical Reports Server (NTRS)

    Patel, Hemil; Lau, Derek; Kulkarni, Deepak

    2003-01-01

    Design, operations and maintenance activities in aviation involve analysis of variety of aviation data. This data is typically in disparate formats making it difficult to use with different software packages. Use of a self-describing and extensible standard called XML provides a solution to this interoperability problem. XML provides a standardized language for describing the contents of an information stream, performing the same kind of definitional role for Web content as a database schema performs for relational databases. XML data can be easily customized for display using Extensible Style Sheets (XSL). While self-describing nature of XML makes it easy to reuse, it also increases the size of data significantly. Therefore, transfemng a dataset in XML form can decrease throughput and increase data transfer time significantly. It also increases storage requirements significantly. A natural solution to the problem is to compress the data using suitable algorithm and transfer it in the compressed form. We found that XML-specific compressors such as Xmill and XMLPPM generally outperform traditional compressors. However, optimal use of Xmill requires of discovery of optimal options to use while running Xmill. This, in turn, depends on the nature of data used. Manual disc0ver.y of optimal setting can require an engineer to experiment for weeks. We have devised an XML compression advisory tool that can analyze sample data files and recommend what compression tool would work the best for this data and what are the optimal settings to be used with a XML compression tool.

  12. Get It Together: Integrating Data with XML.

    ERIC Educational Resources Information Center

    Miller, Ron

    2003-01-01

    Discusses the use of XML for data integration to move data across different platforms, including across the Internet, from a variety of sources. Topics include flexibility; standards; organizing databases; unstructured data and the use of meta tags to encode it with XML information; cost effectiveness; and eliminating client software licenses.…

  13. Recursive least-squares learning algorithms for neural networks

    SciTech Connect

    Lewis, P.S. ); Hwang, Jenq-Neng . Dept. of Electrical Engineering)

    1990-01-01

    This paper presents the development of a pair of recursive least squares (RLS) algorithms for online training of multilayer perceptrons, which are a class of feedforward artificial neural networks. These algorithms incorporate second order information about the training error surface in order to achieve faster learning rates than are possible using first order gradient descent algorithms such as the generalized delta rule. A least squares formulation is derived from a linearization of the training error function. Individual training pattern errors are linearized about the network parameters that were in effect when the pattern was presented. This permits the recursive solution of the least squares approximation, either via conventional RLS recursions or by recursive QR decomposition-based techniques. The computational complexity of the update is in the order of (N{sup 2}), where N is the number of network parameters. This is due to the estimation of the N {times} N inverse Hessian matrix. Less computationally intensive approximations of the RLS algorithms can be easily derived by using only block diagonal elements of this matrix, thereby partitioning the learning into independent sets. A simulation example is presented in which a neural network is trained to approximate a two dimensional Gaussian bump. In this example, RLS training required an order of magnitude fewer iterations on average (527) than did training with the generalized delta rule (6331). 14 refs., 3 figs.

  14. The counting recursive digital filter.

    NASA Technical Reports Server (NTRS)

    Zohar, S.

    1973-01-01

    Analysis of the bit-level operations involved in the convolutions realizing recursive digital filters leads to hardware designs of such filters based on the operation of counting. Various designs realizing both the canonic and 'direct' forms are presented with particular emphasis on low-cost low-speed high-flexibility machines.

  15. How Learning Logic Programming Affects Recursion Comprehension

    ERIC Educational Resources Information Center

    Haberman, Bruria

    2004-01-01

    Recursion is a central concept in computer science, yet it is difficult for beginners to comprehend. Israeli high-school students learn recursion in the framework of a special modular program in computer science (Gal-Ezer & Harel, 1999). Some of them are introduced to the concept of recursion in two different paradigms: the procedural programming…

  16. An Introduction to the Extensible Markup Language (XML).

    ERIC Educational Resources Information Center

    Bryan, Martin

    1998-01-01

    Describes Extensible Markup Language (XML), a subset of the Standard Generalized Markup Language (SGML) that is designed to make it easy to interchange structured documents over the Internet. Topics include Document Type Definition (DTD), components of XML, the use of XML, text and non-text elements, and uses for XML-coded files. (LRW)

  17. XML, TEI, and Digital Libraries in the Humanities.

    ERIC Educational Resources Information Center

    Nellhaus, Tobin

    2001-01-01

    Describes the history and major features of XML and TEI, discusses their potential utility for the creation of digital libraries, and focuses on XML's application in the humanities, particularly theater and drama studies. Highlights include HTML and hyperlinks; the impact of XML on text encoding and document access; and XML and academic…

  18. XML Format for SESAME and LEOS

    SciTech Connect

    Durrenberger, J K; Neely, J R; Sterne, P A

    2009-04-29

    The objective of this document is to describe the XML format used by LLNL and LANL to represent the equation-of-state and related material information in the LEOS and SESAME data libraries. The primary purpose of this document is to describe a specific XML format for representing EOS data that is tailored to the nature of the underlying data and is amenable to conversion to both legacy SESAME and LEOS binary formats. The secondary purpose is to describe an XML format that lends itself to a 'natural' representation in a binary file format of the SESAME, pdb or hdf5 form so that this format and related tools can be used for the rapid and efficient development and implementation of prototype data structures. This document describes the XML format only. A working knowledge of LEOS and SESAME formats is assumed.

  19. MedlinePlus XML Data Sources

    MedlinePlus

    ... format. The Web service provides access to all English and Spanish MedlinePlus health topics and offers a ... Health Topic Group XML These files cover all English and Spanish health topics and can be found ...

  20. Core Recursive Hierarchical Image Segmentation

    NASA Technical Reports Server (NTRS)

    Tilton, James

    2011-01-01

    The Recursive Hierarchical Image Segmentation (RHSEG) software has been repackaged to provide a version of the RHSEG software that is not subject to patent restrictions and that can be released to the general public through NASA GSFC's Open Source release process. Like the Core HSEG Software Package, this Core RHSEG Software Package also includes a visualization program called HSEGViewer along with a utility program HSEGReader. It also includes an additional utility program called HSEGExtract. The unique feature of the Core RHSEG package is that it is a repackaging of the RHSEG technology designed to specifically avoid the inclusion of the certain software technology. Unlike the Core HSEG package, it includes the recursive portions of the technology, but does not include processing window artifact elimination technology.

  1. XML data compression in web publishing

    NASA Astrophysics Data System (ADS)

    Qiu, Ruiheng; Hu, Wei; Tang, Zhi; Lu, Xiaoqing; Zhang, Lei

    2012-03-01

    XML is widely used in various document formats on the web. But it has caused negative impacts such as expensive document distribution time over the web, and long content jumping and rendering delay, especially on mobile devices. Hence we proposed a Schema-based efficient queryable XML compressor, called XTrim, which significantly improves compression ratio by utilizing optimized information in XML Schema while supporting efficient queries. Firstly, XTrim draws structure information from XML document and corresponding XML Schema. Then a novel technique is used to transform the XML tree-like structure into a compact indexed form to support efficient queries. At the same time, text values are obtained, and a language-based text trim method (LTT) that facilitates language-specific text compressors is adopted to reduce the size of text values in various languages. In LTT a word composition detection method is proposed to better process text in non-Latin languages. To evaluate the performance of XTrim, we have implemented a compressor and query engine prototype. Via extensive experiments, results show that XTrim outperforms XMill and existing queryable alternatives in terms of compression ratio, as well as the query efficiency. By applying XTrim to documents, the storage space can save up to 30% and the content jumping and rendering delay is reduced to less than 100ms from 4 seconds.

  2. XML Flight/Ground Data Dictionary Management

    NASA Technical Reports Server (NTRS)

    Wright, Jesse; Wiklow, Colette

    2007-01-01

    A computer program generates Extensible Markup Language (XML) files that effect coupling between the command- and telemetry-handling software running aboard a spacecraft and the corresponding software running in ground support systems. The XML files are produced by use of information from the flight software and from flight-system engineering. The XML files are converted to legacy ground-system data formats for command and telemetry, transformed into Web-based and printed documentation, and used in developing new ground-system data-handling software. Previously, the information about telemetry and command was scattered in various paper documents that were not synchronized. The process of searching and reading the documents was time-consuming and introduced errors. In contrast, the XML files contain all of the information in one place. XML structures can evolve in such a manner as to enable the addition, to the XML files, of the metadata necessary to track the changes and the associated documentation. The use of this software has reduced the extent of manual operations in developing a ground data system, thereby saving considerable time and removing errors that previously arose in the translation and transcription of software information from the flight to the ground system.

  3. XML Reconstruction View Selection in XML Databases: Complexity Analysis and Approximation Scheme

    NASA Astrophysics Data System (ADS)

    Chebotko, Artem; Fu, Bin

    Query evaluation in an XML database requires reconstructing XML subtrees rooted at nodes found by an XML query. Since XML subtree reconstruction can be expensive, one approach to improve query response time is to use reconstruction views - materialized XML subtrees of an XML document, whose nodes are frequently accessed by XML queries. For this approach to be efficient, the principal requirement is a framework for view selection. In this work, we are the first to formalize and study the problem of XML reconstruction view selection. The input is a tree T, in which every node i has a size c i and profit p i , and the size limitation C. The target is to find a subset of subtrees rooted at nodes i 1, ⋯ , i k respectively such that c_{i_1}+\\cdots +c_{i_k}le C, and p_{i_1}+\\cdots +p_{i_k} is maximal. Furthermore, there is no overlap between any two subtrees selected in the solution. We prove that this problem is NP-hard and present a fully polynomial-time approximation scheme (FPTAS) as a solution.

  4. How good is recursive bisection?

    SciTech Connect

    Simon, H.D.; Teng, S.H.

    1997-09-01

    The most commonly used p-way partitioning method is recursive bisection (RB). It first divides a graph or a mesh into two equal-sized pieces, by a good bisection algorithm, and then recursively divides the two pieces. Ideally, the authors would like to use an optimal bisection algorithm. Because the optimal bisection problem that partitions a graph into two equal-sized subgraphs to minimize the number of edges cut is NP-complete, practical Rb algorithms use more efficient heuristics in place of an optimal bisection algorithm. Most such heuristics are designed to find the best possible bisection within allowed time. They show that the RB method, even when an optimal bisection algorithm is assumed, may produce a p-way partition that is very far away from the optimal one. The negative result is complemented by two positive ones: first they show that for some important classes of graphs that occur in practical applications, such as well-shaped finite-element and finite-difference meshes, RB is within a constant factor of the optimal one almost always. Second, they show that if the balance condition is relaxed so that each block in the p-way partition is bounded by 2n/p, where n is the number of vertices of the graph, then a modified RB finds an approximately balanced p-way partition whose cost is within an O(log p) factor of the cost of the optimal p-way partition.

  5. Modeling geological objects with the XML Schema

    NASA Astrophysics Data System (ADS)

    Babaie, Hassan A.; Babaei, Abbed

    2005-11-01

    Interchange, storage, and management of geological data require the development of knowledge-based, standardized vocabularies and data structures. Concepts modeled and designed with the Unified Markup Language (UML), can be mapped into XML Schema Definition Language (XSDL) to compose modular markup languages for each discipline. Developing such efficient, intra-disciplinary, modular and reusable components, based on the XSDL namespace facility and the principles of object-oriented design, reduces redundancy, increases efficiency, scalability, and extensibility, and simplifies the maintenance and future extension of the code. This paper discusses the best practices of composition and reuse of modular intra-disciplinary components by applying XML Schema namespace syntax. In addition to several small examples given for a variety of geological cases, the paper constructs a UML conceptual model and markup language, applying an XML-type library, for a component of the plate tectonics knowledge base (TectonicsML) that deals with the divergent plate boundary.

  6. XML Based Scientific Data Management Facility

    NASA Technical Reports Server (NTRS)

    Mehrotra, P.; Zubair, M.; Bushnell, Dennis M. (Technical Monitor)

    2002-01-01

    The World Wide Web consortium has developed an Extensible Markup Language (XML) to support the building of better information management infrastructures. The scientific computing community realizing the benefits of XML has designed markup languages for scientific data. In this paper, we propose a XML based scientific data management ,facility, XDMF. The project is motivated by the fact that even though a lot of scientific data is being generated, it is not being shared because of lack of standards and infrastructure support for discovering and transforming the data. The proposed data management facility can be used to discover the scientific data itself, the transformation functions, and also for applying the required transformations. We have built a prototype system of the proposed data management facility that can work on different platforms. We have implemented the system using Java, and Apache XSLT engine Xalan. To support remote data and transformation functions, we had to extend the XSLT specification and the Xalan package.

  7. XML and E-Journals: The State of Play.

    ERIC Educational Resources Information Center

    Wusteman, Judith

    2003-01-01

    Discusses the introduction of the use of XML (Extensible Markup Language) in publishing electronic journals. Topics include standards, including DTDs (Document Type Definition), or document type definitions; aggregator requirements; SGML (Standard Generalized Markup Language); benefits of XML for e-journals; XML metadata; the possibility of…

  8. A Survey in Indexing and Searching XML Documents.

    ERIC Educational Resources Information Center

    Luk, Robert W. P.; Leong, H. V.; Dillon, Tharam S.; Chan, Alvin T. S.; Croft, W. Bruce; Allan, James

    2002-01-01

    Discussion of XML focuses on indexing techniques for XML documents, grouping them into flat-file, semistructured, and structured indexing paradigms. Highlights include searching techniques, including full text search and multistage search; search result presentations; database and information retrieval system integration; XML query languages; and

  9. XML: How It Will Be Applied to Digital Library Systems.

    ERIC Educational Resources Information Center

    Kim, Hyun-Hee; Choi, Chang-Seok

    2000-01-01

    Shows how XML is applied to digital library systems. Compares major features of XML with those of HTML and describes an experimental XML-based metadata retrieval system, which is based on the Dublin Core and is designed as a subsystem of the Korean Virtual Library and Information System (VINIS). (Author/LRW)

  10. XML-based resources for simulation

    NASA Astrophysics Data System (ADS)

    Kelsey, Robert L.; Riese, Jane M.; Young, Ginger A.

    2004-08-01

    As simulations and the machines they run on become larger and more complex the inputs and outputs become more unwieldy. Increased complexity makes the setup of simulation problems difficult. It also contributes to the burden of handling and analyzing large amounts of output results. Another problem is that among a class of simulation codes (such as those for physical system simulation) there is often no single standard format or resource for input data. To run the same problem on different simulations requires a different setup for each simulation code. The eXtensible Markup Language (XML) is used to represent a general set of data resources including physical system problems, materials, and test results. These resources provide a "plug and play" approach to simulation setup. For example, a particular material for a physical system can be selected from a material database. The XML-based representation of the selected material is then converted to the native format of the simulation being run and plugged into the simulation input file. In this manner a user can quickly and more easily put together a simulation setup. In the case of output data, an XML approach to regression testing includes tests and test results with XML-based representations. This facilitates the ability to query for specific tests and make comparisons between results. Also, output results can easily be converted to other formats for publishing online or on paper.

  11. Flight Dynamic Model Exchange using XML

    NASA Technical Reports Server (NTRS)

    Jackson, E. Bruce; Hildreth, Bruce L.

    2002-01-01

    The AIAA Modeling and Simulation Technical Committee has worked for several years to develop a standard by which the information needed to develop physics-based models of aircraft can be specified. The purpose of this standard is to provide a well-defined set of information, definitions, data tables and axis systems so that cooperating organizations can transfer a model from one simulation facility to another with maximum efficiency. This paper proposes using an application of the eXtensible Markup Language (XML) to implement the AIAA simulation standard. The motivation and justification for using a standard such as XML is discussed. Necessary data elements to be supported are outlined. An example of an aerodynamic model as an XML file is given. This example includes definition of independent and dependent variables for function tables, definition of key variables used to define the model, and axis systems used. The final steps necessary for implementation of the standard are presented. Software to take an XML-defined model and import/export it to/from a given simulation facility is discussed, but not demonstrated. That would be the next step in final implementation of standards for physics-based aircraft dynamic models.

  12. XML-based resources for simulation

    SciTech Connect

    Kelsey, R. L.; Riese, J. M.; Young, G. A.

    2004-01-01

    As simulations and the machines they run on become larger and more complex the inputs and outputs become more unwieldy. Increased complexity makes the setup of simulation problems difficult. It also contributes to the burden of handling and analyzing large amounts of output results. Another problem is that among a class of simulation codes (such as those for physical system simulation) there is often no single standard format or resource for input data. To run the same problem on different simulations requires a different setup for each simulation code. The extensible Markup Language (XML) is used to represent a general set of data resources including physical system problems, materials, and test results. These resources provide a 'plug and play' approach to simulation setup. For example, a particular material for a physical system can be selected from a material database. The XML-based representation of the selected material is then converted to the native format of the simulation being run and plugged into the simulation input file. In this manner a user can quickly and more easily put together a simulation setup. In the case of output data, an XML approach to regression testing includes tests and test results with XML-based representations. This facilitates the ability to query for specific tests and make comparisons between results. Also, output results can easily be converted to other formats for publishing online or on paper.

  13. Shuttle-Data-Tape XML Translator

    NASA Technical Reports Server (NTRS)

    Barry, Matthew R.; Osborne, Richard N.

    2005-01-01

    JSDTImport is a computer program for translating native Shuttle Data Tape (SDT) files from American Standard Code for Information Interchange (ASCII) format into databases in other formats. JSDTImport solves the problem of organizing the SDT content, affording flexibility to enable users to choose how to store the information in a database to better support client and server applications. JSDTImport can be dynamically configured by use of a simple Extensible Markup Language (XML) file. JSDTImport uses this XML file to define how each record and field will be parsed, its layout and definition, and how the resulting database will be structured. JSDTImport also includes a client application programming interface (API) layer that provides abstraction for the data-querying process. The API enables a user to specify the search criteria to apply in gathering all the data relevant to a query. The API can be used to organize the SDT content and translate into a native XML database. The XML format is structured into efficient sections, enabling excellent query performance by use of the XPath query language. Optionally, the content can be translated into a Structured Query Language (SQL) database for fast, reliable SQL queries on standard database server computers.

  14. Implementing XML Schema Naming and Design Rules

    SciTech Connect

    Lubell, Joshua; Kulvatunyou, Boonserm; Morris, Katherine; Harvey, Betty

    2006-08-01

    We are building a methodology and tool kit for encoding XML schema Naming and Design Rules (NDRs) in a computer-interpretable fashion, enabling automated rule enforcement and improving schema quality. Through our experience implementing rules from various NDR specifications, we discuss some issues and offer practical guidance to organizations grappling with NDR development.

  15. XTCE. XML Telemetry and Command Exchange Tutorial

    NASA Technical Reports Server (NTRS)

    Rice, Kevin; Kizzort, Brad; Simon, Jerry

    2010-01-01

    An XML Telemetry Command Exchange (XTCE) tutoral oriented towards packets or minor frames is shown. The contents include: 1) The Basics; 2) Describing Telemetry; 3) Describing the Telemetry Format; 4) Commanding; 5) Forgotten Elements; 6) Implementing XTCE; and 7) GovSat.

  16. Application of XML to Journal Table Archiving

    NASA Astrophysics Data System (ADS)

    Shaya, E. J.; Blackwell, J. H.; Gass, J. E.; Kargatis, V. E.; Schneider, G. L.; Weiland, J. L.; Borne, K. D.; White, R. A.; Cheung, C. Y.

    1998-12-01

    The Astronomical Data Center (ADC) at the NASA Goddard Space Flight Center is a major archive for machine-readable astronomical data tables. Many ADC tables are derived from published journal articles. Article tables are reformatted to be machine-readable and documentation is crafted to facilitate proper reuse by researchers. The recent switch of journals to web based electronic format has resulted in the generation of large amounts of tabular data that could be captured into machine-readable archive format at fairly low cost. The large data flow of the tables from all major North American astronomical journals (a factor of 100 greater than the present rate at the ADC) necessitates the development of rigorous standards for the exchange of data between researchers, publishers, and the archives. We have selected a suitable markup language that can fully describe the large variety of astronomical information contained in ADC tables. The eXtensible Markup Language XML is a powerful internet-ready documentation format for data. It provides a precise and clear data description language that is both machine- and human-readable. It is rapidly becoming the standard format for business and information transactions on the internet and it is an ideal common metadata exchange format. By labelling, or "marking up", all elements of the information content, documents are created that computers can easily parse. An XML archive can easily and automatically be maintained, ingested into standard databases or custom software, and even totally restructured whenever necessary. Structuring astronomical data into XML format will enable efficient and focused search capabilities via off-the-shelf software. The ADC is investigating XML's expanded hyperlinking power to enhance connectivity within the ADC data/metadata and developing XSL display scripts to enhance display of astronomical data. The ADC XML Definition Type Document can be viewed at http://messier.gsfc.nasa.gov/dtdhtml/DTD-TREE.html

  17. Time-varying modal parameters identification of a spacecraft with rotating flexible appendage by recursive algorithm

    NASA Astrophysics Data System (ADS)

    Ni, Zhiyu; Mu, Ruinan; Xun, Guangbin; Wu, Zhigang

    2016-01-01

    The rotation of spacecraft flexible appendage may cause changes in modal parameters. For this time-varying system, the computation cost of the frequently-used singular value decomposition (SVD) identification method is high. Some control problems, such as the self-adaptive control, need the latest modal parameters to update the controller parameters in time. In this paper, the projection approximation subspace tracking (PAST) recursive algorithm is applied as an alternative method to identify the time-varying modal parameters. This method avoids the SVD by signal subspace projection and improves the computational efficiency. To verify the ability of this recursive algorithm in spacecraft modal parameters identification, a spacecraft model with rapid rotational appendage, Soil Moisture Active/Passive (SMAP) satellite, is established, and the time-varying modal parameters of the satellite are identified recursively by designing the input and output signals. The results illustrate that this recursive algorithm can obtain the modal parameters in the high signal noise ratio (SNR) and it has better computational efficiency than the SVD method. Moreover, to improve the identification precision of this recursive algorithm in the low SNR, the wavelet de-noising technology is used to decrease the effect of noises.

  18. Is recursion language-specific? Evidence of recursive mechanisms in the structure of intentional action.

    PubMed

    Vicari, Giuseppe; Adenzato, Mauro

    2014-05-01

    In their 2002 seminal paper Hauser, Chomsky and Fitch hypothesize that recursion is the only human-specific and language-specific mechanism of the faculty of language. While debate focused primarily on the meaning of recursion in the hypothesis and on the human-specific and syntax-specific character of recursion, the present work focuses on the claim that recursion is language-specific. We argue that there are recursive structures in the domain of motor intentionality by way of extending John R. Searle's analysis of intentional action. We then discuss evidence from cognitive science and neuroscience supporting the claim that motor-intentional recursion is language-independent and suggest some explanatory hypotheses: (1) linguistic recursion is embodied in sensory-motor processing; (2) linguistic and motor-intentional recursions are distinct and mutually independent mechanisms. Finally, we propose some reflections about the epistemic status of HCF as presenting an empirically falsifiable hypothesis, and on the possibility of testing recursion in different cognitive domains. PMID:24762973

  19. Method for implementation of recursive hierarchical segmentation on parallel computers

    NASA Technical Reports Server (NTRS)

    Tilton, James C. (Inventor)

    2005-01-01

    A method, computer readable storage, and apparatus for implementing a recursive hierarchical segmentation algorithm on a parallel computing platform. The method includes setting a bottom level of recursion that defines where a recursive division of an image into sections stops dividing, and setting an intermediate level of recursion where the recursive division changes from a parallel implementation into a serial implementation. The segmentation algorithm is implemented according to the set levels. The method can also include setting a convergence check level of recursion with which the first level of recursion communicates with when performing a convergence check.

  20. Recursive Inversion Of Externally Defined Linear Systems

    NASA Technical Reports Server (NTRS)

    Bach, Ralph E., Jr.; Baram, Yoram

    1992-01-01

    Technical memorandum discusses mathematical technique described in "Recursive Inversion by Finite-Impulse-Response Filters" (ARC-12247). Technique is recursive algorithm yielding finite-impulse-response approximation of unknown single-input/single-output, causal, time-invariant, linear, real system, response of which is sequence of impulses. Useful in such diverse applications as medical diagnoses, identification of military targets, geophysical exploration, and nondestructive testing.

  1. A Simple XML Producer-Consumer Protocol

    NASA Technical Reports Server (NTRS)

    Smith, Warren; Gunter, Dan; Quesnel, Darcy

    2000-01-01

    This document describes a simple XML-based protocol that can be used for producers of events to communicate with consumers of events. The protocol described here is not meant to be the most efficient protocol, the most logical protocol, or the best protocol in any way. This protocol was defined quickly and it's intent is to give us a reasonable protocol that we can implement relatively easily and then use to gain experience in distributed event services. This experience will help us evaluate proposals for event representations, XML-based encoding of information, and communication protocols. The next section of this document describes how we represent events in this protocol and then defines the two events that we choose to use for our initial experiments. These definitions are made by example so that they are informal and easy to understand. The following section then proceeds to define the producer-consumer protocol we have agreed upon for our initial experiments.

  2. An exponentiation method for XML element retrieval.

    PubMed

    Wichaiwong, Tanakorn

    2014-01-01

    XML document is now widely used for modelling and storing structured documents. The structure is very rich and carries important information about contents and their relationships, for example, e-Commerce. XML data-centric collections require query terms allowing users to specify constraints on the document structure; mapping structure queries and assigning the weight are significant for the set of possibly relevant documents with respect to structural conditions. In this paper, we present an extension to the MEXIR search system that supports the combination of structural and content queries in the form of content-and-structure queries, which we call the Exponentiation function. It has been shown the structural information improve the effectiveness of the search system up to 52.60% over the baseline BM25 at MAP. PMID:24696643

  3. An Exponentiation Method for XML Element Retrieval

    PubMed Central

    2014-01-01

    XML document is now widely used for modelling and storing structured documents. The structure is very rich and carries important information about contents and their relationships, for example, e-Commerce. XML data-centric collections require query terms allowing users to specify constraints on the document structure; mapping structure queries and assigning the weight are significant for the set of possibly relevant documents with respect to structural conditions. In this paper, we present an extension to the MEXIR search system that supports the combination of structural and content queries in the form of content-and-structure queries, which we call the Exponentiation function. It has been shown the structural information improve the effectiveness of the search system up to 52.60% over the baseline BM25 at MAP. PMID:24696643

  4. Internet-based data interchange with XML

    NASA Astrophysics Data System (ADS)

    Fuerst, Karl; Schmidt, Thomas

    2000-12-01

    In this paper, a complete concept for Internet Electronic Data Interchange (EDI) - a well-known buzzword in the area of logistics and supply chain management to enable the automation of the interactions between companies and their partners - using XML (eXtensible Markup Language) will be proposed. This approach is based on Internet and XML, because the implementation of traditional EDI (e.g. EDIFACT, ANSI X.12) is mostly too costly for small and medium sized enterprises, which want to integrate their suppliers and customers in a supply chain. The paper will also present the results of the implementation of a prototype for such a system, which has been developed for an industrial partner to improve the current situation of parts delivery. The main functions of this system are an early warning system to detect problems during the parts delivery process as early as possible, and a transport following system to pursue the transportation.

  5. XML Based Scientific Data Management Facility

    NASA Technical Reports Server (NTRS)

    Mehrotra, Piyush; Zubair, M.; Ziebartt, John (Technical Monitor)

    2001-01-01

    The World Wide Web consortium has developed an Extensible Markup Language (XML) to support the building of better information management infrastructures. The scientific computing community realizing the benefits of HTML has designed markup languages for scientific data. In this paper, we propose a XML based scientific data management facility, XDMF. The project is motivated by the fact that even though a lot of scientific data is being generated, it is not being shared because of lack of standards and infrastructure support for discovering and transforming the data. The proposed data management facility can be used to discover the scientific data itself, the transformation functions, and also for applying the required transformations. We have built a prototype system of the proposed data management facility that can work on different platforms. We have implemented the system using Java, and Apache XSLT engine Xalan. To support remote data and transformation functions, we had to extend the XSLT specification and the Xalan package.

  6. The potential of XML encoding in geomatics converting raster images to XML and SVG

    NASA Astrophysics Data System (ADS)

    Antoniou, Byron; Tsoulos, Lysandros

    2006-03-01

    The evolution of open standards and especially those pertaining to the family of XML technologies, have a considerable impact on the way the Geomatics community addresses the acquisition, storage, analysis and display of spatial data. The most recent version of the GML specification enables the merging of vector and raster data into a single "open" format. The notion of "coverage" as described in GML 3.0 can be the equivalent of a raster multi-band dataset. In addition, vector data storage is also described in detail through the GML Schemas and XML itself can store the values of a raster dataset, as values of a multi-table dataset. Under these circumstances an issue that must be addressed is the transformation of raster data into XML format and their subsequent visualization through SVG. The objective of this paper is to give an overview of the steps that can be followed in order to embody open standards and XML technologies in the raster domain. The last part of the work refers to a case study that suggests a step by step methodology to accomplish classification, an important function in Cartography and Remote Sensing, using the XML-encoded images.

  7. Recursive Branching Simulated Annealing Algorithm

    NASA Technical Reports Server (NTRS)

    Bolcar, Matthew; Smith, J. Scott; Aronstein, David

    2012-01-01

    This innovation is a variation of a simulated-annealing optimization algorithm that uses a recursive-branching structure to parallelize the search of a parameter space for the globally optimal solution to an objective. The algorithm has been demonstrated to be more effective at searching a parameter space than traditional simulated-annealing methods for a particular problem of interest, and it can readily be applied to a wide variety of optimization problems, including those with a parameter space having both discrete-value parameters (combinatorial) and continuous-variable parameters. It can take the place of a conventional simulated- annealing, Monte-Carlo, or random- walk algorithm. In a conventional simulated-annealing (SA) algorithm, a starting configuration is randomly selected within the parameter space. The algorithm randomly selects another configuration from the parameter space and evaluates the objective function for that configuration. If the objective function value is better than the previous value, the new configuration is adopted as the new point of interest in the parameter space. If the objective function value is worse than the previous value, the new configuration may be adopted, with a probability determined by a temperature parameter, used in analogy to annealing in metals. As the optimization continues, the region of the parameter space from which new configurations can be selected shrinks, and in conjunction with lowering the annealing temperature (and thus lowering the probability for adopting configurations in parameter space with worse objective functions), the algorithm can converge on the globally optimal configuration. The Recursive Branching Simulated Annealing (RBSA) algorithm shares some features with the SA algorithm, notably including the basic principles that a starting configuration is randomly selected from within the parameter space, the algorithm tests other configurations with the goal of finding the globally optimal solution, and the region from which new configurations can be selected shrinks as the search continues. The key difference between these algorithms is that in the SA algorithm, a single path, or trajectory, is taken in parameter space, from the starting point to the globally optimal solution, while in the RBSA algorithm, many trajectories are taken; by exploring multiple regions of the parameter space simultaneously, the algorithm has been shown to converge on the globally optimal solution about an order of magnitude faster than when using conventional algorithms. Novel features of the RBSA algorithm include: 1. More efficient searching of the parameter space due to the branching structure, in which multiple random configurations are generated and multiple promising regions of the parameter space are explored; 2. The implementation of a trust region for each parameter in the parameter space, which provides a natural way of enforcing upper- and lower-bound constraints on the parameters; and 3. The optional use of a constrained gradient- search optimization, performed on the continuous variables around each branch s configuration in parameter space to improve search efficiency by allowing for fast fine-tuning of the continuous variables within the trust region at that configuration point.

  8. Using Heaps in Recursive Hierarchical Segmentation of Data

    NASA Technical Reports Server (NTRS)

    Tilton, James C.

    2006-01-01

    A modification to increase processing speed has been made in the algorithm and implementing software reported in Modified Recursive Hierarchical Segmentation of Data (GSC-14681-1), NASA Tech Briefs, Vol. 30, No. 6 (June 2006), page 51. That software performs recursive hierarchical segmentation of data having spatial characteristics (e.g., spectral-image data). The segmentation process includes an iterative subprocess, in each iteration of which it is necessary to determine a best pair of regions to merge [merges being justified by one or more measure(s) similarity of pixels in the regions]. In the previously reported version of the algorithm and software, the choice of a best pair of regions to merge involved the use of a fully sorted list of regions. That version was computationally inefficient because a fully sorted list is not needed: what is needed is only the identity of the pair of regions characterized by the smallest measure of dissimilarity. The present modification replaces the use of a fully sorted list with the use of data heaps, which are computationally more efficient for performing the required comparisons among dissimilarity measures. The modification includes the incorporation of standard and modified functions for creating and updating data heaps

  9. Processing heterogeneous XML data from multi-source

    NASA Astrophysics Data System (ADS)

    Wang, Tong; Liu, Da-Xin; Sun, Wei; Lin, Xuanzuo

    2006-04-01

    Recently XML heterogeneity has become a new challenge. In this paper, a novel clustering strategy is proposed to regroup these heterogeneous XML sources, for searching in a relatively smaller space with certain similarity can reduce cost. The strategy consists of four steps. We at first extract features about paths and map them into High-dimension Vector Space (HDVS). In the data pre-process, two algorithms are applied to diminish the redundancies in XML sources. Then heterogeneous documents are clustered. Finally, Multivalued Dependency (MVD) is introduced, for MVD can be redefined according to the range of constraints of XML. This paper also proposes a novel algorithm that discovering minimal MVD, based on the rough set handling non-integrity data. It can solve the problem that non-integrity data of XML influence on finding the MVD of XML, thus patterns can be extracted from each cluster.

  10. δ-dependency for privacy-preserving XML data publishing.

    PubMed

    Landberg, Anders H; Nguyen, Kinh; Pardede, Eric; Rahayu, J Wenny

    2014-08-01

    An ever increasing amount of medical data such as electronic health records, is being collected, stored, shared and managed in large online health information systems and electronic medical record systems (EMR) (Williams et al., 2001; Virtanen, 2009; Huang and Liou, 2007) [1-3]. From such rich collections, data is often published in the form of census and statistical data sets for the purpose of knowledge sharing and enabling medical research. This brings with it an increasing need for protecting individual people privacy, and it becomes an issue of great importance especially when information about patients is exposed to the public. While the concept of data privacy has been comprehensively studied for relational data, models and algorithms addressing the distinct differences and complex structure of XML data are yet to be explored. Currently, the common compromise method is to convert private XML data into relational data for publication. This ad hoc approach results in significant loss of useful semantic information previously carried in the private XML data. Health data often has very complex structure, which is best expressed in XML. In fact, XML is the standard format for exchanging (e.g. HL7 version 3(1)) and publishing health information. Lack of means to deal directly with data in XML format is inevitably a serious drawback. In this paper we propose a novel privacy protection model for XML, and an algorithm for implementing this model. We provide general rules, both for transforming a private XML schema into a published XML schema, and for mapping private XML data to the new privacy-protected published XML data. In addition, we propose a new privacy property, δ-dependency, which can be applied to both relational and XML data, and that takes into consideration the hierarchical nature of sensitive data (as opposed to "quasi-identifiers"). Lastly, we provide an implementation of our model, algorithm and privacy property, and perform an experimental analysis, to demonstrate the proposed privacy scheme in practical application. PMID:24518557

  11. Software Development Of XML Parser Based On Algebraic Tools

    NASA Astrophysics Data System (ADS)

    Georgiev, Bozhidar; Georgieva, Adriana

    2011-12-01

    In this paper, is presented one software development and implementation of an algebraic method for XML data processing, which accelerates XML parsing process. Therefore, the proposed in this article nontraditional approach for fast XML navigation with algebraic tools contributes to advanced efforts in the making of an easier user-friendly API for XML transformations. Here the proposed software for XML documents processing (parser) is easy to use and can manage files with strictly defined data structure. The purpose of the presented algorithm is to offer a new approach for search and restructuring hierarchical XML data. This approach permits fast XML documents processing, using algebraic model developed in details in previous works of the same authors. So proposed parsing mechanism is easy accessible to the web consumer who is able to control XML file processing, to search different elements (tags) in it, to delete and to add a new XML content as well. The presented various tests show higher rapidity and low consumption of resources in comparison with some existing commercial parsers.

  12. Information persistence using XML database technology

    NASA Astrophysics Data System (ADS)

    Clark, Thomas A.; Lipa, Brian E. G.; Macera, Anthony R.; Staskevich, Gennady R.

    2005-05-01

    The Joint Battlespace Infosphere (JBI) Information Management (IM) services provide information exchange and persistence capabilities that support tailored, dynamic, and timely access to required information, enabling near real-time planning, control, and execution for DoD decision making. JBI IM services will be built on a substrate of network centric core enterprise services and when transitioned, will establish an interoperable information space that aggregates, integrates, fuses, and intelligently disseminates relevant information to support effective warfighter business processes. This virtual information space provides individual users with information tailored to their specific functional responsibilities and provides a highly tailored repository of, or access to, information that is designed to support a specific Community of Interest (COI), geographic area or mission. Critical to effective operation of JBI IM services is the implementation of repositories, where data, represented as information, is represented and persisted for quick and easy retrieval. This paper will address information representation, persistence and retrieval using existing database technologies to manage structured data in Extensible Markup Language (XML) format as well as unstructured data in an IM services-oriented environment. Three basic categories of database technologies will be compared and contrasted: Relational, XML-Enabled, and Native XML. These technologies have diverse properties such as maturity, performance, query language specifications, indexing, and retrieval methods. We will describe our application of these evolving technologies within the context of a JBI Reference Implementation (RI) by providing some hopefully insightful anecdotes and lessons learned along the way. This paper will also outline future directions, promising technologies and emerging COTS products that can offer more powerful information management representations, better persistence mechanisms and improved retrieval techniques.

  13. A Simple XML Producer-Consumer Protocol

    NASA Technical Reports Server (NTRS)

    Smith, Warren; Gunter, Dan; Quesnel, Darcy; Biegel, Bryan (Technical Monitor)

    2001-01-01

    There are many different projects from government, academia, and industry that provide services for delivering events in distributed environments. The problem with these event services is that they are not general enough to support all uses and they speak different protocols so that they cannot interoperate. We require such interoperability when we, for example, wish to analyze the performance of an application in a distributed environment. Such an analysis might require performance information from the application, computer systems, networks, and scientific instruments. In this work we propose and evaluate a standard XML-based protocol for the transmission of events in distributed systems. One recent trend in government and academic research is the development and deployment of computational grids. Computational grids are large-scale distributed systems that typically consist of high-performance compute, storage, and networking resources. Examples of such computational grids are the DOE Science Grid, the NASA Information Power Grid (IPG), and the NSF Partnerships for Advanced Computing Infrastructure (PACIs). The major effort to deploy these grids is in the area of developing the software services to allow users to execute applications on these large and diverse sets of resources. These services include security, execution of remote applications, managing remote data, access to information about resources and services, and so on. There are several toolkits for providing these services such as Globus, Legion, and Condor. As part of these efforts to develop computational grids, the Global Grid Forum is working to standardize the protocols and APIs used by various grid services. This standardization will allow interoperability between the client and server software of the toolkits that are providing the grid services. The goal of the Performance Working Group of the Grid Forum is to standardize protocols and representations related to the storage and distribution of performance data. These standard protocols and representations must support tasks such as profiling parallel applications, monitoring the status of computers and networks, and monitoring the performance of services provided by a computational grid. This paper describes a proposed protocol and data representation for the exchange of events in a distributed system. The protocol exchanges messages formatted in XML and it can be layered atop any low-level communication protocol such as TCP or UDP Further, we describe Java and C++ implementations of this protocol and discuss their performance. The next section will provide some further background information. Section 3 describes the main communication patterns of our protocol. Section 4 describes how we represent events and related information using XML. Section 5 describes our protocol and Section 6 discusses the performance of two implementations of the protocol. Finally, an appendix provides the XML Schema definition of our protocol and event information.

  14. Experimental Evaluation of Processing Time for the Synchronization of XML-Based Business Objects

    NASA Astrophysics Data System (ADS)

    Ameling, Michael; Wolf, Bernhard; Springer, Thomas; Schill, Alexander

    Business objects (BOs) are data containers for complex data structures used in business applications such as Supply Chain Management and Customer Relationship Management. Due to the replication of application logic, multiple copies of BOs are created which have to be synchronized and updated. This is a complex and time consuming task because BOs rigorously vary in their structure according to the distribution, number and size of elements. Since BOs are internally represented as XML documents, the parsing of XML is one major cost factor which has to be considered for minimizing the processing time during synchronization. The prediction of the parsing time for BOs is an significant property for the selection of an efficient synchronization mechanism. In this paper, we present a method to evaluate the influence of the structure of BOs on their parsing time. The results of our experimental evaluation incorporating four different XML parsers examine the dependencies between the distribution of elements and the parsing time. Finally, a general cost model will be validated and simplified according to the results of the experimental setup.

  15. Recursive Implementations of the Consider Filter

    NASA Technical Reports Server (NTRS)

    Zanetti, Renato; DSouza, Chris

    2012-01-01

    One method to account for parameters errors in the Kalman filter is to consider their effect in the so-called Schmidt-Kalman filter. This work addresses issues that arise when implementing a consider Kalman filter as a real-time, recursive algorithm. A favorite implementation of the Kalman filter as an onboard navigation subsystem is the UDU formulation. A new way to implement a UDU consider filter is proposed. The non-optimality of the recursive consider filter is also analyzed, and a modified algorithm is proposed to overcome this limitation.

  16. Parallel scheduling of recursively defined arrays

    NASA Technical Reports Server (NTRS)

    Myers, T. J.; Gokhale, M. B.

    1986-01-01

    A new method of automatic generation of concurrent programs which constructs arrays defined by sets of recursive equations is described. It is assumed that the time of computation of an array element is a linear combination of its indices, and integer programming is used to seek a succession of hyperplanes along which array elements can be computed concurrently. The method can be used to schedule equations involving variable length dependency vectors and mutually recursive arrays. Portions of the work reported here have been implemented in the PS automatic program generation system.

  17. Systolic MVDR beamforming with inverse updating

    NASA Astrophysics Data System (ADS)

    Moonen, M.

    1993-06-01

    A stable alternative is described for the 'standard' systolic MVDR (minimum variance distortionless response) beamforming algorithm of McWhirter and Shepherd (1986), which is shown to be numerically unstable. This alternative algorithm is similar to covariance-type recursive least squares algorithms that employ 'inverse updating'. Required a posteriori residuals for updating are computed from the stored inverse matrix together with the Kalman gain vector. The beamforming problem is shown to fit on a systolic array for inverse updating.

  18. A Recursive Method for Calculating Certain Partition Functions.

    ERIC Educational Resources Information Center

    Woodrum, Luther; And Others

    1978-01-01

    Describes a simple recursive method for calculating the partition function and average energy of a system consisting of N electrons and L energy levels. Also, presents an efficient APL computer program to utilize the recursion relation. (Author/GA)

  19. A recursive Bayesian approach for fatigue damage prognosis: An experimental validation at the reliability component level

    NASA Astrophysics Data System (ADS)

    Gobbato, Maurizio; Kosmatka, John B.; Conte, Joel P.

    2014-04-01

    Fatigue-induced damage is one of the most uncertain and highly unpredictable failure mechanisms for a large variety of mechanical and structural systems subjected to cyclic and random loads during their service life. A health monitoring system capable of (i) monitoring the critical components of these systems through non-destructive evaluation (NDE) techniques, (ii) assessing their structural integrity, (iii) recursively predicting their remaining fatigue life (RFL), and (iv) providing a cost-efficient reliability-based inspection and maintenance plan (RBIM) is therefore ultimately needed. In contribution to these objectives, the first part of the paper provides an overview and extension of a comprehensive reliability-based fatigue damage prognosis methodology — previously developed by the authors — for recursively predicting and updating the RFL of critical structural components and/or sub-components in aerospace structures. In the second part of the paper, a set of experimental fatigue test data, available in the literature, is used to provide a numerical verification and an experimental validation of the proposed framework at the reliability component level (i.e., single damage mechanism evolving at a single damage location). The results obtained from this study demonstrate (i) the importance and the benefits of a nearly continuous NDE monitoring system, (ii) the efficiency of the recursive Bayesian updating scheme, and (iii) the robustness of the proposed framework in recursively updating and improving the RFL estimations. This study also demonstrates that the proposed methodology can lead to either an extent of the RFL (with a consequent economical gain without compromising the minimum safety requirements) or an increase of safety by detecting a premature fault and therefore avoiding a very costly catastrophic failure.

  20. Adding XML to the MIS Curriculum: Lessons from the Classroom

    ERIC Educational Resources Information Center

    Wagner, William P.; Pant, Vik; Hilken, Ralph

    2008-01-01

    eXtensible Markup Language (XML) is a new technology that is currently being extolled by many industry experts and software vendors. Potentially it represents a platform independent language for sharing information over networks in a way that is much more seamless than with previous technologies. It is extensible in that XML serves as a "meta"

  1. Data Manipulation in an XML-Based Digital Image Library

    ERIC Educational Resources Information Center

    Chang, Naicheng

    2005-01-01

    Purpose: To help to clarify the role of XML tools and standards in supporting transition and migration towards a fully XML-based environment for managing access to information. Design/methodology/approach: The Ching Digital Image Library, built on a three-tier architecture, is used as a source of examples to illustrate a number of methods of data

  2. A Typed Text Retrieval Query Language for XML Documents.

    ERIC Educational Resources Information Center

    Colazzo, Dario; Sartiani, Carlo; Albano, Antonio; Manghi, Paolo; Ghelli, Giorgio; Lini, Luca; Paoli, Michele

    2002-01-01

    Discussion of XML focuses on a description of Tequyla-TX, a typed text retrieval query language for XML documents that can search on both content and structures. Highlights include motivations; numerous examples; word-based and char-based searches; tag-dependent full-text searches; text normalization; query algebra; data models and term language;…

  3. An XML file format for exchanging singlet lens specifications

    NASA Astrophysics Data System (ADS)

    Gay, Shawn C.; Gangadhara, Sanjay

    2015-10-01

    Zemax has developed an XML schema for the distribution of singlet lens specifications based on the ISO 10110 standard. In OpticStudio 15, this kind of XML data can be exported from the ISO Element Drawing analysis. The data file is then used in a feature that automates exchange of lens data between designer and manufacturer, the Cost Estimator. This Cost Estimator feature submits the XML data to various manufacturers to obtain cost estimates for prototype lens production. The workflow centered on the XML data exchange facilitates rapid cost estimate retrieval and eliminates the need for redundant manual data entry. The XML Schema Definition (XSD) for the XML format can be used with Microsoft developer tools to automatically create .NET classes to serialize and deserialize the singlet lens data to/from XML files. The format provides flexible unit specification for most parameters. Choosing XML as the basis for the file format has provided several benefits, such as the above mentioned automated serialization capabilities in .NET, a human-readable text-based format, and ready support for consumption by web services.

  4. EquiX-A Search and Query Language for XML.

    ERIC Educational Resources Information Center

    Cohen, Sara; Kanza, Yaron; Kogan, Yakov; Sagiv, Yehoshua; Nutt, Werner; Serebrenik, Alexander

    2002-01-01

    Describes EquiX, a search language for XML that combines querying with searching to query the data and the meta-data content of Web pages. Topics include search engines; a data model for XML documents; search query syntax; search query semantics; an algorithm for evaluating a query on a document; and indexing EquiX queries. (LRW)

  5. An Expressive and Efficient Language for XML Information Retrieval.

    ERIC Educational Resources Information Center

    Chinenyanga, Taurai Tapiwa; Kushmerick, Nicholas

    2002-01-01

    Discusses XML and information retrieval and describes a query language, ELIXIR (expressive and efficient language for XML information retrieval), with a textual similarity operator that can be used for similarity joins. Explains the algorithm for answering ELIXIR queries to generate intermediate relational data. (Author/LRW)

  6. Data Manipulation in an XML-Based Digital Image Library

    ERIC Educational Resources Information Center

    Chang, Naicheng

    2005-01-01

    Purpose: To help to clarify the role of XML tools and standards in supporting transition and migration towards a fully XML-based environment for managing access to information. Design/methodology/approach: The Ching Digital Image Library, built on a three-tier architecture, is used as a source of examples to illustrate a number of methods of data…

  7. Adding XML to the MIS Curriculum: Lessons from the Classroom

    ERIC Educational Resources Information Center

    Wagner, William P.; Pant, Vik; Hilken, Ralph

    2008-01-01

    eXtensible Markup Language (XML) is a new technology that is currently being extolled by many industry experts and software vendors. Potentially it represents a platform independent language for sharing information over networks in a way that is much more seamless than with previous technologies. It is extensible in that XML serves as a "meta"…

  8. Does Being Technical Matter? XML, Single Source, and Technical Communication.

    ERIC Educational Resources Information Center

    Sapienza, Filipp

    2002-01-01

    Describes XML, a recent Web design language that will enable technical communicators to produce documentation that can reuse information and present it across multiple types of media for diverse audiences. Argues that XML requires more interdisciplinary approaches toward the teaching and research of technical communication, particularly with…

  9. RECURSIVE PARAMETER ESTIMATION OF HYDROLOGIC MODELS

    EPA Science Inventory

    Proposed is a nonlinear filtering approach to recursive parameter estimation of conceptual watershed response models in state-space form. he conceptual model state is augmented by the vector of free parameters which are to be estimated from input-output data, and the extended Kal...

  10. Recursive harmonic analysis for computing Hansen coefficients

    NASA Astrophysics Data System (ADS)

    Adel Sharaf, Mohamed; Hassan Selim, Hadia

    2010-12-01

    We report on a simple pure numerical method developed for computing Hansen coefficients by using a recursive harmonic analysis technique. The precision criteria of the computations are very satisfactory and provide materials for computing Hansen's and Hansen's like expansions, and also to check the accuracy of some existing algorithms.

  11. A Recursive Approach to Office Object Modelling.

    ERIC Educational Resources Information Center

    Lamersdorf, W.; And Others

    1986-01-01

    Identifies requirements for representing structural aspects of nonconventional information system applications and demonstrates how recursive data models support information systems better than record-based database modelling tools. Structural concepts for representing office data objects are discussed and applied to an office example. Language…

  12. Recursive Generation of Space-Times

    NASA Astrophysics Data System (ADS)

    Marks, Dennis

    2015-04-01

    Space-times can be generated recursively from a time-like unit basis vector T and a space-like one S. T is unique up to sign, corresponding to particles and antiparticles. S has the form of qubits. Qubits can make quantum transitions, suggesting spontaneous generation of space-time. Recursive generation leads from 2 dimensions to 4, with grades of the resulting algebra corresponding to space-time, spin-area, momentum-energy, and action. Dimensions can be open (like space-time) or closed. A closed time-like dimension has the symmetry of electromagnetism; 3 closed space-like dimensions have the symmetry of the weak force. The 4 open dimensions and the 4 closed dimensions produce an 8-dimensional space with a symmetry that is the product of the Yang regularization of the Heisenberg-Poincaré group and the GUT regularization of the Standard Model. After 8 dimensions, the pattern of real geometric algebras repeats itself, producing a recursive lattice of spontaneously expanding space-time with the physics of the Standard Model at each point of the lattice, implying conservation laws by Noether's theorem. The laws of nature are not preexistent; rather, they are consequences of the uniformity of space-time. The uniformity of space-time is a consequence of its recursive generation.

  13. On the design of recursive digital filters

    NASA Technical Reports Server (NTRS)

    Shenoi, K.; Narasimha, M. J.; Peterson, A. M.

    1976-01-01

    A change of variables is described which transforms the problem of designing a recursive digital filter to that of approximation by a ratio of polynomials on a finite interval. Some analytic techniques for the design of low-pass filters are presented, illustrating the use of the transformation. Also considered are methods for the design of phase equalizers.

  14. A recursive filter for despeckling SAR images.

    PubMed

    Subrahmanyam, G R K S; Rajagopalan, A N; Aravind, R

    2008-10-01

    This correspondence proposes a recursive algorithm for noise reduction in synthetic aperture radar imagery. Excellent despeckling in conjunction with feature preservation is achieved by incorporating a discontinuity-adaptive Markov random field prior within the unscented Kalman filter framework through importance sampling. The performance of this method is demonstrated on both synthetic and real examples. PMID:18784043

  15. Causal Models in Educational Research: Recursive Models.

    ERIC Educational Resources Information Center

    Anderson, James G.; Evans, Francis B.

    Causal modelling procedures developed in economics and biology provide social scientists with powerful methodological tools that permit them to bridge the gap between theory and research. In this paper one type of causal modelling technique involving a structural set of equations that are recursive in form has been used to reanalyze the data from…

  16. Recursive inversion of externally defined linear systems

    NASA Technical Reports Server (NTRS)

    Bach, Ralph E., Jr.; Baram, Yoram

    1988-01-01

    The approximate inversion of an internally unknown linear system, given by its impulse response sequence, by an inverse system having a finite impulse response, is considered. The recursive least squares procedure is shown to have an exact initialization, based on the triangular Toeplitz structure of the matrix involved. The proposed approach also suggests solutions to the problems of system identification and compensation.

  17. Recursive splicing in long vertebrate genes

    PubMed Central

    Blazquez, Lorea; Faro, Ana; Haberman, Nejc; Briese, Michael; Trabzuni, Daniah; Ryten, Mina; Weale, Michael E; Hardy, John; Modic, Miha; Curk, Tomaž; Wilson, Stephen W; Plagnol, Vincent; Ule, Jernej

    2015-01-01

    It is generally believed that splicing removes introns as single units from pre-mRNA transcripts. However, some long D. melanogaster introns contain a cryptic site, called a recursive splice site (RS-site), that enables a multi-step process of intron removal termed recursive splicing1,2. The extent to which recursive splicing occurs in other species and its mechanistic basis remain unclear. Here we identify highly conserved RS-sites in genes expressed in the mammalian brain that encode proteins functioning in neuronal development. Moreover, the RS-sites are found in some of the longest introns across vertebrates. We find that vertebrate recursive splicing requires initial definition of a “RS-exon” that follows the RS-site. The RS-exon is then excluded from the dominant mRNA isoform due to competition with a reconstituted 5′ splice site formed at the RS-site after the first splicing step. Conversely, the RS-exon is included when preceded by cryptic exons or promoters that are prevalent in long introns, but which fail to reconstitute an efficient 5′ splice site. Most RS-exons contain a premature stop codon such that their inclusion may decrease mRNA stability. Thus, by establishing a binary splicing switch, RS-sites demarcate different mRNA isoforms emerging from long genes by coupling inclusion of cryptic elements with RS-exons. PMID:25970246

  18. Recursive splicing in long vertebrate genes.

    PubMed

    Sibley, Christopher R; Emmett, Warren; Blazquez, Lorea; Faro, Ana; Haberman, Nejc; Briese, Michael; Trabzuni, Daniah; Ryten, Mina; Weale, Michael E; Hardy, John; Modic, Miha; Curk, Tomaž; Wilson, Stephen W; Plagnol, Vincent; Ule, Jernej

    2015-05-21

    It is generally believed that splicing removes introns as single units from precursor messenger RNA transcripts. However, some long Drosophila melanogaster introns contain a cryptic site, known as a recursive splice site (RS-site), that enables a multi-step process of intron removal termed recursive splicing. The extent to which recursive splicing occurs in other species and its mechanistic basis have not been examined. Here we identify highly conserved RS-sites in genes expressed in the mammalian brain that encode proteins functioning in neuronal development. Moreover, the RS-sites are found in some of the longest introns across vertebrates. We find that vertebrate recursive splicing requires initial definition of an 'RS-exon' that follows the RS-site. The RS-exon is then excluded from the dominant mRNA isoform owing to competition with a reconstituted 5' splice site formed at the RS-site after the first splicing step. Conversely, the RS-exon is included when preceded by cryptic promoters or exons that fail to reconstitute an efficient 5' splice site. Most RS-exons contain a premature stop codon such that their inclusion can decrease mRNA stability. Thus, by establishing a binary splicing switch, RS-sites demarcate different mRNA isoforms emerging from long genes by coupling cryptic elements with inclusion of RS-exons. PMID:25970246

  19. TORTIS (Toddler's Own Recursive Turtle Interpreter System).

    ERIC Educational Resources Information Center

    Perlman, Radia

    TORTIS (Toddler's Own Recursive Turtle Interpreter System) is a device which can be used to study or nurture the cognitive development of preschool children. The device consists of a "turtle" which the child can control by use of buttons on a control panel. The "turtle" can be made to move in prescribed directions, to take a given number of paces,…

  20. Converting European-Pulsar-Network Data to XML

    NASA Astrophysics Data System (ADS)

    Seiradakis, J. H.; Radioti, E.

    2006-04-01

    The European Pulsar Network (EPN) format for pulsar data was discussed, reviewed and established about 10 years ago, during the EU Human Capital and Mobility project. Since then it has been extensively used in most pulsar observing facilities worldwide. It is a flexible format, able to accommodate the various data-structures of pulsar data. Its use has enhanced collaborative pulsar research and corroborated our understanding of these exotic compact stars. Three years ago it was decided to create a new tool to convert EPN-formatted data to XML (eXtensible Markup Language). The tool automatically converts EPN files to XML Physical Structure (xml) and XML Document Type Definition (dtd) files. The XML-formatted pulsar data allows the use of a wide range of analysis packages to be applied to pulsar research.

  1. A Survey on Teaching and Learning Recursive Programming

    ERIC Educational Resources Information Center

    Rinderknecht, Christian

    2014-01-01

    We survey the literature about the teaching and learning of recursive programming. After a short history of the advent of recursion in programming languages and its adoption by programmers, we present curricular approaches to recursion, including a review of textbooks and some programming methodology, as well as the functional and imperative…

  2. Scheduling Topics for Improved Student Comprehension of Recursion

    ERIC Educational Resources Information Center

    Zmuda, Michael; Hatch, Melanie

    2007-01-01

    This paper presents the results of an experiment conducted to assess the affects of teaching recursion in two disjoint, non-consecutive units of instruction. One group of students was taught basic and advanced recursion topics in four consecutive class periods, while a second group was taught recursion in two two-period blocks that were separated…

  3. BCFW recursion relations and string theory

    NASA Astrophysics Data System (ADS)

    Cheung, Clifford; O'Connell, Donal; Wecht, Brian

    2010-09-01

    We demonstrate that all tree-level string theory amplitudes can be computed using the BCFW recursion relations. Our proof utilizes the pomeron vertex operator introduced by Brower, Polchinski, Strassler, and Tan. Surprisingly, we find that in a particular large complex momentum limit, the asymptotic expansion of massless string amplitudes is identical in form to that of the corresponding field theory amplitudes. This observation makes manifest the fact that field-theoretic Yang-Mills and graviton amplitudes obey KLT-like relations. Moreover, we conjecture that in this large momentum limit certain string theory and field theory amplitudes are identical, and provide evidence for this conjecture. Additionally, we find a new recursion relation which relates tachyon amplitudes to lower-point tachyon amplitudes.

  4. The recursion relation in Lagrangian perturbation theory

    SciTech Connect

    Rampf, Cornelius

    2012-12-01

    We derive a recursion relation in the framework of Lagrangian perturbation theory, appropriate for studying the inhomogeneities of the large scale structure of the universe. We use the fact that the perturbative expansion of the matter density contrast is in one-to-one correspondence with standard perturbation theory (SPT) at any order. This correspondence has been recently shown to be valid up to fourth order for a non-relativistic, irrotational and dust-like component. Assuming it to be valid at arbitrary (higher) order, we express the Lagrangian displacement field in terms of the perturbative kernels of SPT, which are itself given by their own and well-known recursion relation. We argue that the Lagrangian solution always contains more non-linear information in comparison with the SPT solution, (mainly) if the non-perturbative density contrast is restored after the displacement field is obtained.

  5. Geomagnetic modeling by optimal recursive filtering

    NASA Technical Reports Server (NTRS)

    Gibbs, B. P.; Estes, R. H.

    1981-01-01

    The results of a preliminary study to determine the feasibility of using Kalman filter techniques for geomagnetic field modeling are given. Specifically, five separate field models were computed using observatory annual means, satellite, survey and airborne data for the years 1950 to 1976. Each of the individual field models used approximately five years of data. These five models were combined using a recursive information filter (a Kalman filter written in terms of information matrices rather than covariance matrices.) The resulting estimate of the geomagnetic field and its secular variation was propogated four years past the data to the time of the MAGSAT data. The accuracy with which this field model matched the MAGSAT data was evaluated by comparisons with predictions from other pre-MAGSAT field models. The field estimate obtained by recursive estimation was found to be superior to all other models.

  6. Acoustic multiple scattering using recursive algorithms

    NASA Astrophysics Data System (ADS)

    Amirkulova, Feruza A.; Norris, Andrew N.

    2015-10-01

    Acoustic multiple scattering by a cluster of cylinders in an acoustic medium is considered. A fast recursive technique is described which takes advantage of the multilevel Block Toeplitz structure of the linear system. A parallelization technique is described that enables efficient application of the proposed recursive algorithm for solving multilevel Block Toeplitz systems on high performance computer clusters. Numerical comparisons of CPU time and total elapsed time taken to solve the linear system using the direct LAPACK and TOEPLITZ libraries on Intel FORTRAN, show the advantage of the TOEPLITZ solver. Computations are optimized by multi-threading which displays improved efficiency of the TOEPLITZ solver with the increase of the number of scatterers and frequency.

  7. Recursive dynamics for geared robot manipulators

    NASA Technical Reports Server (NTRS)

    Jain, A.; Rodriguez, G.

    1990-01-01

    The authors consider the dynamical modeling of robot manipulators whose joint actuators consist of motors driving the joints through gears. The dynamical models for such manipulators are significantly more complex than those for direct drive manipulators. The authors develop recursive O(n) inverse and forward dynamics algorithms as well as recursive O(n2) algorithms for the computation of the mass matrix for geared manipulators. It is shown that, despite the added complexity of the dynamical models for geared manipulators, the algorithms closely resemble the corresponding algorithms for direct drive manipulators, and that the additional algorithmic or computational complexity is relatively insignificant. As a consequence, with little additional cost, existing direct drive algorithms can be easily extended to handle the effects of gearing at the joints.

  8. Weighted Laplacians, cocycles and recursion relations

    NASA Astrophysics Data System (ADS)

    Krasnov, Kirill; Scarinci, Carlos

    2013-11-01

    Hodge's formula represents the gravitational MHV amplitude as the determinant of a minor of a certain matrix. When expanded, this determinant becomes a sum over weighted trees, which is the form of the MHV formula first obtained by Bern, Dixon, Perelstein, Rozowsky and rediscovered by Nguyen, Spradlin, Volovich and Wen. The gravity MHV amplitude satisfies the Britto, Cachazo, Feng and Witten recursion relation. The main building block of the MHV amplitude, the so-called half-soft function, satisfies a different, Berends-Giele-type recursion relation. We show that all these facts are illustrations to a more general story. We consider a weighted Laplacian for a complete graph of n vertices. The matrix tree theorem states that its diagonal minor determinants are all equal and given by a sum over spanning trees. We show that, for any choice of a cocycle on the graph, the minor determinants satisfy a Berends-Giele as well as Britto-Cachazo-Feng-Witten type recursion relation. Our proofs are purely combinatorial.

  9. Using XML to Separate Content from the Presentation Software in eLearning Applications

    ERIC Educational Resources Information Center

    Merrill, Paul F.

    2005-01-01

    This paper has shown how XML (extensible Markup Language) can be used to mark up content. Since XML documents, with meaningful tags, can be interpreted easily by humans as well as computers, they are ideal for the interchange of information. Because XML tags can be defined by an individual or organization, XML documents have proven useful in a

  10. A Study of XML in the Library Science Curriculum in Taiwan and South East Asia

    ERIC Educational Resources Information Center

    Chang, Naicheng; Huang, Yuhui; Hopkinson, Alan

    2011-01-01

    This paper aims to investigate the current XML-related courses available in 96 LIS schools in South East Asia and Taiwan's 9 LIS schools. Also, this study investigates the linkage of library school graduates in Taiwan who took different levels of XML-related education (that is XML arranged as an individual course or XML arranged as a section unit…

  11. A Study of XML in the Library Science Curriculum in Taiwan and South East Asia

    ERIC Educational Resources Information Center

    Chang, Naicheng; Huang, Yuhui; Hopkinson, Alan

    2011-01-01

    This paper aims to investigate the current XML-related courses available in 96 LIS schools in South East Asia and Taiwan's 9 LIS schools. Also, this study investigates the linkage of library school graduates in Taiwan who took different levels of XML-related education (that is XML arranged as an individual course or XML arranged as a section unit

  12. Using XML to Separate Content from the Presentation Software in eLearning Applications

    ERIC Educational Resources Information Center

    Merrill, Paul F.

    2005-01-01

    This paper has shown how XML (extensible Markup Language) can be used to mark up content. Since XML documents, with meaningful tags, can be interpreted easily by humans as well as computers, they are ideal for the interchange of information. Because XML tags can be defined by an individual or organization, XML documents have proven useful in a…

  13. XML DTD and Schemas for HDF-EOS

    NASA Technical Reports Server (NTRS)

    Ullman, Richard; Yang, Jingli

    2008-01-01

    An Extensible Markup Language (XML) document type definition (DTD) standard for the structure and contents of HDF-EOS files and their contents, and an equivalent standard in the form of schemas, have been developed.

  14. On the usage of XML file format in geodynamic calculations

    NASA Astrophysics Data System (ADS)

    Choliy, V.

    2012-12-01

    We recommend the extended usage of XML data format for the representation of geodynamic observations and processing results. A short introduction to the technology and a simple example for a Consolidated Laser Ranging Data Format (CRD) data file are presented.

  15. Citing geospatial feature inventories with XML manifests

    NASA Astrophysics Data System (ADS)

    Bose, R.; McGarva, G.

    2006-12-01

    Today published scientific papers include a growing number of citations for online information sources that either complement or replace printed journals and books. We anticipate this same trend for cartographic citations used in the geosciences, following advances in web mapping and geographic feature-based services. Instead of using traditional libraries to resolve citations for print material, the geospatial citation life cycle will include requesting inventories of objects or geographic features from distributed geospatial data repositories. Using a case study from the UK Ordnance Survey MasterMap database, which is illustrative of geographic object-based products in general, we propose citing inventories of geographic objects using XML feature manifests. These manifests: (1) serve as a portable listing of sets of versioned features; (2) could be used as citations within the identification portion of an international geospatial metadata standard; (3) could be incorporated into geospatial data transfer formats such as GML; but (4) can be resolved only with comprehensive, curated repositories of current and historic data. This work has implications for any researcher who foresees the need to make or resolve references to online geospatial databases.

  16. An XML cytometry standard based on DICOM

    NASA Astrophysics Data System (ADS)

    Leif, Robert C.

    2009-02-01

    Introduction: The International Society for the Advancement of Cytometry (ISAC) Data Standards Task Force (DSTF) is developing a new Advanced Cytometry Specification (ACS). DICOM has developed and is extending a pathology extension. The work of both groups is complementary with some overlap. Interoperation would benefit both groups and permit each to benefit from the other's expertise. Methods: The design and implementation of the CytometryML version of the ACS schemas have been based on each schema describing one object (modularity), iterative (spiral) development, inheritance, and reuse of data-types and their definitions from DICOM, Flow Cytometry Standard, and other standards. Results: These schemas have been validated with two tools and XML pages were generated from highest level schemas. Binary image data and its associated metadata are stored together in a zip file based container. A schema for a table of contents, which is one of the metadata files of this container, has recently been developed and reported upon. The binary image data is placed in one file in the container; and the metadata associated with an image in another. The schema for the image metadata file includes elements that are based on the DICOM design. This image schema includes descriptions of the acquisition context, image (including information on compression), specimen, slide, transmission medium, major optical parts, optical elements in one or more optical channels, detectors, and pixel format. The image schema describes both conventional camera systems and scanning or confocal systems.

  17. Specifics on a XML Data Format for Scientific Data

    NASA Astrophysics Data System (ADS)

    Shaya, E.; Thomas, B.; Cheung, C.

    An XML-based data format for interchange and archiving of scientific data would benefit in many ways from the features standardized in XML. Foremost of these features is the world-wide acceptance and adoption of XML. Applications, such as browsers, XQL and XSQL advanced query, XML editing, or CSS or XSLT transformation, that are coming out of industry and academia can be easily adopted and provide startling new benefits and features. We have designed a prototype of a core format for holding, in a very general way, parameters, tables, scalar and vector fields, atlases, animations and complex combinations of these. This eXtensible Data Format (XDF) makes use of XML functionalities such as: self-validation of document structure, default values for attributes, XLink hyperlinks, entity replacements, internal referencing, inheritance, and XSLT transformation. An API is available to aid in detailed assembly, extraction, and manipulation. Conversion tools to and from FITS and other existing data formats are under development. In the future, we hope to provide object oriented interfaces to C++, Java, Python, IDL, Mathematica, Maple, and various databases. http://xml.gsfc.nasa.gov/XDF

  18. Path Query Processing in Large-Scale XML Databases

    NASA Astrophysics Data System (ADS)

    Haw, Su-Cheng; Radha Krishna Rao, G. S. V.

    With the ever-increasing popularity of XML (e-Xtensible Markup Language) as data representation and exchange on the Internet, querying XML data has become an important issue to be address. In Native XML Database (NXD), XML documents are usually modeled as trees and XML queries are typically specified in path expression. In path expression, the primitive structural relationships are Parent-Child (P-C) and Ancestor-Descendant (A-D). Thus, finding all occurrences of these relationships is crucial for XML query processing. Current methods for query processing on NXD usually employ either sequential traversing of tree-structured model or a decomposition-matching-merging processes. We adopt the later approach and propose a novel hybrid query optimization technique, INLAB comprising both indexing and labeling technologies. Furthermore, we also propose several algorithms to create INLAB encoding and analyze the path query. We implemented our technique and present performance results over several benchmarking datasets, which prove the viability of our approach.

  19. Structural Query Optimization in Native XML Databases: A Hybrid Approach

    NASA Astrophysics Data System (ADS)

    Haw, Su-Cheng; Lee, Chien-Sing

    As XML (eXtensible Mark-up Language) is gaining its popularity in data exchange over the Web, querying XML data has become an important issue to be addressed. In native XML databases (NXD), XML documents are usually modeled as trees and XML queries are typically specified in path expression. The primitive structural relationships are Parent-Child (P-C), Ancestor-Descendant (A-D), sibling and ordered query. Thus, a suitable and compact labeling scheme is crucial to identify these relationships and henceforth to process the query efficiently. We propose a novel labeling scheme consisting of < self-level:parent> to support all these relationships efficiently. Besides, we adopt the decomposition-matching-merging approach for structural query processing and propose a hybrid query optimization technique, TwigINLAB to process and optimize the twig query evaluation. Experimental results indicate that TwigINLAB can process all types of XML queries 15% better than the TwigStack algorithm in terms of execution time in most test cases.

  20. Fast bilateral filtering using recursive moving sum

    NASA Astrophysics Data System (ADS)

    Igarashi, Masaki; Ikebe, Masayuki; Shimoyama, Sohsuke; Motohisa, Junichi

    We propose a constant-time algorithm for a bilateral filter. Bilateral filter can be converted into the operation of three-dimensional (3D) convolution. By using recursive moving sum, we can reduce the number of calculations needed to construct a pseudo-Gaussian filter. Applying one-dimensional Gaussian filter to the 3D convolution, we achieved a constant-time bilateral filter. We used a 3-GHz CPU without SIMD instructions, or multi-thread operations. We confirmed our proposed bilateral filter to be processed in constant time. In practical conditions, high PSNR values over 40 dB are obtained.

  1. Elimination and recursions in the scattering equations

    NASA Astrophysics Data System (ADS)

    Cardona, Carlos; Kalousios, Chrysostomos

    2016-05-01

    We use the elimination theory to explicitly construct the (n - 3) ! order polynomial in one of the variables of the scattering equations. The answer can be given either in terms of a determinant of Sylvester type of dimension (n - 3) ! or a determinant of Bézout type of dimension (n - 4) !. We present a recursive formula for the Sylvester determinant. Expansion of the determinants yields expressions in terms of Plücker coordinates. Elimination of the rest of the variables of the scattering equations is also presented.

  2. Recursion Theoretic Operators for Function Complexity Classes

    NASA Astrophysics Data System (ADS)

    Ueno, Kenya

    We characterize the gap between time and space complexity of functions by operators and completeness. First, we introduce a new notion of operators for function complexity classes based on recursive function theory and construct an operator which generates FPS PACE from FP. Then, we introduce new function classes composed of functions whose output lengths are bounded by the input length plus some constant. We characterize FP and FPS PACE by using these classes and operators. Finally, we define a new notion of completeness for FPS PACE and show a FPS PACE-complete function.

  3. Modified Recursive Hierarchical Segmentation of Data

    NASA Technical Reports Server (NTRS)

    Tilton, James C.

    2006-01-01

    An algorithm and a computer program that implements the algorithm that performs recursive hierarchical segmentation (RHSEG) of data have been developed. While the current implementation is for two-dimensional data having spatial characteristics (e.g., image, spectral, or spectral-image data), the generalized algorithm also applies to three-dimensional or higher dimensional data and also to data with no spatial characteristics. The algorithm and software are modified versions of a prior RHSEG algorithm and software, the outputs of which often contain processing-window artifacts including, for example, spurious segmentation-image regions along the boundaries of processing-window edges.

  4. XML-BSPM: an XML format for storing Body Surface Potential Map recordings

    PubMed Central

    2010-01-01

    Background The Body Surface Potential Map (BSPM) is an electrocardiographic method, for recording and displaying the electrical activity of the heart, from a spatial perspective. The BSPM has been deemed more accurate for assessing certain cardiac pathologies when compared to the 12-lead ECG. Nevertheless, the 12-lead ECG remains the most popular ECG acquisition method for non-invasively assessing the electrical activity of the heart. Although data from the 12-lead ECG can be stored and shared using open formats such as SCP-ECG, no open formats currently exist for storing and sharing the BSPM. As a result, an innovative format for storing BSPM datasets has been developed within this study. Methods The XML vocabulary was chosen for implementation, as opposed to binary for the purpose of human readability. There are currently no standards to dictate the number of electrodes and electrode positions for recording a BSPM. In fact, there are at least 11 different BSPM electrode configurations in use today. Therefore, in order to support these BSPM variants, the XML-BSPM format was made versatile. Hence, the format supports the storage of custom torso diagrams using SVG graphics. This diagram can then be used in a 2D coordinate system for retaining electrode positions. Results This XML-BSPM format has been successfully used to store the Kornreich-117 BSPM dataset and the Lux-192 BSPM dataset. The resulting file sizes were in the region of 277 kilobytes for each BSPM recording and can be deemed suitable for example, for use with any telemonitoring application. Moreover, there is potential for file sizes to be further reduced using basic compression algorithms, i.e. the deflate algorithm. Finally, these BSPM files have been parsed and visualised within a convenient time period using a web based BSPM viewer. Conclusions This format, if widely adopted could promote BSPM interoperability, knowledge sharing and data mining. This work could also be used to provide conceptual solutions and inspire existing formats such as DICOM, SCP-ECG and aECG to support the storage of BSPMs. In summary, this research provides initial ground work for creating a complete BSPM management system. PMID:20470392

  5. A recursive algorithm for Zernike polynomials

    NASA Technical Reports Server (NTRS)

    Davenport, J. W.

    1982-01-01

    The analysis of a function defined on a rotationally symmetric system, with either a circular or annular pupil is discussed. In order to numerically analyze such systems it is typical to expand the given function in terms of a class of orthogonal polynomials. Because of their particular properties, the Zernike polynomials are especially suited for numerical calculations. Developed is a recursive algorithm that can be used to generate the Zernike polynomials up to a given order. The algorithm is recursively defined over J where R(J,N) is the Zernike polynomial of degree N obtained by orthogonalizing the sequence R(J), R(J+2), ..., R(J+2N) over (epsilon, 1). The terms in the preceding row - the (J-1) row - up to the N+1 term is needed for generating the (J,N)th term. Thus, the algorith generates an upper left-triangular table. This algorithm was placed in the computer with the necessary support program also included.

  6. Recursion Relations for Double Ramification Hierarchies

    NASA Astrophysics Data System (ADS)

    Buryak, Alexandr; Rossi, Paolo

    2015-12-01

    In this paper we study various properties of the double ramification hierarchy, an integrable hierarchy of hamiltonian PDEs introduced in Buryak (CommunMath Phys 336(3):1085-1107, 2015) using intersection theory of the double ramification cycle in the moduli space of stable curves. In particular, we prove a recursion formula that recovers the full hierarchy starting from just one of the Hamiltonians, the one associated to the first descendant of the unit of a cohomological field theory. Moreover, we introduce analogues of the topological recursion relations and the divisor equation both for the Hamiltonian densities and for the string solution of the double ramification hierarchy. This machinery is very efficient and we apply it to various computations for the trivial and Hodge cohomological field theories, and for the r -spin Witten's classes. Moreover, we prove the Miura equivalence between the double ramification hierarchy and the Dubrovin-Zhang hierarchy for the Gromov-Witten theory of the complex projective line (extended Toda hierarchy).

  7. Recursive delay calculation unit for parametric beamformer

    NASA Astrophysics Data System (ADS)

    Nikolov, Svetoslav I.; Jensen, Jørgen A.; Tomov, Borislav

    2006-03-01

    This paper presents a recursive approach for parametric delay calculations for a beamformer. The suggested calculation procedure is capable of calculating the delays for any image line defined by an origin and arbitrary direction. It involves only add and shift operations making it suitable for hardware implementation. One delaycalculation unit (DCU) needs 4 parameters, and all operations can be implemented using fixed-point arithmetics. An N-channel system needs N+ 1 DCUs per line - one for the distance from the transmit origin to the image point and N for the distances from the image point to each of the receivers. Each DCU recursively calculates the square of the distance between a transducer element and a point on the beamformed line. Then it finds the approximate square root. The distance to point i is used as an initial guess for point i + 1. Using fixed-point calculations with 36-bit precision gives an error in the delay calculations on the order of 1/64 samples, at a sampling frequency of f s = 40 MHz. The circuit has been synthesized for a Virtex II Pro device speed grade 6 in two versions - a pipelined and a non-pipelined producing 150 and 30 million delays per second, respectively. The non-pipelined circuit occupies about 0.5 % of the FPGA resources and the pipelined one about 1 %. When the square root is found with a pipelined CORDIC processor, 2 % of the FPGA slices are used to deliver 150 million delays per second.

  8. Recursion Relations for Double Ramification Hierarchies

    NASA Astrophysics Data System (ADS)

    Buryak, Alexandr; Rossi, Paolo

    2016-03-01

    In this paper we study various properties of the double ramification hierarchy, an integrable hierarchy of hamiltonian PDEs introduced in Buryak (CommunMath Phys 336(3):1085-1107, 2015) using intersection theory of the double ramification cycle in the moduli space of stable curves. In particular, we prove a recursion formula that recovers the full hierarchy starting from just one of the Hamiltonians, the one associated to the first descendant of the unit of a cohomological field theory. Moreover, we introduce analogues of the topological recursion relations and the divisor equation both for the Hamiltonian densities and for the string solution of the double ramification hierarchy. This machinery is very efficient and we apply it to various computations for the trivial and Hodge cohomological field theories, and for the r -spin Witten's classes. Moreover, we prove the Miura equivalence between the double ramification hierarchy and the Dubrovin-Zhang hierarchy for the Gromov-Witten theory of the complex projective line (extended Toda hierarchy).

  9. Improvement in Recursive Hierarchical Segmentation of Data

    NASA Technical Reports Server (NTRS)

    Tilton, James C.

    2006-01-01

    A further modification has been made in the algorithm and implementing software reported in Modified Recursive Hierarchical Segmentation of Data (GSC- 14681-1), NASA Tech Briefs, Vol. 30, No. 6 (June 2006), page 51. That software performs recursive hierarchical segmentation of data having spatial characteristics (e.g., spectral-image data). The output of a prior version of the software contained artifacts, including spurious segmentation-image regions bounded by processing-window edges. The modification for suppressing the artifacts, mentioned in the cited article, was addition of a subroutine that analyzes data in the vicinities of seams to find pairs of regions that tend to lie adjacent to each other on opposite sides of the seams. Within each such pair, pixels in one region that are more similar to pixels in the other region are reassigned to the other region. The present modification provides for a parameter ranging from 0 to 1 for controlling the relative priority of merges between spatially adjacent and spatially non-adjacent regions. At 1, spatially-adjacent-/spatially- non-adjacent-region merges have equal priority. At 0, only spatially-adjacent-region merges (no spectral clustering) are allowed. Between 0 and 1, spatially-adjacent- region merges have priority over spatially- non-adjacent ones.

  10. Realization Of Algebraic Processor For XML Documents Processing

    SciTech Connect

    Georgiev, Bozhidar; Georgieva, Adriana

    2010-10-25

    In this paper, are presented some possibilities concerning the implementation of an algebraic method for XML hierarchical data processing which makes faster the XML search mechanism. Here is offered a different point of view for creation of advanced algebraic processor (with all necessary software tools and programming modules respectively). Therefore, this nontraditional approach for fast XML navigation with the presented algebraic processor may help to build an easier user-friendly interface provided XML transformations, which can avoid the difficulties in the complicated language constructions of XSL, XSLT and XPath. This approach allows comparatively simple search of XML hierarchical data by means of the following types of functions: specification functions and so named build-in functions. The choice of programming language Java may appear strange at first, but it isn't when you consider that the applications can run on different kinds of computers. The specific search mechanism based on the linear algebra theory is faster in comparison with MSXML parsers (on the basis of the developed examples with about 30%). Actually, there exists the possibility for creating new software tools based on the linear algebra theory, which cover the whole navigation and search techniques characterizing XSLT/XPath. The proposed method is able to replace more complicated operations in other SOA components.

  11. The SGML Standardization Framework and the Introduction of XML

    PubMed Central

    Grütter, Rolf

    2000-01-01

    Extensible Markup Language (XML) is on its way to becoming a global standard for the representation, exchange, and presentation of information on the World Wide Web (WWW). More than that, XML is creating a standardization framework, in terms of an open network of meta-standards and mediators that allows for the definition of further conventions and agreements in specific business domains. Such an approach is particularly needed in the healthcare domain; XML promises to especially suit the particularities of patient records and their lifelong storage, retrieval, and exchange. At a time when change rather than steadiness is becoming the faithful feature of our society, standardization frameworks which support a diversified growth of specifications that are appropriate to the actual needs of the users are becoming more and more important; and efforts should be made to encourage this new attempt at standardization to grow in a fruitful direction. Thus, the introduction of XML reflects a standardization process which is neither exclusively based on an acknowledged standardization authority, nor a pure market standard. Instead, a consortium of companies, academic institutions, and public bodies has agreed on a common recommendation based on an existing standardization framework. The consortium's process of agreeing to a standardization framework will doubtlessly be successful in the case of XML, and it is suggested that it should be considered as a generic model for standardization processes in the future. PMID:11720931

  12. The SGML standardization framework and the introduction of XML.

    PubMed

    Fierz, W; Grtter, R

    2000-01-01

    Extensible Markup Language (XML) is on its way to becoming a global standard for the representation, exchange, and presentation of information on the World Wide Web (WWW). More than that, XML is creating a standardization framework, in terms of an open network of meta-standards and mediators that allows for the definition of further conventions and agreements in specific business domains. Such an approach is particularly needed in the healthcare domain; XML promises to especially suit the particularities of patient records and their lifelong storage, retrieval, and exchange. At a time when change rather than steadiness is becoming the faithful feature of our society, standardization frameworks which support a diversified growth of specifications that are appropriate to the actual needs of the users are becoming more and more important; and efforts should be made to encourage this new attempt at standardization to grow in a fruitful direction. Thus, the introduction of XML reflects a standardization process which is neither exclusively based on an acknowledged standardization authority, nor a pure market standard. Instead, a consortium of companies, academic institutions, and public bodies has agreed on a common recommendation based on an existing standardization framework. The consortium's process of agreeing to a standardization framework will doubtlessly be successful in the case of XML, and it is suggested that it should be considered as a generic model for standardization processes in the future. PMID:11720931

  13. Language, Mind, Practice: Families of Recursive Thinking in Human Reasoning

    ERIC Educational Resources Information Center

    Josephson, Marika

    2011-01-01

    In 2002, Chomsky, Hauser, and Fitch asserted that recursion may be the one aspect of the human language faculty that makes human language unique in the narrow sense--unique to language and unique to human beings. They also argue somewhat more quietly (as do Pinker and Jackendoff 2005) that recursion may be possible outside of language: navigation,…

  14. Recursive Frame Analysis: A Practitioner's Tool for Mapping Therapeutic Conversation

    ERIC Educational Resources Information Center

    Keeney, Hillary; Keeney, Bradford; Chenail, Ronald J.

    2012-01-01

    Recursive frame analysis (RFA), both a practical therapeutic tool and an advanced qualitative research method that maps the structure of therapeutic conversation, is introduced with a clinical case vignette. We present and illustrate a means of mapping metaphorical themes that contextualize the performance taking place in the room, recursively

  15. The Real Performance Drivers behind XML Lock Protocols

    NASA Astrophysics Data System (ADS)

    Bächle, Sebastian; Härder, Theo

    Fine-grained lock protocols should allow for highly concurrent transaction processing on XML document trees, which is addressed by the taDOM lock protocol family enabling specific lock modes and lock granules adjusted to the various XML processing models. We have already proved its operational flexibility and performance superiority when compared to competitor protocols. Here, we outline our experiences gained during the implementation and optimization of these protocols. We figure out their performance drivers to maximize throughput while keeping the response times at an acceptable level and perfectly exploiting the advantages of our tailor-made lock protocols for XML trees. Because we have implemented all options and alternatives in our prototype system XTC, benchmark runs for all “drivers” allow for comparisons in identical environments and illustrate the benefit of all implementation decisions. Finally, they reveal that careful lock protocol optimization pays off.

  16. Development Life Cycle and Tools for XML Content Models

    SciTech Connect

    Kulvatunyou, Boonserm; Morris, Katherine; Buhwan, Jeong; Goyal, Puja

    2004-11-01

    Many integration projects today rely on shared semantic models based on standards represented using Extensible Mark up Language (XML) technologies. Shared semantic models typically evolve and require maintenance. In addition, to promote interoperability and reduce integration costs, the shared semantics should be reused as much as possible. Semantic components must be consistent and valid in terms of agreed upon standards and guidelines. In this paper, we describe an activity model for creation, use, and maintenance of a shared semantic model that is coherent and supports efficient enterprise integration. We then use this activity model to frame our research and the development of tools to support those activities. We provide overviews of these tools primarily in the context of the W3C XML Schema. At the present, we focus our work on the W3C XML Schema as the representation of choice, due to its extensive adoption by industry.

  17. An XML-based method for astronomy software designing

    NASA Astrophysics Data System (ADS)

    Liao, Mingxue; Aili, Yusupu; Zhang, Jin

    XML-based method for standardization of software designing is introduced and analyzed and successfully applied to renovating the hardware and software of the digital clock at Urumqi Astronomical Station. Basic strategy for eliciting time information from the new digital clock of FT206 in the antenna control program is introduced. By FT206, the need to compute how many centuries passed since a certain day with sophisticated formulas is eliminated and it is no longer necessary to set right UT time for the computer holding control over antenna because the information about year, month, day are all deduced from Julian day dwelling in FT206, rather than from computer time. With XML-based method and standard for software designing, various existing designing methods are unified, communications and collaborations between developers are facilitated, and thus Internet-based mode of developing software becomes possible. The trend of development of XML-based designing method is predicted.

  18. Utilizing the Structure and Content Information for XML Document Clustering

    NASA Astrophysics Data System (ADS)

    Tran, Tien; Kutty, Sangeetha; Nayak, Richi

    This paper reports on the experiments and results of a clustering approach used in the INEX 2008 document mining challenge. The clustering approach utilizes both the structure and content information of the Wikipedia XML document collection. A latent semantic kernel (LSK) is used to measure the semantic similarity between XML documents based on their content features. The construction of a latent semantic kernel involves the computing of singular vector decomposition (SVD). On a large feature space matrix, the computation of SVD is very expensive in terms of time and memory requirements. Thus in this clustering approach, the dimension of the document space of a term-document matrix is reduced before performing SVD. The document space reduction is based on the common structural information of the Wikipedia XML document collection. The proposed clustering approach has shown to be effective on the Wikipedia collection in the INEX 2008 document mining challenge.

  19. Research of home networking system based on XML/BACnet

    NASA Astrophysics Data System (ADS)

    Wang, Zhongming

    2008-11-01

    To standardize home networking information and simplify its management, this paper form a universal information module of various devices in home networking by adopting XML technology and BACnet protocol(XML/BACnet). Then, a software architecture of home networking based on this module is designed, having the function like auto management and maintenance, safety, real-time and remote controlling. Consequently, a home networking system based on this architecture is completed. Tested and evaluated, this system is one easy-using, easy-realizing, nice real-time system with strong heterogeneity and stable safety system.

  20. XML in an Adaptive Framework for Instrument Control

    NASA Technical Reports Server (NTRS)

    Ames, Troy J.

    2004-01-01

    NASA Goddard Space Flight Center is developing an extensible framework for instrument command and control, known as Instrument Remote Control (IRC), that combines the platform independent processing capabilities of Java with the power of the Extensible Markup Language (XML). A key aspect of the architecture is software that is driven by an instrument description, written using the Instrument Markup Language (IML). IML is an XML dialect used to describe interfaces to control and monitor the instrument, command sets and command formats, data streams, communication mechanisms, and data processing algorithms.

  1. Bayesian recursive data-pattern tomography

    NASA Astrophysics Data System (ADS)

    Mikhalychev, Alexander; Mogilevtsev, Dmitri; Teo, Yong Siah; Řeháček, Jaroslav; Hradil, Zdeněk

    2015-11-01

    We present a simple and efficient Bayesian recursive algorithm for the data-pattern scheme for quantum state reconstruction, which is applicable to situations where measurement settings can be controllably varied efficiently. The algorithm predicts the best measurements required to accurately reconstruct the unknown signal state in terms of a fixed set of probe states. In each iterative step, this algorithm seeks the measurement setting that minimizes the variance of the data-pattern estimator, which essentially measures the reconstruction accuracy, with the help of a data-pattern bank that was acquired prior to the signal reconstruction. We show that, with this algorithm, it is possible to minimize the number of measurement settings required to obtain a reasonably accurate state estimator by using just the optimal settings and, at the same time, increasing the numerical efficiency of the data-pattern reconstruction.

  2. XML does Real Programmers a Service

    SciTech Connect

    Gorton, Ian

    2008-09-01

    As the sun slowly sets of this first decade of the new millenium, it seems appropriate to update the sojourn of the real programmers as they adapt to their ever changing technical and business environment. Real Programmers were perfectly characterized and differentiated from their quiche-eating, Pascal programming brethren in Ed Post’s seminal “Real Programmers Don’t Use Pascal” (Datamation, 1983). My follow-up ("Real programmers do use Delphi," Software, IEEE , vol.12, no.6, pp.8, 10, 12-, Nov 1995) charted their evolution from FORTRAN-only programmers to embracing a wider range of mainstream languages and tools that still afforded ample opportunity for creativity, game-playing, irregular work hours, and importantly, long-term job security.

  3. Applying Analogical Reasoning Techniques for Teaching XML Document Querying Skills in Database Classes

    ERIC Educational Resources Information Center

    Mitri, Michel

    2012-01-01

    XML has become the most ubiquitous format for exchange of data between applications running on the Internet. Most Web Services provide their information to clients in the form of XML. The ability to process complex XML documents in order to extract relevant information is becoming as important a skill for IS students to master as querying…

  4. Applying Analogical Reasoning Techniques for Teaching XML Document Querying Skills in Database Classes

    ERIC Educational Resources Information Center

    Mitri, Michel

    2012-01-01

    XML has become the most ubiquitous format for exchange of data between applications running on the Internet. Most Web Services provide their information to clients in the form of XML. The ability to process complex XML documents in order to extract relevant information is becoming as important a skill for IS students to master as querying

  5. New NED XML/VOtable Services and Client Interface Applications

    NASA Astrophysics Data System (ADS)

    Pevunova, O.; Good, J.; Mazzarella, J.; Berriman, G. B.; Madore, B.

    2005-12-01

    The NASA/IPAC Extragalactic Database (NED) provides data and cross-identifications for over 7 million extragalactic objects fused from thousands of survey catalogs and journal articles. The data cover all frequencies from radio through gamma rays and include positions, redshifts, photometry and spectral energy distributions (SEDs), sizes, and images. NED services have traditionally supplied data in HTML format for connections from Web browsers, and a custom ASCII data structure for connections by remote computer programs written in the C programming language. We describe new services that provide responses from NED queries in XML documents compliant with the international virtual observatory VOtable protocol. The XML/VOtable services support cone searches, all-sky searches based on object attributes (survey names, cross-IDs, redshifts, flux densities), and requests for detailed object data. Initial services have been inserted into the NVO registry, and others will follow soon. The first client application is a Style Sheet specification for rendering NED VOtable query results in Web browsers that support XML. The second prototype application is a Java applet that allows users to compare multiple SEDs. The new XML/VOtable output mode will also simplify the integration of data from NED into visualization and analysis packages, software agents, and other virtual observatory applications. We show an example SED from NED plotted using VOPlot. The NED website is: http://nedwww.ipac.caltech.edu.

  6. Personalization of XML Content Browsing Based on User Preferences

    ERIC Educational Resources Information Center

    Encelle, Benoit; Baptiste-Jessel, Nadine; Sedes, Florence

    2009-01-01

    Personalization of user interfaces for browsing content is a key concept to ensure content accessibility. In this direction, we introduce concepts that result in the generation of personalized multimodal user interfaces for browsing XML content. User requirements concerning the browsing of a specific content type can be specified by means of

  7. Personalization of XML Content Browsing Based on User Preferences

    ERIC Educational Resources Information Center

    Encelle, Benoit; Baptiste-Jessel, Nadine; Sedes, Florence

    2009-01-01

    Personalization of user interfaces for browsing content is a key concept to ensure content accessibility. In this direction, we introduce concepts that result in the generation of personalized multimodal user interfaces for browsing XML content. User requirements concerning the browsing of a specific content type can be specified by means of…

  8. Chapter 57: Advanced XML Technologies: Schema, Xpath, XQuery, and XSL

    NASA Astrophysics Data System (ADS)

    Plante, R. L.

    Much of what happens behind the scenes in the VO happens using XML A major reason for choosing to use XML is to take advantage of "off-the-shelf" standards and technologies that can help us manage our metadata. Most VO users will not need to know anything about various manipulations of XML going on underneath their applications. However, users who begin to delve into programming for the VO, be it scripting to gather data for a VO research project or developing a general application, will start to see some of these technologies at work under the hood. In four sections in this chapter, we'll look at four of the most useful XML technologies. With each one, we'll start by highlighting how you might find it useful. The intent is not to make you proficient in these tools. Rather, by getting a general sense how these technologies work you will at least have some ability to debug your application when things go wrong. In some cases, you may acquire enough familiarity to edit and use pre-existing samples. In this chapter, you will get a chance to make some of those edits and try out some tools you can find on the CD. We will use some tools from the adqllib package. If you tried out the exercises in Chapter 36, then you may have already built these tools. If not, you can do that now

  9. Recursive estimation of prior probabilities using the mixture approach

    NASA Technical Reports Server (NTRS)

    Kazakos, D.

    1974-01-01

    The problem of estimating the prior probabilities q sub k of a mixture of known density functions f sub k(X), based on a sequence of N statistically independent observations is considered. It is shown that for very mild restrictions on f sub k(X), the maximum likelihood estimate of Q is asymptotically efficient. A recursive algorithm for estimating Q is proposed, analyzed, and optimized. For the M = 2 case, it is possible for the recursive algorithm to achieve the same performance with the maximum likelihood one. For M 2, slightly inferior performance is the price for having a recursive algorithm. However, the loss is computable and tolerable.

  10. Progress on an implementation of MIFlowCyt in XML

    NASA Astrophysics Data System (ADS)

    Leif, Robert C.; Leif, Stephanie H.

    2015-03-01

    Introduction: The International Society for Advancement of Cytometry (ISAC) Data Standards Task Force (DSTF) has created a standard for the Minimum Information about a Flow Cytometry Experiment (MIFlowCyt 1.0). The CytometryML schemas, are based in part upon the Flow Cytometry Standard and Digital Imaging and Communication (DICOM) standards. CytometryML has and will be extended and adapted to include MIFlowCyt, as well as to serve as a common standard for flow and image cytometry (digital microscopy). Methods: The MIFlowCyt data-types were created, as is the rest of CytometryML, in the XML Schema Definition Language (XSD1.1). Individual major elements of the MIFlowCyt schema were translated into XML and filled with reasonable data. A small section of the code was formatted with HTML formatting elements. Results: The differences in the amount of detail to be recorded for 1) users of standard techniques including data analysts and 2) others, such as method and device creators, laboratory and other managers, engineers, and regulatory specialists required that separate data-types be created to describe the instrument configuration and components. A very substantial part of the MIFlowCyt element that describes the Experimental Overview part of the MIFlowCyt and substantial parts of several other major elements have been developed. Conclusions: The future use of structured XML tags and web technology should facilitate searching of experimental information, its presentation, and inclusion in structured research, clinical, and regulatory documents, as well as demonstrate in publications adherence to the MIFlowCyt standard. The use of CytometryML together with XML technology should also result in the textual and numeric data being published using web technology without any change in composition. Preliminary testing indicates that CytometryML XML pages can be directly formatted with the combination of HTML and CSS.

  11. Statistical Error Analysis for Digital Recursive Filters

    NASA Astrophysics Data System (ADS)

    Wu, Kevin Chi-Rung

    The study of arithmetic roundoff error has attracted many researchers to investigate how the signal-to-noise ratio (SNR) is affected by algorithmic parameters, especially since the VLSI (Very Large Scale Integrated circuits) technologies have become more promising for digital signal processing. Typically, digital signal processing involving, either with or without matrix inversion, will have tradeoffs on speed and processor cost. Hence, the problems of an area-time efficient matrix computation and roundoff error behavior analysis will play an important role in this dissertation. A newly developed non-Cholesky square-root matrix will be discussed which precludes the arithmetic roundoff error over some interesting operations, such as complex -valued matrix inversion with its SNR analysis and error propagation effects. A non-CORDIC parallelism approach for complex-valued matrix will be presented to upgrade speed at the cost of moderate increase of processor. The lattice filter will also be looked into, in such a way, that one can understand the SNR behavior under the conditions of different inputs in the joint process system. Pipelining technique will be demonstrated to manifest the possibility of high-speed non-matrix-inversion lattice filter. Floating point arithmetic modelings used in this study have been focused on effective methodologies that have been proved to be reliable and feasible. With the models in hand, we study the roundoff error behavior based on some statistical assumptions. Results are demonstrated by carrying out simulation to show the feasibility of SNR analysis. We will observe that non-Cholesky square-root matrix has advantage of saving a time of O(n^3) as well as a reduced realization cost. It will be apparent that for a Kalman filter the register size is increasing significantly, if pole of the system matrix is moving closer to the edge of the unit circle. By comparing roundoff error effect due to floating-point and fixed-point arithmetics, we can see that floating-point arithmetic is less sensitive to maximum/minimum value of a matrix. Implementation of parallelism on complex-valued matrix can be justified in the sense of simpler hardware and higher speed. It will be illustrated that the more correlated inputs are present in the lattice filter, the SNR of the order-recursive filter will be degraded faster. Pipelining technique will be proved to be effective in speed upgradation, especially for the order-recursive lattice filter whose order is usually very high.

  12. Recursive Construction of Operator Product Expansion Coefficients

    NASA Astrophysics Data System (ADS)

    Holland, Jan; Hollands, Stefan

    2015-06-01

    We derive a novel formula for the derivative of operator product expansion (OPE) coefficients with respect to a coupling constant. The formula involves just the OPE coefficients themselves but no further input, and is in this sense self-consistent. Furthermore, unlike other formal identities of this general nature in quantum field theory (such as the formal expression for the Lagrangian perturbation of a correlation function), our formula requires no further UV-renormalization, i.e., it is completely well-defined from the start. This feature is a result of a cancelation of UV- and IR-divergences between various terms in our identity. Our proof, and an analysis of the features of the identity, is given for the example of massive, Euclidean theory in 4 dimensional Euclidean space. It relies on the renormalization group flow equation method and is valid to arbitrary, but finite orders in perturbation theory. The final formula, however, makes neither explicit reference to the renormalization group flow, nor to perturbation theory, and we conjecture that it also holds non-perturbatively. Our identity can be applied constructively because it gives a novel recursive algorithm for the computation of OPE coefficients to arbitrary (finite) perturbation order in terms of the zeroth order coefficients corresponding to the underlying free field theory, which in turn are trivial to obtain. We briefly illustrate the relation of this method to more standard methods for computing the OPE in some simple examples.

  13. Recursive frame integration of limited data: RAFAIL

    NASA Astrophysics Data System (ADS)

    Rafailov, Michael K.; Soli, Robert A.

    2005-08-01

    Real time infrared imaging and tracking usually requires a high probability of target detection along with a low false alarm rate, achievable only with a high "Signal-to-Noise Ratio" (SNR). Frame integration--summing of non-correlated frames--is commonly used to improve the SNR. But conventional frame integration requires significant processing to store full frames and integrate intermediate results, normalize frame data, etc. It may drive acquisition of highly specialized hardware, faster processors, dedicated frame integration circuit cards and extra memory cards. Non-stationary noise, low frequency noise correlation, non-ergodic noise, scene dynamics, or pointing accuracy may also limit performance. Recursive frame integration of limited data--RAFAIL, is proposed as a means to improve frame integration performance and mitigate the issues. The technique applies two thresholds--one tuned for optimum probability of detection, the other to manage required false alarm rate--and allows a non-linear integration process that, along with optimal noise management, provides system designers more capability where cost, weight, or power considerations limit system data rate, processing, or memory capability.

  14. On-Shell Recursion Relations for Effective Field Theories

    NASA Astrophysics Data System (ADS)

    Cheung, Clifford; Kampf, Karol; Novotny, Jiri; Shen, Chia-Hsien; Trnka, Jaroslav

    2016-01-01

    We derive the first ever on-shell recursion relations applicable to effective field theories. Based solely on factorization and the soft behavior of amplitudes, these recursion relations employ a new rescaling momentum shift to construct all tree-level scattering amplitudes in the nonlinear sigma model, Dirac-Born-Infeld theory, and the Galileon. Our results prove that all theories with enhanced soft behavior are on-shell constructible.

  15. Recursive graphs with small-world scale-free properties

    NASA Astrophysics Data System (ADS)

    Comellas, Francesc; Fertin, Guillaume; Raspaud, André

    2004-03-01

    We discuss a category of graphs, recursive clique trees, which have small-world and scale-free properties and allow a fine tuning of the clustering and the power-law exponent of their discrete degree distribution. We determine relevant characteristics of those graphs: the diameter, degree distribution, and clustering parameter. The graphs have also an interesting recursive property, and generalize recent constructions with fixed degree distributions.

  16. XML-based MPEG-4 Video Representation, Streaming and Error Resilience

    NASA Astrophysics Data System (ADS)

    Sun, Xiaoming; Shi, Zhi; Kuo, C.-C. Jay

    2002-07-01

    A new XML-based MPEG-4 coding system for video streaming is proposed in this research. The XML technology is applied to MPEG-4 coded video contents to provide more flexibility for video manipulation. To be more specific, for every compressed video file, we generate the corresponding XML document that works as an indexing file as well as a streaming facilitator. This XML-based MPEG-4 coded bitstream can be transmitted over the high level HTTP protocol or the low level IP or RTP protocol. In this work, we will examine the design of this XML-based streaming system, including the overhead of the XML document, its transmission, the processing requirement, and error resilience, etc. Experimental results will be provided to compare the performances of the traditional MPEG-4 streaming solution and the proposed XML-based solution.

  17. XML: James Webb Space Telescope Database Issues, Lessons, and Status

    NASA Technical Reports Server (NTRS)

    Detter, Ryan; Mooney, Michael; Fatig, Curtis

    2003-01-01

    This paper will present the current concept using extensible Markup Language (XML) as the underlying structure for the James Webb Space Telescope (JWST) database. The purpose of using XML is to provide a JWST database, independent of any portion of the ground system, yet still compatible with the various systems using a variety of different structures. The testing of the JWST Flight Software (FSW) started in 2002, yet the launch is scheduled for 2011 with a planned 5-year mission and a 5-year follow on option. The initial database and ground system elements, including the commands, telemetry, and ground system tools will be used for 19 years, plus post mission activities. During the Integration and Test (I&T) phases of the JWST development, 24 distinct laboratories, each geographically dispersed, will have local database tools with an XML database. Each of these laboratories database tools will be used for the exporting and importing of data both locally and to a central database system, inputting data to the database certification process, and providing various reports. A centralized certified database repository will be maintained by the Space Telescope Science Institute (STScI), in Baltimore, Maryland, USA. One of the challenges for the database is to be flexible enough to allow for the upgrade, addition or changing of individual items without effecting the entire ground system. Also, using XML should allow for the altering of the import and export formats needed by the various elements, tracking the verification/validation of each database item, allow many organizations to provide database inputs, and the merging of the many existing database processes into one central database structure throughout the JWST program. Many National Aeronautics and Space Administration (NASA) projects have attempted to take advantage of open source and commercial technology. Often this causes a greater reliance on the use of Commercial-Off-The-Shelf (COTS), which is often limiting. In our review of the database requirements and the COTS software available, only very expensive COTS software will meet 90% of requirements. Even with the high projected initial cost of COTS, the development and support for custom code over the 19-year mission period was forecasted to be higher than the total licensing costs. A group did look at reusing existing database tools and formats. If the JWST database was already in a mature state, the reuse made sense, but with the database still needing to handing the addition of different types of command and telemetry structures, defining new spacecraft systems, accept input and export to systems which has not been defined yet, XML provided the flexibility desired. It remains to be determined whether the XML database will reduce the over all cost for the JWST mission.

  18. Recursion to food plants by free-ranging Bornean elephant

    PubMed Central

    Gillespie, Graeme; Goossens, Benoit; Ismail, Sulaiman; Ancrenaz, Marc; Linklater, Wayne

    2015-01-01

    Plant recovery rates after herbivory are thought to be a key factor driving recursion by herbivores to sites and plants to optimise resource-use but have not been investigated as an explanation for recursion in large herbivores. We investigated the relationship between plant recovery and recursion by elephants (Elephas maximus borneensis) in the Lower Kinabatangan Wildlife Sanctuary, Sabah. We identified 182 recently eaten food plants, from 30 species, along 14 × 50 m transects and measured their recovery growth each month over nine months or until they were re-browsed by elephants. The monthly growth in leaf and branch or shoot length for each plant was used to calculate the time required (months) for each species to recover to its pre-eaten length. Elephant returned to all but two transects with 10 eaten plants, a further 26 plants died leaving 146 plants that could be re-eaten. Recursion occurred to 58% of all plants and 12 of the 30 species. Seventy-seven percent of the re-eaten plants were grasses. Recovery times to all plants varied from two to twenty months depending on the species. Recursion to all grasses coincided with plant recovery whereas recursion to most browsed plants occurred four to twelve months before they had recovered to their previous length. The small sample size of many browsed plants that received recursion and uneven plant species distribution across transects limits our ability to generalise for most browsed species but a prominent pattern in plant-scale recursion did emerge. Plant recovery time was a good predictor of time to recursion but varied as a function of growth form (grass, ginger, palm, liana and woody) and differences between sites. Time to plant recursion coincided with plant recovery time for the elephant’s preferred food, grasses, and perhaps also gingers, but not the other browsed species. Elephants are bulk feeders so it is likely that they time their returns to bulk feed on these grass species when quantities have recovered sufficiently to meet their intake requirements. The implications for habitat and elephant management are discussed. PMID:26290779

  19. voevent-parse: Parse, manipulate, and generate VOEvent XML packets

    NASA Astrophysics Data System (ADS)

    Staley, Tim D.

    2014-11-01

    voevent-parse, written in Python, parses, manipulates, and generates VOEvent XML packets; it is built atop lxml.objectify. Details of transients detected by many projects, including Fermi, Swift, and the Catalina Sky Survey, are currently made available as VOEvents, which is also the standard alert format by future facilities such as LSST and SKA. However, working with XML and adhering to the sometimes lengthy VOEvent schema can be a tricky process. voevent-parse provides convenience routines for common tasks, while allowing the user to utilise the full power of the lxml library when required. An earlier version of voevent-parse was part of the pysovo (ascl:1411.002) library.

  20. A proposed referral centre based on HL7/XML.

    PubMed

    Chen, T S; Liao, B S; Lee, C H; Gough, T G

    2002-01-01

    With the growth of the Inteernet, hospitals have also applied HL7 (Health Level Seven) to exchange data between them. The referral system is identified as an appropriate application system. The effect of referral is to transfer the patient to a suitable hospital in a timely fashion, and to arrange appropriate treatment for the patient. Taking advantage of the Internet to exchange referral data can, not only accelerate the process of patient referral, but also avoid the unnecessary repeat examinations to decrease the waste of medical resources. This article builds up a referral-related message according to the HL7 standard, and develops a referral centre using the Internet environment, making use of XML (eXtensible Markup Language) standard to transform the referral-related data to XML format and exchange referral data between platforms. This electronic referral mechanism is expected to offer other hospitals experience of improved referral practice. PMID:15460680

  1. T:XML: A Tool Supporting User Interface Model Transformation

    NASA Astrophysics Data System (ADS)

    López-Jaquero, Víctor; Montero, Francisco; González, Pascual

    Model driven development of user interfaces is based on the transformation of an abstract specification into the final user interface the user will interact with. The design of transformation rules to carry out this transformation process is a key issue in any model-driven user interface development approach. In this paper, we introduce T:XML, an integrated development environment for managing, creating and previewing transformation rules. The tool supports the specification of transformation rules by using a graphical notation that works on the basis of the transformation of the input model into a graph-based representation. T:XML allows the design and execution of transformation rules in an integrated development environment. Furthermore, the designer can also preview how the generated user interface looks like after the transformations have been applied. These previewing capabilities can be used to quickly create prototypes to discuss with the users in user-centered design methods.

  2. The XBabelPhish MAGE-ML and XML Translator

    PubMed Central

    Maier, Don; Wymore, Farrell; Sherlock, Gavin; Ball, Catherine A

    2008-01-01

    Background MAGE-ML has been promoted as a standard format for describing microarray experiments and the data they produce. Two characteristics of the MAGE-ML format compromise its use as a universal standard: First, MAGE-ML files are exceptionally large – too large to be easily read by most people, and often too large to be read by most software programs. Second, the MAGE-ML standard permits many ways of representing the same information. As a result, different producers of MAGE-ML create different documents describing the same experiment and its data. Recognizing all the variants is an unwieldy software engineering task, resulting in software packages that can read and process MAGE-ML from some, but not all producers. This Tower of MAGE-ML Babel bars the unencumbered exchange of microarray experiment descriptions couched in MAGE-ML. Results We have developed XBabelPhish – an XQuery-based technology for translating one MAGE-ML variant into another. XBabelPhish's use is not restricted to translating MAGE-ML documents. It can transform XML files independent of their DTD, XML schema, or semantic content. Moreover, it is designed to work on very large (> 200 Mb.) files, which are common in the world of MAGE-ML. Conclusion XBabelPhish provides a way to inter-translate MAGE-ML variants for improved interchange of microarray experiment information. More generally, it can be used to transform most XML files, including very large ones that exceed the capacity of most XML tools. PMID:18205924

  3. XSemantic: An Extension of LCA Based XML Semantic Search

    NASA Astrophysics Data System (ADS)

    Supasitthimethee, Umaporn; Shimizu, Toshiyuki; Yoshikawa, Masatoshi; Porkaew, Kriengkrai

    One of the most convenient ways to query XML data is a keyword search because it does not require any knowledge of XML structure or learning a new user interface. However, the keyword search is ambiguous. The users may use different terms to search for the same information. Furthermore, it is difficult for a system to decide which node is likely to be chosen as a return node and how much information should be included in the result. To address these challenges, we propose an XML semantic search based on keywords called XSemantic. On the one hand, we give three definitions to complete in terms of semantics. Firstly, the semantic term expansion, our system is robust from the ambiguous keywords by using the domain ontology. Secondly, to return semantic meaningful answers, we automatically infer the return information from the user queries and take advantage of the shortest path to return meaningful connections between keywords. Thirdly, we present the semantic ranking that reflects the degree of similarity as well as the semantic relationship so that the search results with the higher relevance are presented to the users first. On the other hand, in the LCA and the proximity search approaches, we investigated the problem of information included in the search results. Therefore, we introduce the notion of the Lowest Common Element Ancestor (LCEA) and define our simple rule without any requirement on the schema information such as the DTD or XML Schema. The first experiment indicated that XSemantic not only properly infers the return information but also generates compact meaningful results. Additionally, the benefits of our proposed semantics are demonstrated by the second experiment.

  4. Enhance Reuse of Standard e-Business XML Schema Documents

    SciTech Connect

    Buhwan, Jeong; Kulvatunyou, Boonserm; Ivezic, Nenad; Jones, Albert

    2005-07-01

    Ideally, e-Business application interfaces would be built from highly reusable specifications of business document standards. Since many of these specifications are poorly understood, users often create new ones or customize existing ones every time a new integration problem arises. Consequently, even though there is a potential for reuse, the lack of a component discovery tool means that the cost of reuse is still prohibitively high. In this paper, we explore the potential of using similarity metrics to discover standard XML Schema documents. Our goal is to enhance reuse of XML Schema document/component standards in new integration contexts through the discovery process. We are motivated by the increasing access to the application interface specifications expressed in the form of XML Schema. These specifications are created to facilitate business documents exchange among software applications. Reuse can reduce both the proliferation of standards and the interoperability costs. To demonstrate these potential benefits, we propose and position our research based on an experimental scenario and a novel evaluation approach to qualify alternative similarity metrics on schema discovery. The edge equality in the evaluation method provides a conservative quality measure. We review a number of fundamental approaches to developing similarity metrics, and we organize these metrics into lexical, structural, and logical categories. For each of the metrics, we discuss its relevance and potential issues in its application to the XML Schema discovery task. We conclude that each of the similarity measures has its own strengths and weaknesses and each is expected to yield different results in different search situations. It is important, in the context of an application of these measures to e-Business standards that a schema discovery engine capable of assigning appropriate weights to different similarity measures be used when the search conditions change. This is a subject of our future experimental work.

  5. HDX Data Model: FITS, NDF and XML Implementation

    NASA Astrophysics Data System (ADS)

    Giaretta, D.; Taylor, M.; Draper, P.; Gray, N.; McIlwrath, B.

    A highly adaptable data model, HDX, based on the concepts embodied in FITS and various proposed XML-based formats, as well as Starlink's NDF and HDS will be described, together with the Java software that has been developed to support it. The aim is to provide a flexible model which is compatible with FITS, can be extended to accommodate VO requirements, but which maintains enough mandatory structure to make application-level interoperability relatively easy.

  6. XTCE: XML Telemetry and Command Exchange Tutorial, XTCE Version 1

    NASA Technical Reports Server (NTRS)

    Rice, Kevin; Kizzort, Brad

    2008-01-01

    These presentation slides are a tutorial on XML Telemetry and Command Exchange (XTCE). The goal of XTCE is to provide an industry standard mechanism for describing telemetry and command streams (particularly from satellites.) it wiill lower cost and increase validation over traditional formats, and support exchange or native format.XCTE is designed to describe bit streams, that are typical of telemetry and command in the historic space domain.

  7. Phase Response Design of Recursive All-Pass Digital Filters Using a Modified PSO Algorithm

    PubMed Central

    Chang, Wei-Der

    2015-01-01

    This paper develops a new design scheme for the phase response of an all-pass recursive digital filter. A variant of particle swarm optimization (PSO) algorithm will be utilized for solving this kind of filter design problem. It is here called the modified PSO (MPSO) algorithm in which another adjusting factor is more introduced in the velocity updating formula of the algorithm in order to improve the searching ability. In the proposed method, all of the designed filter coefficients are firstly collected to be a parameter vector and this vector is regarded as a particle of the algorithm. The MPSO with a modified velocity formula will force all particles into moving toward the optimal or near optimal solution by minimizing some defined objective function of the optimization problem. To show the effectiveness of the proposed method, two different kinds of linear phase response design examples are illustrated and the general PSO algorithm is compared as well. The obtained results show that the MPSO is superior to the general PSO for the phase response design of digital recursive all-pass filter. PMID:26366168

  8. An XML Driven Graphical User Interface and Application Management Toolkit

    SciTech Connect

    White, Greg R

    2002-01-18

    In the past, the features of a user interface were limited by those available in the existing graphical widgets it used. Now, improvements in processor speed have fostered the emergence of interpreted languages, in which the appropriate method to render a given data object can be loaded at runtime. XML can be used to precisely describe the association of data types with their graphical handling (beans), and Java provides an especially rich environment for programming the graphics. We present a graphical user interface builder based on Java Beans and XML, in which the graphical screens are described textually (in files or a database) in terms of their screen components. Each component may be a simple text read back, or a complex plot. The programming model provides for dynamic data pertaining to a component to be forwarded synchronously or asynchronously, to the appropriate handler, which may be a built-in method, or a complex applet. This work was initially motivated by the need to move the legacy VMS display interface of the SLAC Control Program to another platform while preserving all of its existing functionality. However the model allows us a powerful and generic system for adding new kinds of graphics, such as Matlab, data sources, such as EPICS, middleware, such as AIDA[1], and transport, such as XML and SOAP. The system will also include a management console, which will be able to report on the present usage of the system, for instance who is running it where and connected to which channels.

  9. Experiments with recursive estimation in astronomical image processing

    NASA Technical Reports Server (NTRS)

    Busko, I.

    1992-01-01

    Recursive estimation concepts were applied to image enhancement problems since the 70's. However, very few applications in the particular area of astronomical image processing are known. These concepts were derived, for 2-dimensional images, from the well-known theory of Kalman filtering in one dimension. The historic reasons for application of these techniques to digital images are related to the images' scanned nature, in which the temporal output of a scanner device can be processed on-line by techniques borrowed directly from 1-dimensional recursive signal analysis. However, recursive estimation has particular properties that make it attractive even in modern days, when big computer memories make the full scanned image available to the processor at any given time. One particularly important aspect is the ability of recursive techniques to deal with non-stationary phenomena, that is, phenomena which have their statistical properties variable in time (or position in a 2-D image). Many image processing methods make underlying stationary assumptions either for the stochastic field being imaged, for the imaging system properties, or both. They will underperform, or even fail, when applied to images that deviate significantly from stationarity. Recursive methods, on the contrary, make it feasible to perform adaptive processing, that is, to process the image by a processor with properties tuned to the image's local statistical properties. Recursive estimation can be used to build estimates of images degraded by such phenomena as noise and blur. We show examples of recursive adaptive processing of astronomical images, using several local statistical properties to drive the adaptive processor, as average signal intensity, signal-to-noise and autocorrelation function. Software was developed under IRAF, and as such will be made available to interested users.

  10. Solving Electronic Structure Problems with the Recursion Method

    NASA Astrophysics Data System (ADS)

    Haydock, Roger

    1994-02-01

    The goal of these lectures is to achieve a working knowledge of how to apply the recursion method to problems in electronic structure. The first lecture introduces the basic issues of the electronic structure of condensed matter and explains how the recursion method addresses them. In a macroscopic system it is only possible to compute local quantities which converge exponentially with computational effort due to a quantum analogue of the black body radiation theorem. The second lecture develops the recursion method from quantum path counting, leading to the chain model and its physical interpretation. The third lecture is about the calculation of electronic energies, bounds on them, and differences in total energies. The last lecture is on the recursion method in finite precision, ghosts and how to eliminate their consequences. In summary, the contents of the lectures may be described as: 1. Local environment and the quantum black body theorem, 2. Quantum path counting and The Recursion Method, 3. Orbital peeling-microscopic differences of macroscopic energies, 4. Finite precision and ghosts.

  11. Knot Invariants from Topological Recursion on Augmentation Varieties

    NASA Astrophysics Data System (ADS)

    Gu, Jie; Jockers, Hans; Klemm, Albrecht; Soroush, Masoud

    2015-06-01

    Using the duality between Wilson loop expectation values of SU( N) Chern-Simons theory on S 3 and topological open-string amplitudes on the local mirror of the resolved conifold, we study knots on S 3 and their invariants encoded in colored HOMFLY polynomials by means of topological recursion. In the context of the local mirror Calabi-Yau threefold of the resolved conifold, we generalize the topological recursion of the remodelled B-model in order to study branes beyond the class of toric Harvey-Lawson special Lagrangians—as required for analyzing non-trivial knots on S 3. The basic ingredients for the proposed recursion are the spectral curve, given by the augmentation variety of the knot, and the calibrated annulus kernel, encoding the topological annulus amplitudes associated to the knot. We present an explicit construction of the calibrated annulus kernel for torus knots and demonstrate the validity of the topological recursion. We further argue that—if an explicit form of the calibrated annulus kernel is provided for any other knot—the proposed topological recursion should still be applicable. We study the implications of our proposal for knot theory, which exhibit interesting consequences for colored HOMFLY polynomials of mutant knots.

  12. Robust recursive impedance estimation for automotive lithium-ion batteries

    NASA Astrophysics Data System (ADS)

    Fridholm, Björn; Wik, Torsten; Nilsson, Magnus

    2016-02-01

    Recursive algorithms, such as recursive least squares (RLS) or Kalman filters, are commonly used in battery management systems to estimate the electrical impedance of the battery cell. However, these algorithms can in some cases run into problems with bias and even divergence of the estimates. This article illuminates problems that can arise in the online estimation using recursive methods, and lists modifications to handle these issues. An algorithm is also proposed that estimates the impedance by separating the problem in two parts; one estimating the ohmic resistance with an RLS approach, and another one where the dynamic effects are estimated using an adaptive Kalman filter (AKF) that is novel in the battery field. The algorithm produces robust estimates of ohmic resistance and time constant of the battery cell in closed loop with SoC estimation, as demonstrated by both in simulations and with experimental data from a lithium-ion battery cell.

  13. Theory and applications of sampled analog recursive CTD comb filters

    NASA Technical Reports Server (NTRS)

    Tao, T. F.; Holmes, S. V.; Ejaz, A.; Piazza, F.; Saetre, L. T.; Freund, B.

    1976-01-01

    Sampled analog comb filter using a recursive filter implementation was studied. Charge transfer device delay lines were used as the delay elements. The similarities and differences between sampled analog recursive filter and digital recursive filter were pointed out. Both canceller type and integrator type comb filter using the first order or the second order canonical circuits were studied. Experimental results of frequency response are presented to show their difference from that of digital filters. A theoretical analysis was presented based essentially on the digital filter theory but modified to take into account the differences described above. The usefulness of this theory in analyzing the comb filter characteristics was discussed. Several applications of both the canceller type comb filter and the integrator type comb filter were demonstrated.

  14. The limits on combining recursive horn rules with description logics

    SciTech Connect

    Levy, A.Y.; Rousset, M.C.

    1996-12-31

    Horn rule languages have formed the basis for many Artificial Intelligence application languages, but are not expressive enough to model domains with a rich hierarchical structure. Description logics have been designed especially to model rich hierarchies. Several applications would significantly benefit from combining the expressive power of both formalisms. This paper focuses on combining recursive function-free Horn rules with the expressive description logic ALCNR, and shows exactly when a hybrid language with decidable inference can be obtained. First, we show that several of the core constructors of description logics lead by themselves to undecidability of inference when combined with recursive function-free Horn rules. We then show that without these constructors we obtain a maximal subset of ALCNRR that yields a decidable hybrid language. Finally, we describe a restriction on the Horn rules that guarantees decidable inference when combined with all of ALCNR, and covers many of the common usages of recursive rules.

  15. Elucidating the stop bands of structurally colored systems through recursion

    NASA Astrophysics Data System (ADS)

    Amir, Ariel; Vukusic, Peter

    2013-04-01

    Interference is the source of some of the spectacular colors of animals and plants in nature. In some of these systems, the physical structure consists of an ordered array of layers with alternating high and low refractive indices. This periodicity leads to an optical band structure that is analogous to the electronic band structure encountered in semiconductor physics: specific bands of wavelengths (the stop bands) are perfectly reflected. Here, we present a minimal model for optical band structure in a periodic multilayer structure and solve it using recursion relations. The stop bands emerge in the limit of an infinite number of layers by finding the fixed point of the recursion. We compare to experimental data for various beetles, whose optical structure resembles the proposed model. Thus, using only the phenomenon of interference and the idea of recursion, we are able to elucidate the concept of band structure in the context of the experimentally observed high reflectance and iridescent appearance of structurally colored beetles.

  16. The Recursive Dual-Net and Its Applications

    NASA Astrophysics Data System (ADS)

    Li, Yamin; Peng, Shietung; Chu, Wanming

    In this paper, we propose a universal network, called recursive dual-net (RDN). It can be used as a candidate of effective interconnection networks for massively parallel computers. The RDN is generated by recursively applying dual-construction on a base-network. Given a regular and symmetric graph of size n and node-degree d, the dual-construction generates a regular and symmetric graph of size 2n 2 and node-degree d + 1. The RDN has many interesting properties including low node-degree and small diameter. For example, we can construct an RDN connecting more than 3-million nodes with only 6 links per node and a diameter of 22. We investigate the topological properties of the RDN and compare it to other networks including 3D torus, WK-recursive network, hypercube, cube-connected-cycle, and dual-cube. We also describe an efficient routing algorithm for RDN.

  17. Haydock's recursive solution of self-adjoint problems. Discrete spectrum

    NASA Astrophysics Data System (ADS)

    Moroz, Alexander

    2014-12-01

    Haydock's recursive solution is shown to underline a number of different concepts such as (i) quasi-exactly solvable models, (ii) exactly solvable models, (iii) three-term recurrence solutions based on Schweber's quantization criterion in Hilbert spaces of entire analytic functions, and (iv) a discrete quantum mechanics of Odake and Sasaki. A recurrent theme of Haydock's recursive solution is that the spectral properties of any self-adjoint problem can be mapped onto a corresponding sequence of polynomials {pn(E) } in energy variable E. The polynomials {pn(E) } are orthonormal with respect to the density of states n0(E) and energy eigenstate | E > is the generating function of {pn(E) } . The generality of Haydock's recursive solution enables one to see the different concepts from a unified perspective and mutually benefiting from each other. Some results obtained within the particular framework of any of (i) to (iv) may have much broader significance.

  18. XML — an opportunity for data standards in the geosciences

    NASA Astrophysics Data System (ADS)

    Houlding, Simon W.

    2001-08-01

    Extensible markup language (XML) is a recently introduced meta-language standard on the Web. It provides the rules for development of metadata (markup) standards for information transfer in specific fields. XML allows development of markup languages that describe what information is rather than how it should be presented. This allows computer applications to process the information in intelligent ways. In contrast hypertext markup language (HTML), which fuelled the initial growth of the Web, is a metadata standard concerned exclusively with presentation of information. Besides its potential for revolutionizing Web activities, XML provides an opportunity for development of meaningful data standards in specific application fields. The rapid endorsement of XML by science, industry and e-commerce has already spawned new metadata standards in such fields as mathematics, chemistry, astronomy, multi-media and Web micro-payments. Development of XML-based data standards in the geosciences would significantly reduce the effort currently wasted on manipulating and reformatting data between different computer platforms and applications and would ensure compatibility with the new generation of Web browsers. This paper explores the evolution, benefits and status of XML and related standards in the more general context of Web activities and uses this as a platform for discussion of its potential for development of data standards in the geosciences. Some of the advantages of XML are illustrated by a simple, browser-compatible demonstration of XML functionality applied to a borehole log dataset. The XML dataset and the associated stylesheet and schema declarations are available for FTP download.

  19. XTaGe: A Flexible Generation System for Complex XML Collections

    NASA Astrophysics Data System (ADS)

    Pérez, María; Sanz, Ismael; Berlanga, Rafael

    We introduce XTaGe (XML Tester and Generator), a system for the synthesis of XML collections meant for testing and micro-benchmarking applications. In contrast with existing approaches, XTaGe focuses on complex collections, by providing a highly extensible framework to introduce controlled variability in XML structures. In this paper we present the theoretical foundation, internal architecture and main features of our generator; we describe its implementation, which includes a GUI to facilitate the specification of collections; we discuss how XTaGe’s features compare with those in other XML generation systems; finally, we illustrate its usage by presenting a use case in the Bioinformatics domain.

  20. The XSD-Builder Specification Language—Toward a Semantic View of XML Schema Definition

    NASA Astrophysics Data System (ADS)

    Fong, Joseph; Cheung, San Kuen

    In the present database market, XML database model is a main structure for the forthcoming database system in the Internet environment. As a conceptual schema of XML database, XML Model has its limitation on presenting its data semantics. System analyst has no toolset for modeling and analyzing XML system. We apply XML Tree Model (shown in Figure 2) as a conceptual schema of XML database to model and analyze the structure of an XML database. It is important not only for visualizing, specifying, and documenting structural models, but also for constructing executable systems. The tree model represents inter-relationship among elements inside different logical schema such as XML Schema Definition (XSD), DTD, Schematron, XDR, SOX, and DSD (shown in Figure 1, an explanation of the terms in the figure are shown in Table 1). The XSD-Builder consists of XML Tree Model, source language, translator, and XSD. The source language is called XSD-Source which is mainly for providing an environment with concept of user friendliness while writing an XSD. The source language will consequently be translated by XSD-Translator. Output of XSD-Translator is an XSD which is our target and is called as an object language.

  1. A decoupled recursive approach for constrained flexible multibody system dynamics

    NASA Technical Reports Server (NTRS)

    Lai, Hao-Jan; Kim, Sung-Soo; Haug, Edward J.; Bae, Dae-Sung

    1989-01-01

    A variational-vector calculus approach is employed to derive a recursive formulation for dynamic analysis of flexible multibody systems. Kinematic relationships for adjacent flexible bodies are derived in a companion paper, using a state vector notation that represents translational and rotational components simultaneously. Cartesian generalized coordinates are assigned for all body and joint reference frames, to explicitly formulate deformation kinematics under small deformation kinematics and an efficient flexible dynamics recursive algorithm is developed. Dynamic analysis of a closed loop robot is performed to illustrate efficiency of the algorithm.

  2. Recursive dynamics algorithm for multibody systems with prescribed motion

    NASA Technical Reports Server (NTRS)

    Jain, Abhinandan; Rodriguez, Guillermo

    1993-01-01

    This paper uses spatial operator techniques to develop a new algorithm for the dynamics of multibody systems with hinges undergoing prescribed motion. This algorithm is spatially recursive, and its computational complexity grows only linearly with the number of degrees of freedom in the system. Its structure is a hybrid of known recursive forward and inverse dynamics algorithms for regular multibody systems. Changes to the prescribed/nonprescribed nature of hinges can be implemented during run time since they are handled with very low overhead in the algorithm.

  3. Task allocation onto a hypercube by recursive mincut bipartitioning

    SciTech Connect

    Ercal, F.; Ramanujam, J.; Sadayappan, P. . Computer and Information Science Research Center)

    1990-09-01

    An efficient recursive task allocation scheme, based on the Kernighan-Lin mincut bisection heuristic, is proposed for the effective mapping of tasks of a parallel program onto a hypercube parallel computer. It is evaluated by comparison with an adaptive, scaled simulated annealing method. The recursive allocation scheme is shown to be effective on a number of large test task graphs---its solution quality is nearly as good as that produced by simulated annealing, and its computation time is several orders of magnitude less.

  4. Efficient Scheduling of Recursive Control Flow on GPUs

    SciTech Connect

    Huo, Xin; Krishnamoorthy, Sriram; Agrawal, Gagan

    2013-06-10

    Graphics processing units (GPUs) have rapidly emerged as a very significant player in high performance computing. Single instruction multiple thread (SIMT) pipelines are typically used in GPUs to exploit parallelism and maximize performance. Although support for unstructured control flow has been included in GPUs, efficiently managing thread divergence for arbitrary parallel programs remains a critical challenge. In this paper, we focus on the problem of supporting recursion in modern GPUs. We design and comparatively evaluate various algorithms to manage thread divergence encountered in recursive programs. The results improve upon traditional post-dominator based reconvergence mechanisms designed to handle thread divergence due to control flow within a procedure.

  5. Research on Heterogeneous Data Exchange based on XML

    NASA Astrophysics Data System (ADS)

    Li, Huanqin; Liu, Jinfeng

    Integration of multiple data sources is becoming increasingly important for enterprises that cooperate closely with their partners for e-commerce. OLAP enables analysts and decision makers fast access to various materialized views from data warehouses. However, many corporations have internal business applications deployed on different platforms. This paper introduces a model for heterogeneous data exchange based on XML. The system can exchange and share the data among the different sources. The method used to realize the heterogeneous data exchange is given in this paper.

  6. An XML-Based Protocol for Distributed Event Services

    NASA Technical Reports Server (NTRS)

    Smith, Warren; Gunter, Dan; Quesnel, Darcy; Biegel, Bryan (Technical Monitor)

    2001-01-01

    A recent trend in distributed computing is the construction of high-performance distributed systems called computational grids. One difficulty we have encountered is that there is no standard format for the representation of performance information and no standard protocol for transmitting this information. This limits the types of performance analysis that can be undertaken in complex distributed systems. To address this problem, we present an XML-based protocol for transmitting performance events in distributed systems and evaluate the performance of this protocol.

  7. RosettaLigand docking with flexible XML protocols

    PubMed Central

    Lemmon, Gordon; Meiler, Jens

    2013-01-01

    Summary RosettaLigand is premiere software for predicting how a protein and a small molecule interact. Benchmark studies demonstrate that 70% of the top scoring RosettaLigand predicted interfaces are within 2 Å RMSD from the crystal structure (1). The latest release of Rosetta ligand software includes many new features, such as (1) docking of multiple ligands simultaneously, (2) representing ligands as fragments for greater flexibility, (3) redesign of the interface during docking, and (4) an XML script based interface that gives the user full control of the ligand docking protocol. PMID:22183535

  8. A Low-Storage-Consumption XML Labeling Method for Efficient Structural Information Extraction

    NASA Astrophysics Data System (ADS)

    Liang, Wenxin; Takahashi, Akihiro; Yokota, Haruo

    Recently, labeling methods to extract and reconstruct the structural information of XML data, which are important for many applications such as XPath query and keyword search, are becoming more attractive. To achieve efficient structural information extraction, in this paper we propose C-DO-VLEI code, a novel update-friendly bit-vector encoding scheme, based on register-length bit operations combining with the properties of Dewey Order numbers, which cannot be implemented in other relevant existing schemes such as ORDPATH. Meanwhile, the proposed method also achieves lower storage consumption because it does not require either prefix schema or any reserved codes for node insertion. We performed experiments to evaluate and compare the performance and storage consumption of the proposed method with those of the ORDPATH method. Experimental results show that the execution times for extracting depth information and parent node labels using the C-DO-VLEI code are about 25% and 15% less, respectively, and the average label size using the C-DO-VLEI code is about 24% smaller, comparing with ORDPATH.

  9. Cluster update for tensor network states

    NASA Astrophysics Data System (ADS)

    Wang, Ling; Verstraete, Frank

    2012-02-01

    We propose a novel recursive way of updating the tensors in projected entangled pair states by evolving the tensor in imaginary time evolution on clusters of different sizes. This generalizes the so-called simple update method of Jiang et al. [Phys. Rev. Lett. 101, 090603 (2008)] and the updating schemes in the single layer picture of Pizorn et al. [Phys. Rev. A 83, 052321 (2011)]. A finite-size scaling of the observables as a function of the cluster size provides a remarkable improvement in the accuracy as compared to the simple update scheme. We benchmark our results on the hand of the spin 1/2 staggered dimerized antiferromagnetic model on the square lattice, and accurate results for the magnetization and the critical exponents are determined. Reference L. Wang and F. Verstraete, arXiv:1110.4362.

  10. Semantics Boosts Syntax in Artificial Grammar Learning Tasks with Recursion

    ERIC Educational Resources Information Center

    Fedor, Anna; Varga, Mate; Szathmary, Eors

    2012-01-01

    Center-embedded recursion (CER) in natural language is exemplified by sentences such as "The malt that the rat ate lay in the house." Parsing center-embedded structures is in the focus of attention because this could be one of the cognitive capacities that make humans distinct from all other animals. The ability to parse CER is usually tested by…

  11. Exploring the Recursive Nature of Food and Family Communication

    ERIC Educational Resources Information Center

    Manning, Linda D.

    2006-01-01

    Family meals act as a barometer to signify the changing nature of family life. The primary objective of this activity is to allow students to experience the many ways in which a recursive relationship exists between the food families eat and the patterns of communication families enact. Through this activity, students experience how food and…

  12. Recursion and the Competence/Performance Distinction in AGL Tasks

    ERIC Educational Resources Information Center

    Lobina, David J.

    2011-01-01

    The term "recursion" is used in at least four distinct theoretical senses within cognitive science. Some of these senses in turn relate to the different levels of analysis described by David Marr some 20 years ago; namely, the underlying competence capacity (the "computational" level), the performance operations used in real-time processing (the…

  13. A recursive Bayesian beamforming for steering vector uncertainties

    NASA Astrophysics Data System (ADS)

    Han, Yubing; Zhang, Dongqing

    2013-12-01

    A recursive Bayesian approach to narrowband beamforming for an uncertain steering vector of interest signal is presented. In this paper, the interference-plus-noise covariance matrix and signal power are assumed to be known. The steering vector is modeled as a complex Gaussian random vector that characterizes the level of steering vector uncertainty. Applying the Bayesian model, a recursive algorithm for minimum mean square error (MMSE) estimation is developed. It can be viewed as a mixture of conditional MMSE estimates weighted by the posterior probability density function of the random steering vector given the observed data. The proposed recursive Bayesian beamformer can make use of the information about the steering vector brought by all the observed data until the current short-term integration window and can estimate the mean and covariance of the steering vector recursively. Numerical simulations show that the proposed beamformer with the known signal power and interference-plus-noise covariance matrix outperforms the linearly constrained minimum variance, subspace projection, and other three Bayesian beamformers. After convergence, it has similar performance to the optimal Max-SINR beamformer with the true steering vector.

  14. A Scale-Invariant Treatment for Recursive Path Models.

    ERIC Educational Resources Information Center

    McDonald, Roderick P.; And Others

    1993-01-01

    A reparameterization is formulated that yields estimates of scale-invariant parameters in recursive path models with latent variables, and (asymptotically) correct standard errors, without the use of constrained optimization. The method is based on the logical structure of the reticular action model. (Author)

  15. Unifying Recursion in Partial, Total and General Correctness

    NASA Astrophysics Data System (ADS)

    Guttmann, Walter

    We give an algebraic semantics of non-deterministic, sequential programs which is valid for partial, total and general correctness. It covers full recursion based on a unified approximation order. We provide explicit solutions in terms of the refinement order. As an application, we systematically derive a semantics of while-programs common to the three correctness approaches.

  16. Recursive Vocal Pattern Learning and Generalization in Starlings

    ERIC Educational Resources Information Center

    Bloomfield, Tiffany Corinna

    2012-01-01

    Among known communication systems, human language alone exhibits open-ended productivity of meaning. Interest in the psychological mechanisms supporting this ability, and their evolutionary origins, has resurged following the suggestion that the only uniquely human ability underlying language is a mechanism of recursion. This "Unique…

  17. Augmented-space recursion for partially disordered systems

    NASA Astrophysics Data System (ADS)

    Chakrabarti, Atisdipankar; Mookerjee, Abhijit

    2001-11-01

    Off-stoichiometric alloys exhibit partial disorder, in the sense that only some of the sublattices of the stoichiometric ordered alloy become disordered. This paper puts forward a generalization of the augmented-space recursion introduced earlier by Saha and Mookerjee (Saha T and Mookerjee A 1997 J. Phys.: Condens. Matter 10 2179) for systems with many atoms per unit cell.

  18. Recursive Inversion By Finite-Impulse-Response Filters

    NASA Technical Reports Server (NTRS)

    Bach, Ralph E., Jr.; Baram, Yoram

    1991-01-01

    Recursive approximation gives least-squares best fit to exact response. Algorithm yields finite-impulse-response approximation of unknown single-input/single-output, causal, time-invariant, linear, real system, response of which is sequence of impulses. Applicable to such system-inversion problems as suppression of echoes and identification of target from its scatter response to incident impulse.

  19. Recursive inversion of externally defined linear systems by FIR filters

    NASA Technical Reports Server (NTRS)

    Bach, Ralph E., Jr.; Baram, Yoram

    1989-01-01

    The approximate inversion of an internally unknown linear system, given by its impulse response sequence, by an inverse system having a finite impulse response, is considered. The recursive least-squares procedure is shown to have an exact initialization, based on the triangular Toeplitz structure of the matrix involved. The proposed approach also suggests solutions to the problem of system identification and compensation.

  20. Recursion and the Competence/Performance Distinction in AGL Tasks

    ERIC Educational Resources Information Center

    Lobina, David J.

    2011-01-01

    The term "recursion" is used in at least four distinct theoretical senses within cognitive science. Some of these senses in turn relate to the different levels of analysis described by David Marr some 20 years ago; namely, the underlying competence capacity (the "computational" level), the performance operations used in real-time processing (the

  1. Recursivity: A Working Paper on Rhetoric and "Mnesis"

    ERIC Educational Resources Information Center

    Stormer, Nathan

    2013-01-01

    This essay proposes the genealogical study of remembering and forgetting as recursive rhetorical capacities that enable discourse to place itself in an ever-changing present. "Mnesis" is a meta-concept for the arrangements of remembering and forgetting that enable rhetoric to function. Most of the essay defines the materiality of "mnesis", first

  2. Recursivity: A Working Paper on Rhetoric and "Mnesis"

    ERIC Educational Resources Information Center

    Stormer, Nathan

    2013-01-01

    This essay proposes the genealogical study of remembering and forgetting as recursive rhetorical capacities that enable discourse to place itself in an ever-changing present. "Mnesis" is a meta-concept for the arrangements of remembering and forgetting that enable rhetoric to function. Most of the essay defines the materiality of "mnesis", first…

  3. XML schemas and mark-up practices of taxonomic literature

    PubMed Central

    Penev, Lyubomir; Lyal, Christopher HC; Weitzman, Anna; Morse, David R.; King, David; Sautter, Guido; Georgiev, Teodor; Morris, Robert A.; Catapano, Terry; Agosti, Donat

    2011-01-01

    Abstract We review the three most widely used XML schemas used to mark-up taxonomic texts, TaxonX, TaxPub and taXMLit. These are described from the viewpoint of their development history, current status, implementation, and use cases. The concept of “taxon treatment” from the viewpoint of taxonomy mark-up into XML is discussed. TaxonX and taXMLit are primarily designed for legacy literature, the former being more lightweight and with a focus on recovery of taxon treatments, the latter providing a much more detailed set of tags to facilitate data extraction and analysis. TaxPub is an extension of the National Library of Medicine Document Type Definition (NLM DTD) for taxonomy focussed on layout and recovery and, as such, is best suited for mark-up of new publications and their archiving in PubMedCentral. All three schemas have their advantages and shortcomings and can be used for different purposes. PMID:22207808

  4. XML-Based Visual Specification of Multidisciplinary Applications

    NASA Technical Reports Server (NTRS)

    Al-Theneyan, Ahmed; Jakatdar, Amol; Mehrotra, Piyush; Zubair, Mohammad

    2001-01-01

    The advancements in the Internet and Web technologies have fueled a growing interest in developing a web-based distributed computing environment. We have designed and developed Arcade, a web-based environment for designing, executing, monitoring, and controlling distributed heterogeneous applications, which is easy to use and access, portable, and provides support through all phases of the application development and execution. A major focus of the environment is the specification of heterogeneous, multidisciplinary applications. In this paper we focus on the visual and script-based specification interface of Arcade. The web/browser-based visual interface is designed to be intuitive to use and can also be used for visual monitoring during execution. The script specification is based on XML to: (1) make it portable across different frameworks, and (2) make the development of our tools easier by using the existing freely available XML parsers and editors. There is a one-to-one correspondence between the visual and script-based interfaces allowing users to go back and forth between the two. To support this we have developed translators that translate a script-based specification to a visual-based specification, and vice-versa. These translators are integrated with our tools and are transparent to users.

  5. Towards health care process description framework: an XML DTD design.

    PubMed

    Staccini, P; Joubert, M; Quaranta, J F; Aymard, S; Fieschi, D; Fieschi, M

    2001-01-01

    The development of health care and hospital information systems has to meet users needs as well as requirements such as the tracking of all care activities and the support of quality improvement. The use of process-oriented analysis is of-value to provide analysts with: (i) a systematic description of activities; (ii) the elicitation of the useful data to perform and record care tasks; (iii) the selection of relevant decision-making support. But paper-based tools are not a very suitable way to manage and share the documentation produced during this step. The purpose of this work is to propose a method to implement the results of process analysis according to XML techniques (eXtensible Markup Language). It is based on the IDEF0 activity modeling language (Integration DEfinition for Function modeling). A hierarchical description of a process and its components has been defined through a flat XML file with a grammar of proper metadata tags. Perspectives of this method are discussed. PMID:11825265

  6. Teaching and Learning Recursive Programming: A Review of the Research Literature

    ERIC Educational Resources Information Center

    McCauley, Renée; Grissom, Scott; Fitzgerald, Sue; Murphy, Laurie

    2015-01-01

    Hundreds of articles have been published on the topics of teaching and learning recursion, yet fewer than 50 of them have published research results. This article surveys the computing education research literature and presents findings on challenges students encounter in learning recursion, mental models students develop as they learn recursion,…

  7. Integrated Syntactic/Semantic XML Data Validation with a Reusable Software Component

    ERIC Educational Resources Information Center

    Golikov, Steven

    2013-01-01

    Data integration is a critical component of enterprise system integration, and XML data validation is the foundation for sound data integration of XML-based information systems. Since B2B e-commerce relies on data validation as one of the critical components for enterprise integration, it is imperative for financial industries and e-commerce

  8. Integrated Syntactic/Semantic XML Data Validation with a Reusable Software Component

    ERIC Educational Resources Information Center

    Golikov, Steven

    2013-01-01

    Data integration is a critical component of enterprise system integration, and XML data validation is the foundation for sound data integration of XML-based information systems. Since B2B e-commerce relies on data validation as one of the critical components for enterprise integration, it is imperative for financial industries and e-commerce…

  9. On the convergence improvement in the metadynamics simulations: a Wang-Landau recursion approach.

    PubMed

    Min, Donghong; Liu, Yusong; Carbone, Irina; Yang, Wei

    2007-05-21

    As a popular tool in exploring free energy landscapes, the metadynamics method has been widely applied to elucidate various chemical or biochemical processes. As deeply discussed by Laio et al. [J. Phys. Chem. B 109, 6714 (2005)], the size of the updating Gaussian function is pivotal to the free energy convergence toward the target free energy surface. For instance, a greater Gaussian height can facilitate the quick visit of a conformation region of interest; however, it may lead to a larger error of the calculated free energy surface. In contrast, a lower Gaussian height can guarantee a better resolution of the calculated free energy surface; however, it will take longer time for such a simulation to navigate through the defined conformational region. In order to reconcile such confliction, the authors present a method by implementing the Wang-Landau recursion scheme in the metadynamics simulations to adaptively update the height of the unit Gaussian function. As demonstrated in their model studies on both a toy system, and a realistic molecular system treated with the hybrid quantum mechanical and molecular mechanical (QMMM) potential, the present approach can quickly result in more decently converged free energy surfaces, compared with the classical metadynamics simulations employing the fixed Gaussian heights. PMID:17523795

  10. An investigation of a manipulative simulation in the learning of recursive programming

    NASA Astrophysics Data System (ADS)

    Bower, Randall Wayne

    Recursion is a fundamentally important topic in computer science. Even so, it is often omitted in introductory courses, or discussed only briefly. This is likely due, at least in part, to the fact that teaching recursion has been difficult. Perhaps the biggest problem in teaching recursion is that there are few, if any, naturally existing examples of recursion in our lives. However, successful simulations have shown that the computer may hold the key to solving this problem. A simulation of recursion presented to students before formal classroom instruction can provide a foundation of concrete experiences to build upon. The challenge is to develop an appropriate simulation and lesson plan for introducing recursion to students early in their programming experience. This research reviews previous attempts at teaching recursion, including detailed lesson plans, mental models of recursion, and other simulations. Then, a new simulation and lesson plan for its use are described. The effectiveness of the simulation is studied using two groups of students enrolled in a college-level, introductory programming course. Results indicate that students who used the simulation as their first exposure to recursion gained a deeper understanding of recursion than students receiving a lecture-based introduction to recursion. Specifically, students who used the simulation required fewer attempts to complete a set of recursive programming exercises and performed better on a follow-up exam given six weeks after the experiment. This research concludes with a discussion of two important questions: How should students think about recursion and how do they think about recursion. The simulation's strengths and shortcomings in fostering effective ways of thinking about recursion are also discussed.

  11. XML at the ADC: Steps to a Next Generation Data Repository

    NASA Astrophysics Data System (ADS)

    Shaya, E.; Gass, J.; Blackwell, J.; Thomas, B.; Holmes, B.; Cheung, C. Y.

    The staff of the Astronomical Data Center at GSFC (ADC, http://adc.gsfc.nasa.gov) is involved in a research project to define the XML format for the metadata of an astronomical repository and for large data tables. In the process, an XML tool box is being developed for importation, enhancement, and distribution of published data and their metadata documents. There is now a working draft Document Type Definition (DTD, http://messier.gsfc.nasa.gov/xml/dataset.dtd) which specifies the required elements of content and their attributes. The documentation for each data set will be viewable in several different styles via eXtensible Stylesheet Language Transformations (XSLT, (http://www.w3.org/TR/xslt) scripts. The ADC is actively creating designs for the flow of data through automated pipelines from authors and journal presses into an XML archive, as well as data retrieval through the web via the XML Query Language.

  12. Using XML-Schema to Model Data from Present and Future Astronomical Databases

    NASA Astrophysics Data System (ADS)

    Debray, Bernard

    XML (eXtensible Markup Language) has now become an almost inescapable tool for the exchange of formatted data over the World Wide Web and is now becoming extensively used in astronomical data bases to interweave them, therefore making it an important tool for the building of the Virtual Observatory. Document Type Definitions (DTD) have been initially used to describe XML documents formatting; they offer very limited possibility about data typing, constraints or bounds for the handling of data-oriented documents. These limitations can now be overcome thanks to the XML Schema recommendation. The potential benefit of XML Schema for astronomical databases is illustrated through an ongoing implementation of XML Schema for formatted data output in the Besancon Double and Multiple Star database.

  13. Creating preservation metadata from XML-metadata profiles

    NASA Astrophysics Data System (ADS)

    Ulbricht, Damian; Bertelmann, Roland; Gebauer, Petra; Hasler, Tim; Klump, Jens; Kirchner, Ingo; Peters-Kottig, Wolfgang; Mettig, Nora; Rusch, Beate

    2014-05-01

    Registration of dataset DOIs at DataCite makes research data citable and comes with the obligation to keep data accessible in the future. In addition, many universities and research institutions measure data that is unique and not repeatable like the data produced by an observational network and they want to keep these data for future generations. In consequence, such data should be ingested in preservation systems, that automatically care for file format changes. Open source preservation software that is developed along the definitions of the ISO OAIS reference model is available but during ingest of data and metadata there are still problems to be solved. File format validation is difficult, because format validators are not only remarkably slow - due to variety in file formats different validators return conflicting identification profiles for identical data. These conflicts are hard to resolve. Preservation systems have a deficit in the support of custom metadata. Furthermore, data producers are sometimes not aware that quality metadata is a key issue for the re-use of data. In the project EWIG an university institute and a research institute work together with Zuse-Institute Berlin, that is acting as an infrastructure facility, to generate exemplary workflows for research data into OAIS compliant archives with emphasis on the geosciences. The Institute for Meteorology provides timeseries data from an urban monitoring network whereas GFZ Potsdam delivers file based data from research projects. To identify problems in existing preservation workflows the technical work is complemented by interviews with data practitioners. Policies for handling data and metadata are developed. Furthermore, university teaching material is created to raise the future scientists awareness of research data management. As a testbed for ingest workflows the digital preservation system Archivematica [1] is used. During the ingest process metadata is generated that is compliant to the Metadata Encoding and Transmission Standard (METS). To find datasets in future portals and to make use of this data in own scientific work, proper selection of discovery metadata and application metadata is very important. Some XML-metadata profiles are not suitable for preservation, because version changes are very fast and make it nearly impossible to automate the migration. For other XML-metadata profiles schema definitions are changed after publication of the profile or the schema definitions become inaccessible, which might cause problems during validation of the metadata inside the preservation system [2]. Some metadata profiles are not used widely enough and might not even exist in the future. Eventually, discovery and application metadata have to be embedded into the mdWrap-subtree of the METS-XML. [1] http://www.archivematica.org [2] http://dx.doi.org/10.2218/ijdc.v7i1.215

  14. On recursive least-squares filtering algorithms and implementations. Ph.D. Thesis

    NASA Technical Reports Server (NTRS)

    Hsieh, Shih-Fu

    1990-01-01

    In many real-time signal processing applications, fast and numerically stable algorithms for solving least-squares problems are necessary and important. In particular, under non-stationary conditions, these algorithms must be able to adapt themselves to reflect the changes in the system and take appropriate adjustments to achieve optimum performances. Among existing algorithms, the QR-decomposition (QRD)-based recursive least-squares (RLS) methods have been shown to be useful and effective for adaptive signal processing. In order to increase the speed of processing and achieve high throughput rate, many algorithms are being vectorized and/or pipelined to facilitate high degrees of parallelism. A time-recursive formulation of RLS filtering employing block QRD will be considered first. Several methods, including a new non-continuous windowing scheme based on selectively rejecting contaminated data, were investigated for adaptive processing. Based on systolic triarrays, many other forms of systolic arrays are shown to be capable of implementing different algorithms. Various updating and downdating systolic algorithms and architectures for RLS filtering are examined and compared in details, which include Householder reflector, Gram-Schmidt procedure, and Givens rotation. A unified approach encompassing existing square-root-free algorithms is also proposed. For the sinusoidal spectrum estimation problem, a judicious method of separating the noise from the signal is of great interest. Various truncated QR methods are proposed for this purpose and compared to the truncated SVD method. Computer simulations provided for detailed comparisons show the effectiveness of these methods. This thesis deals with fundamental issues of numerical stability, computational efficiency, adaptivity, and VLSI implementation for the RLS filtering problems. In all, various new and modified algorithms and architectures are proposed and analyzed; the significance of any of the new method depends crucially on specific application.

  15. Teaching object concepts for XML-based representations.

    SciTech Connect

    Kelsey, R. L.

    2002-01-01

    Students learned about object-oriented design concepts and knowledge representation through the use of a set of toy blocks. The blocks represented a limited and focused domain of knowledge and one that was physical and tangible. The blocks helped the students to better visualize, communicate, and understand the domain of knowledge as well as how to perform object decomposition. The blocks were further abstracted to an engineering design kit for water park design. This helped the students to work on techniques for abstraction and conceptualization. It also led the project from tangible exercises into software and programming exercises. Students employed XML to create object-based knowledge representations and Java to use the represented knowledge. The students developed and implemented software allowing a lay user to design and create their own water slide and then to take a simulated ride on their slide.

  16. DNA Data Bank of Japan (DDBJ) in XML.

    PubMed

    Miyazaki, S; Sugawara, H; Gojobori, T; Tateno, Y

    2003-01-01

    The DNA Data Bank of Japan (DDBJ, http://www.ddbj.nig.ac.jp) has collected and released more entries and bases than last year. This is mainly due to large-scale submissions from Japanese sequencing teams on mouse, rice, chimpanzee, nematoda and other organisms. The contributions of DDBJ over the past year are 17.3% (entries) and 10.3% (bases) of the combined outputs of the International Nucleotide Sequence Databases (INSD). Our complete genome sequence database, Genome Information Broker (GIB), has been improved by incorporating XML. It is now possible to perform a more sophisticated database search against the new GIB than the ordinary BLAST or FASTA search. PMID:12519938

  17. Incorporating Load Balancing Spatial Analysis Into Xml-Based Webgis

    NASA Astrophysics Data System (ADS)

    Huang, H.

    2012-07-01

    This article aims to introduce load balancing spatial analysis into XML-based WebGIS. In contrast to other approaches that implement spatial queries and analyses solely on server or browser sides, load balancing spatial analysis carries out spatial analysis on either the server or the browser sides depending on the execution costs (i.e., network transmission costs and computational costs). In this article, key elements of load balancing middlewares are investigated, and relevant solution is proposed. The comparison with server-side solution, browse-side solution, and our former solution shows that the proposed solution can optimize the execution of spatial analysis, greatly ease the network transmission load between the server and the browser sides, and therefore lead to a better performance. The proposed solution enables users to access high-performance spatial analysis simply via a web browser.

  18. Managing and querying image annotation and markup in XML

    NASA Astrophysics Data System (ADS)

    Wang, Fusheng; Pan, Tony; Sharma, Ashish; Saltz, Joel

    2010-03-01

    Proprietary approaches for representing annotations and image markup are serious barriers for researchers to share image data and knowledge. The Annotation and Image Markup (AIM) project is developing a standard based information model for image annotation and markup in health care and clinical trial environments. The complex hierarchical structures of AIM data model pose new challenges for managing such data in terms of performance and support of complex queries. In this paper, we present our work on managing AIM data through a native XML approach, and supporting complex image and annotation queries through native extension of XQuery language. Through integration with xService, AIM databases can now be conveniently shared through caGrid.

  19. Using XML and Java Technologies for Astronomical Instrument Control

    NASA Technical Reports Server (NTRS)

    Ames, Troy; Case, Lynne; Powers, Edward I. (Technical Monitor)

    2001-01-01

    Traditionally, instrument command and control systems have been highly specialized, consisting mostly of custom code that is difficult to develop, maintain, and extend. Such solutions are initially very costly and are inflexible to subsequent engineering change requests, increasing software maintenance costs. Instrument description is too tightly coupled with details of implementation. NASA Goddard Space Flight Center, under the Instrument Remote Control (IRC) project, is developing a general and highly extensible framework that applies to any kind of instrument that can be controlled by a computer. The software architecture combines the platform independent processing capabilities of Java with the power of the Extensible Markup Language (XML), a human readable and machine understandable way to describe structured data. A key aspect of the object-oriented architecture is that the software is driven by an instrument description, written using the Instrument Markup Language (IML), a dialect of XML. IML is used to describe the command sets and command formats of the instrument, communication mechanisms, format of the data coming from the instrument, and characteristics of the graphical user interface to control and monitor the instrument. The IRC framework allows the users to define a data analysis pipeline which converts data coming out of the instrument. The data can be used in visualizations in order for the user to assess the data in real-time, if necessary. The data analysis pipeline algorithms can be supplied by the user in a variety of forms or programming languages. Although the current integration effort is targeted for the High-resolution Airborne Wideband Camera (HAWC) and the Submillimeter and Far Infrared Experiment (SAFIRE), first-light instruments of the Stratospheric Observatory for Infrared Astronomy (SOFIA), the framework is designed to be generic and extensible so that it can be applied to any instrument. Plans are underway to test the framework with other types of instruments, such as remote sensing earth science instruments.

  20. Common Data Format: New XML and Conversion Tools

    NASA Astrophysics Data System (ADS)

    Han, D. B.; Liu, M. H.; McGuire, R. E.

    2002-12-01

    Common Data Format (CDF) is a self-describing platform-independent data format for storing, accessing, and manipulating scalar and multidimensional scientific data sets. Significant benefit has accrued to specific science communities from their use of standard formats within those communities. Examples include the International Solar Terrestrial Physics (ISTP) community in using CDF for traditional space physics data (fields, particles and plasma, waves, and images), the worldwide astronomical community in using FITS (Flexible Image Transport System) for solar data (primarily spectral images), the NASA Planetary community in using Planetary Data System (PDS) Labels, and the earth science community in using Hierarchical Data Format (HDF). Scientific progress in solar-terrestrial physics continues to be impeded by the multiplicity of available standards for data formats and dearth of general data format translators. As a result, scientists today spend a significant amount of time translating data into the format they are familiar with for their research. To minimize this unnecessary data translation time and to allow more research time, the CDF office located at GSFC National Space Science Data Center (NSSDC) has developed HDF-to-CDF and FITS-to-CDF translators, and employed the eXtensible Markup Language (XML) technology to facilitate and promote data interoperability within the space science community. We will present the current status of the CDF work including the conversion tools that have been recently developed, conversion tools that are planned in the near future, share some of the XML experiences, and use the discussion to gain community feedback to our planned future work.

  1. Development of a recursion RNG-based turbulence model

    NASA Technical Reports Server (NTRS)

    Zhou, YE; Vahala, George; Thangam, S.

    1993-01-01

    Reynolds stress closure models based on the recursion renormalization group theory are developed for the prediction of turbulent separated flows. The proposed model uses a finite wavenumber truncation scheme to account for the spectral distribution of energy. In particular, the model incorporates effects of both local and nonlocal interactions. The nonlocal interactions are shown to yield a contribution identical to that from the epsilon-renormalization group (RNG), while the local interactions introduce higher order dispersive effects. A formal analysis of the model is presented and its ability to accurately predict separated flows is analyzed from a combined theoretical and computational stand point. Turbulent flow past a backward facing step is chosen as a test case and the results obtained based on detailed computations demonstrate that the proposed recursion -RNG model with finite cut-off wavenumber can yield very good predictions for the backstep problem.

  2. Topological recursion for Gaussian means and cohomological field theories

    NASA Astrophysics Data System (ADS)

    Andersen, J. E.; Chekhov, L. O.; Norbury, P.; Penner, R. C.

    2015-12-01

    We introduce explicit relations between genus-filtrated s-loop means of the Gaussian matrix model and terms of the genus expansion of the Kontsevich-Penner matrix model (KPMM), which is the generating function for volumes of discretized (open) moduli spaces M g,s disc (discrete volumes). Using these relations, we express Gaussian means in all orders of the genus expansion as polynomials in special times weighted by ancestor invariants of an underlying cohomological field theory. We translate the topological recursion of the Gaussian model into recurrence relations for the coefficients of this expansion, which allows proving that they are integers and positive. We find the coefficients in the first subleading order for M g,1 for all g in three ways: using the refined Harer-Zagier recursion, using the Givental-type decomposition of the KPMM, and counting diagrams explicitly.

  3. Recursive utility in a Markov environment with stochastic growth.

    PubMed

    Hansen, Lars Peter; Scheinkman, José A

    2012-07-24

    Recursive utility models that feature investor concerns about the intertemporal composition of risk are used extensively in applied research in macroeconomics and asset pricing. These models represent preferences as the solution to a nonlinear forward-looking difference equation with a terminal condition. In this paper we study infinite-horizon specifications of this difference equation in the context of a Markov environment. We establish a connection between the solution to this equation and to an arguably simpler Perron-Frobenius eigenvalue equation of the type that occurs in the study of large deviations for Markov processes. By exploiting this connection, we establish existence and uniqueness results. Moreover, we explore a substantive link between large deviation bounds for tail events for stochastic consumption growth and preferences induced by recursive utility. PMID:22778428

  4. Recursion relations and branching rules for simple Lie algebras

    NASA Astrophysics Data System (ADS)

    Lyakhovsky, V. D.; Melnikov, S. Yu

    1996-03-01

    The branching rules between simple Lie algebras and their regular (maximal) simple subalgebras are studied. Two types of recursion relations for anomalous relative multiplicities are obtained. One of them is proved to be the factorized version of the other. The factorization property is based on the existence of the set of weights 0305-4470/29/5/020/img1 specific for each injection. The structure of 0305-4470/29/5/020/img1 is easily deduced from the correspondence between the root systems of the algebra and subalgebra. The recursion relations thus obtained give rise to a simple and effective algorithm for branching rules. The details are illustrated by performing the explicit decomposition procedure for the injection 0305-4470/29/5/020/img3.

  5. Functional Dual Adaptive Control with Recursive Gaussian Process Model

    NASA Astrophysics Data System (ADS)

    Prüher, Jakub; Král, Ladislav

    2015-11-01

    The paper deals with dual adaptive control problem, where the functional uncertainties in the system description are modelled by a non-parametric Gaussian process regression model. Current approaches to adaptive control based on Gaussian process models are severely limited in their practical applicability, because the model is re-adjusted using all the currently available data, which keeps growing with every time step. We propose the use of recursive Gaussian process regression algorithm for significant reduction in computational requirements, thus bringing the Gaussian process-based adaptive controllers closer to their practical applicability. In this work, we design a bi-criterial dual controller based on recursive Gaussian process model for discrete-time stochastic dynamic systems given in an affine-in-control form. Using Monte Carlo simulations, we show that the proposed controller achieves comparable performance with the full Gaussian process-based controller in terms of control quality while keeping the computational demands bounded.

  6. ROOM: A recursive object oriented method for information systems development

    SciTech Connect

    Thelliez, T.; Donahue, S.

    1994-02-09

    Although complementary for the development of complex systems, top-down structured design and object oriented approach are still opposed and not integrated. As the complexity of the systems are still growing, and the so-called software crisis still not solved, it is urgent to provide a framework mixing the two paradigms. This paper presents an elegant attempt in this direction through our Recursive Object-Oriented Method (ROOM) in which a top-down approach divides the complexity of the system and an object oriented method studies a given level of abstraction. Illustrating this recursive schema with a simple example, we demonstrate that we achieve the goal of creating loosely coupled and reusable components.

  7. EEG and MEG source localization using recursively applied (RAP) MUSIC

    SciTech Connect

    Mosher, J.C.; Leahy, R.M.

    1996-12-31

    The multiple signal characterization (MUSIC) algorithm locates multiple asynchronous dipolar sources from electroencephalography (EEG) and magnetoencephalography (MEG) data. A signal subspace is estimated from the data, then the algorithm scans a single dipole model through a three-dimensional head volume and computes projections onto this subspace. To locate the sources, the user must search the head volume for local peaks in the projection metric. Here we describe a novel extension of this approach which we refer to as RAP (Recursively APplied) MUSIC. This new procedure automatically extracts the locations of the sources through a recursive use of subspace projections, which uses the metric of principal correlations as a multidimensional form of correlation analysis between the model subspace and the data subspace. The dipolar orientations, a form of `diverse polarization,` are easily extracted using the associated principal vectors.

  8. Exact calculations of first-passage quantities on recursive networks

    NASA Astrophysics Data System (ADS)

    Meyer, B.; Agliari, E.; Bénichou, O.; Voituriez, R.

    2012-02-01

    We present general methods to exactly calculate mean first-passage quantities on self-similar networks defined recursively. In particular, we calculate the mean first-passage time and the splitting probabilities associated to a source and one or several targets; averaged quantities over a given set of sources (e.g., same-connectivity nodes) are also derived. The exact estimate of such quantities highlights the dependency of first-passage processes with respect to the source-target distance, which has recently revealed to be a key parameter in characterizing transport in complex media. We explicitly perform calculations for different classes of recursive networks [finitely ramified fractals, scale-free (trans)fractals, nonfractals, mixtures between fractals and nonfractals, nondecimable hierarchical graphs] of arbitrary size. Our approach unifies and significantly extends the available results in the field.

  9. A fast recursive algorithm for molecular dynamics simulation

    NASA Technical Reports Server (NTRS)

    Jain, A.; Vaidehi, N.; Rodriguez, G.

    1993-01-01

    The present recursive algorithm for solving molecular systems' dynamical equations of motion employs internal variable models that reduce such simulations' computation time by an order of magnitude, relative to Cartesian models. Extensive use is made of spatial operator methods recently developed for analysis and simulation of the dynamics of multibody systems. A factor-of-450 speedup over the conventional O(N-cubed) algorithm is demonstrated for the case of a polypeptide molecule with 400 residues.

  10. Using XML to improve the productivity and robustness in application development in geosciences

    NASA Astrophysics Data System (ADS)

    Mello, Ulisses T.; Xu, Liqing

    2006-12-01

    In this paper, we describe an approach to apply Extensible Markup Language (XML) technologies to improve the robustness of geological and geophysical applications as well as to increase the efficacy in the application development process. Geological and geophysical applications are often data centric, I/O intensive and their development is incremental. Therefore, significant amount of development resources is devoted to the design and reengineering of the container data structures that store data. This process is time consuming, mechanical and error prone. Normally, ad hoc parsers are necessary for reading inputs, as well as numerous filters, or adapters to transform the data for integration with other legacy applications. Most of this can be avoided by using XML-related technologies. XML has a type system schema that can be used to define input parameters and constraints. The XML parser can validate the input data using the constraints defined in the schema. Exporting results in XML format allows the use of Extensible Stylesheet Language Transformations (XSLT) to transform XML output to any other format necessary for integration with legacy applications. Additionally, XML-data binding code can be automatically generated in specified languages such C++ and Java. We used this approach to develop applications for seismic ray-tracing and basin modeling with great success, and the major benefits of this approach were the significant gains in productivity during the developement and application robustness.

  11. Recursive linearization of multibody dynamics equations of motion

    NASA Technical Reports Server (NTRS)

    Lin, Tsung-Chieh; Yae, K. Harold

    1989-01-01

    The equations of motion of a multibody system are nonlinear in nature, and thus pose a difficult problem in linear control design. One approach is to have a first-order approximation through the numerical perturbations at a given configuration, and to design a control law based on the linearized model. Here, a linearized model is generated analytically by following the footsteps of the recursive derivation of the equations of motion. The equations of motion are first written in a Newton-Euler form, which is systematic and easy to construct; then, they are transformed into a relative coordinate representation, which is more efficient in computation. A new computational method for linearization is obtained by applying a series of first-order analytical approximations to the recursive kinematic relationships. The method has proved to be computationally more efficient because of its recursive nature. It has also turned out to be more accurate because of the fact that analytical perturbation circumvents numerical differentiation and other associated numerical operations that may accumulate computational error, thus requiring only analytical operations of matrices and vectors. The power of the proposed linearization algorithm is demonstrated, in comparison to a numerical perturbation method, with a two-link manipulator and a seven degrees of freedom robotic manipulator. Its application to control design is also demonstrated.

  12. Recursive Bayesian electromagnetic refractivity estimation from radar sea clutter

    NASA Astrophysics Data System (ADS)

    Vasudevan, Sathyanarayanan; Anderson, Richard H.; Kraut, Shawn; Gerstoft, Peter; Rogers, L. Ted; Krolik, Jeffrey L.

    2007-04-01

    Estimation of the range- and height-dependent index of refraction over the sea surface facilitates prediction of ducted microwave propagation loss. In this paper, refractivity estimation from radar clutter returns is performed using a Markov state space model for microwave propagation. Specifically, the parabolic approximation for numerical solution of the wave equation is used to formulate the refractivity from clutter (RFC) problem within a nonlinear recursive Bayesian state estimation framework. RFC under this nonlinear state space formulation is more efficient than global fitting of refractivity parameters when the total number of range-varying parameters exceeds the number of basis functions required to represent the height-dependent field at a given range. Moreover, the range-recursive nature of the estimator can be easily adapted to situations where the refractivity modeling changes at discrete ranges, such as at a shoreline. A fast range-recursive solution for obtaining range-varying refractivity is achieved by using sequential importance sampling extensions to state estimation techniques, namely, the forward and Viterbi algorithms. Simulation and real data results from radar clutter collected off Wallops Island, Virginia, are presented which demonstrate the ability of this method to produce propagation loss estimates that compare favorably with ground truth refractivity measurements.

  13. Recursive linear optical networks for realizing quantum algorithms

    NASA Astrophysics Data System (ADS)

    Tabia, Gelo Noel

    Linear optics has played a leading role in the development of practical quantum technologies. In recent years, advances in integrated quantum photonics have significantly improved the functionality and scalability of linear optical devices. In this talk, I present recursive schemes for implementing quantum Fourier transforms and inversion about the mean in Grover's algorithm with photonic integrated circuits. By recursive, I mean that two copies of a d-dimensional unitary operation is used to build the corresponding unitary operation on 2 d modes. The linear optical networks operate on path-encoded qudits and realize d-dimensional unitary operations using O (d2) elements. To demonstrate that the recursive circuits are viable in practice, I conducted simulations of proof-of-principle experiments using a fabrication model of realistic errors in silicon-based photonic integrated devices. The results indicate high-fidelity performance in the circuits for 2-qubit and 3-qubit quantum Fourier transforms, and for quantum search on 4-item and 8-item databases. This work was funded by institutional research grant IUT2-1 from the Estonian Research Council and by the European Union through the European Regional Development Fund.

  14. An Accelerated Recursive Doubling Algorithm for Block Tridiagonal Systems

    SciTech Connect

    Seal, Sudip K

    2014-01-01

    Block tridiagonal systems of linear equations arise in a wide variety of scientific and engineering applications. Recursive doubling algorithm is a well-known prefix computation-based numerical algorithm that requires O(M^3(N/P + log P)) work to compute the solution of a block tridiagonal system with N block rows and block size M on P processors. In real-world applications, solutions of tridiagonal systems are most often sought with multiple, often hundreds and thousands, of different right hand sides but with the same tridiagonal matrix. Here, we show that a recursive doubling algorithm is sub-optimal when computing solutions of block tridiagonal systems with multiple right hand sides and present a novel algorithm, called the accelerated recursive doubling algorithm, that delivers O(R) improvement when solving block tridiagonal systems with R distinct right hand sides. Since R is typically about 100 1000, this improvement translates to very significant speedups in practice. Detailed complexity analyses of the new algorithm with empirical confirmation of runtime improvements are presented. To the best of our knowledge, this algorithm has not been reported before in the literature.

  15. Categorial Compositionality III: F-(co)algebras and the Systematicity of Recursive Capacities in Human Cognition

    PubMed Central

    Phillips, Steven; Wilson, William H.

    2012-01-01

    Human cognitive capacity includes recursively definable concepts, which are prevalent in domains involving lists, numbers, and languages. Cognitive science currently lacks a satisfactory explanation for the systematic nature of such capacities (i.e., why the capacity for some recursive cognitive abilities–e.g., finding the smallest number in a list–implies the capacity for certain others–finding the largest number, given knowledge of number order). The category-theoretic constructs of initial F-algebra, catamorphism, and their duals, final coalgebra and anamorphism provide a formal, systematic treatment of recursion in computer science. Here, we use this formalism to explain the systematicity of recursive cognitive capacities without ad hoc assumptions (i.e., to the same explanatory standard used in our account of systematicity for non-recursive capacities). The presence of an initial algebra/final coalgebra explains systematicity because all recursive cognitive capacities, in the domain of interest, factor through (are composed of) the same component process. Moreover, this factorization is unique, hence no further (ad hoc) assumptions are required to establish the intrinsic connection between members of a group of systematically-related capacities. This formulation also provides a new perspective on the relationship between recursive cognitive capacities. In particular, the link between number and language does not depend on recursion, as such, but on the underlying functor on which the group of recursive capacities is based. Thus, many species (and infants) can employ recursive processes without having a full-blown capacity for number and language. PMID:22514704

  16. Metaphor, recursive systems, and paradox in science and developmental theory.

    PubMed

    Overton, W F

    1991-01-01

    The contradictions found at any level of abstraction among concepts such as subject-object, whole-part, synthesis-analysis, metaphor-observation, organicism-mechanism, and interpretationism-realism cannot be eliminated or resolved at that level. They can, however, be reconciled into productive paradoxes by recognizing them as components of recursive systems. The resolution of the paradox occurs only at the next higher level of abstraction where a synthesis can be established. However, this synthesis at the next higher level entails its own contradictions. These can again be reconciled into productive paradoxes through the recognition of broader recursive systems. This progressive solution continues at each iteration, or level or recursion, and it is illustrated in Fig. 5. In the figure, the innermost cycle represents the knowing organism, knowing in the paradoxical cycle of metaphor-assumptions-concepts-observations. This knowing organism is explained and hence understood, and the paradoxes reconciled, only by moving to the next level of recursive cycle. At this next level, the first level of abstraction, psychological theories operate to explain the phenomenological knowing organism of the first cycle. In fact, it is only by moving to this level that we transcend the paradox of man knowing and explaining himself. And only by moving to this level do we avoid the vicious circularity that could befall the use of recursive systems. However, the theories themselves involve a new paradoxical cycle of metaphor-assumptions-concepts-observation. Thus, explanation of this level requires movement to the next level of abstraction or next outer cycle. Here metatheoretical assumptions provide the transcendence and the opportunity to reconcile the paradoxes of psychological theory. But this level too operates in a cycle of paradoxes and consequently the process continues as it does for any dialectic process. In closing, I should in fairness note that I have outlined only one type of solution to contradictions that are found among approaches to the game called science and the game called developmental psychology. It is a solution that draws heavily on the categories of dialectical method, and it is just this method that is both the reason for, and the consequence of, the organismic metaphor. Thus, my solution generates its own contradiction for I have again, as Scholnick argues, based my solution at some level of organicism: a point that will not escape the discerning realist or, for that matter, the discerning rationalist. PMID:1767726

  17. XML-based information system for planetary sciences

    NASA Astrophysics Data System (ADS)

    Carraro, F.; Fonte, S.; Turrini, D.

    2009-04-01

    EuroPlaNet (EPN in the following) has been developed by the planetological community under the "Sixth Framework Programme" (FP6 in the following), the European programme devoted to the improvement of the European research efforts through the creation of an internal market for science and technology. The goal of the EPN programme is the creation of a European network aimed to the diffusion of data produced by space missions dedicated to the study of the Solar System. A special place within the EPN programme is that of I.D.I.S. (Integrated and Distributed Information Service). The main goal of IDIS is to offer to the planetary science community a user-friendly access to the data and information produced by the various types of research activities, i.e. Earth-based observations, space observations, modeling, theory and laboratory experiments. During the FP6 programme IDIS development consisted in the creation of a series of thematic nodes, each of them specialized in a specific scientific domain, and a technical coordination node. The four thematic nodes are the Atmosphere node, the Plasma node, the Interiors & Surfaces node and the Small Bodies & Dust node. The main task of the nodes have been the building up of selected scientific cases related with the scientific domain of each node. The second work done by EPN nodes have been the creation of a catalogue of resources related to their main scientific theme. Both these efforts have been used as the basis for the development of the main IDIS goal, i.e. the integrated distributed service. An XML-based data model have been developed to describe resources using meta-data and to store the meta-data within an XML-based database called eXist. A search engine has been then developed in order to allow users to search resources within the database. Users can select the resource type and can insert one or more values or can choose a value among those present in a list, depending on selected resource. The system searches for all the resources containing the inserted values within the resources descriptions. An important facility of the IDIS search system is the multi-node search capability. This is due to the capacity of eXist to make queries on remote databases. This allows the system to show all resources which satisfy the search criteria on local node and to show how many resources are found on remote nodes, giving also a link to open the results page on remote nodes. During FP7 the development of the IDIS system will have the main goal to make the service Virtual Observatory compliant.

  18. XML-based approaches for the integration of heterogeneous bio-molecular data

    PubMed Central

    Mesiti, Marco; Jiménez-Ruiz, Ernesto; Sanz, Ismael; Berlanga-Llavori, Rafael; Perlasca, Paolo; Valentini, Giorgio; Manset, David

    2009-01-01

    Background The today's public database infrastructure spans a very large collection of heterogeneous biological data, opening new opportunities for molecular biology, bio-medical and bioinformatics research, but raising also new problems for their integration and computational processing. Results In this paper we survey the most interesting and novel approaches for the representation, integration and management of different kinds of biological data by exploiting XML and the related recommendations and approaches. Moreover, we present new and interesting cutting edge approaches for the appropriate management of heterogeneous biological data represented through XML. Conclusion XML has succeeded in the integration of heterogeneous biomolecular information, and has established itself as the syntactic glue for biological data sources. Nevertheless, a large variety of XML-based data formats have been proposed, thus resulting in a difficult effective integration of bioinformatics data schemes. The adoption of a few semantic-rich standard formats is urgent to achieve a seamless integration of the current biological resources. PMID:19828083

  19. A Tool Kit for Implementing XML Schema Naming and Design Rules

    SciTech Connect

    Lubell, Joshua; Kulvatunyou, Boonserm

    2005-11-01

    A tool kit being developed at the National Institute of Standards and Technology (NIST)encodes XML schema Naming and Design Rules in a computer-interpretable fashion,enabling automated rule enforcement and improving schema quality.

  20. XML based tools for assessing potential impact of advanced technology space validation

    NASA Technical Reports Server (NTRS)

    Some, Raphael R.; Weisbin, Charles

    2004-01-01

    A hierarchical XML database and related analysis tools are being developed by the New Millennium Program to provide guidance on the relative impact, to future NASA missions, of advanced technologies under consideration for developmental funding.

  1. Generic XML-based framework for metadata portals

    NASA Astrophysics Data System (ADS)

    Schindler, Uwe; Diepenbroek, Michael

    2008-12-01

    We present a generic and flexible framework for building geoscientific metadata portals independent of content standards for metadata and protocols. Data can be harvested with commonly used protocols (e.g., Open Archives Initiative Protocol for Metadata Harvesting) and metadata standards like DIF or ISO 19115. The new Java-based portal software supports any XML encoding and makes metadata searchable through Apache Lucene. Software administrators are free to define searchable fields independent of their type using XPath. In addition, by extending the full-text search engine (FTS) Apache Lucene, we have significantly improved queries for numerical and date/time ranges by supplying a new trie-based algorithm, thus, enabling high-performance space/time retrievals in FTS-based geo portals. The harvested metadata are stored in separate indexes, which makes it possible to combine these into different portals. The portal-specific Java API and web service interface is highly flexible and supports custom front-ends for users, provides automatic query completion (AJAX), and dynamic visualization with conventional mapping tools. The software has been made freely available through the open source concept.

  2. Using XML and Java for Astronomical Instrumentation Control

    NASA Technical Reports Server (NTRS)

    Ames, Troy; Koons, Lisa; Sall, Ken; Warsaw, Craig

    2000-01-01

    Traditionally, instrument command and control systems have been highly specialized, consisting mostly of custom code that is difficult to develop, maintain, and extend. Such solutions are initially very costly and are inflexible to subsequent engineering change requests, increasing software maintenance costs. Instrument description is too tightly coupled with details of implementation. NASA Goddard Space Flight Center is developing a general and highly extensible framework that applies to any kind of instrument that can be controlled by a computer. The software architecture combines the platform independent processing capabilities of Java with the power of the Extensible Markup Language (XML), a human readable and machine understandable way to describe structured data. A key aspect of the object-oriented architecture is software that is driven by an instrument description, written using the Instrument Markup Language (IML). ]ML is used to describe graphical user interfaces to control and monitor the instrument, command sets and command formats, data streams, and communication mechanisms. Although the current effort is targeted for the High-resolution Airborne Wideband Camera, a first-light instrument of the Stratospheric Observatory for Infrared Astronomy, the framework is designed to be generic and extensible so that it can be applied to any instrument.

  3. Development of a deterministic XML schema by resolving structure ambiguity of HL7 messages.

    PubMed

    Huang, Ean-Wen; Wang, Da-Wei; Liou, Der-Ming

    2005-10-01

    Health level 7 (HL7) is a standard for medical information exchange. It defines data transfers for the application systems in the healthcare environment. Alternatively, the extensible markup language (XML) is a standard for data exchange using the Internet. If exchange messages follow the content and the sequence defined by HL7 and are expressed in the XML format, the system may benefit from the advantages of both standards. In creating the XML schema, we found ambiguities in HL7 message structures that cause the XML schema to be non-deterministic. These ambiguous expressions are summarized within 12 structures and can be replaced with equivalent or similar unambiguous structures. The finite state automata are used to verify expression equivalence. Applying this schema, an XML document may eliminate redundant segment group definitions and make the structure simple and easy to reproduce. In this paper, we discuss the methods and our experience in resolving ambiguous problems in HL7 messages to generate a deterministic XML schema. PMID:15993979

  4. An adaptable XML based approach for scientific data management and integration

    NASA Astrophysics Data System (ADS)

    Wang, Fusheng; Thiel, Florian; Furrer, Daniel; Vergara-Niedermayr, Cristobal; Qin, Chen; Hackenberg, Georg; Bourgue, Pierre-Emmanuel; Kaltschmidt, David; Wang, Mo

    2008-03-01

    Increased complexity of scientific research poses new challenges to scientific data management. Meanwhile, scientific collaboration is becoming increasing important, which relies on integrating and sharing data from distributed institutions. We develop SciPort, a Web-based platform on supporting scientific data management and integration based on a central server based distributed architecture, where researchers can easily collect, publish, and share their complex scientific data across multi-institutions. SciPort provides an XML based general approach to model complex scientific data by representing them as XML documents. The documents capture not only hierarchical structured data, but also images and raw data through references. In addition, SciPort provides an XML based hierarchical organization of the overall data space to make it convenient for quick browsing. To provide generalization, schemas and hierarchies are customizable with XML-based definitions, thus it is possible to quickly adapt the system to different applications. While each institution can manage documents on a Local SciPort Server independently, selected documents can be published to a Central Server to form a global view of shared data across all sites. By storing documents in a native XML database, SciPort provides high schema extensibility and supports comprehensive queries through XQuery. By providing a unified and effective means for data modeling, data access and customization with XML, SciPort provides a flexible and powerful platform for sharing scientific data for scientific research communities, and has been successfully used in both biomedical research and clinical trials.

  5. A novel recursive Fourier transform for nonuniform sampled signals: application to heart rate variability spectrum estimation.

    PubMed

    Holland, Alexander; Aboy, Mateo

    2009-07-01

    We present a novel method to iteratively calculate discrete Fourier transforms for discrete time signals with sample time intervals that may be widely nonuniform. The proposed recursive Fourier transform (RFT) does not require interpolation of the samples to uniform time intervals, and each iterative transform update of N frequencies has computational order N. Because of the inherent non-uniformity in the time between successive heart beats, an application particularly well suited for this transform is power spectral density (PSD) estimation for heart rate variability. We compare RFT based spectrum estimation with Lomb-Scargle Transform (LST) based estimation. PSD estimation based on the LST also does not require uniform time samples, but the LST has a computational order greater than Nlog(N). We conducted an assessment study involving the analysis of quasi-stationary signals with various levels of randomly missing heart beats. Our results indicate that the RFT leads to comparable estimation performance to the LST with significantly less computational overhead and complexity for applications requiring iterative spectrum estimations. PMID:19247700

  6. Recursive time-varying filter banks for subband image coding

    NASA Technical Reports Server (NTRS)

    Smith, Mark J. T.; Chung, Wilson C.

    1992-01-01

    Filter banks and wavelet decompositions that employ recursive filters have been considered previously and are recognized for their efficiency in partitioning the frequency spectrum. This paper presents an analysis of a new infinite impulse response (IIR) filter bank in which these computationally efficient filters may be changed adaptively in response to the input. The filter bank is presented and discussed in the context of finite-support signals with the intended application in subband image coding. In the absence of quantization errors, exact reconstruction can be achieved and by the proper choice of an adaptation scheme, it is shown that IIR time-varying filter banks can yield improvement over conventional ones.

  7. Association-dissociation process with aging subunits: Recursive solution

    NASA Astrophysics Data System (ADS)

    Niedermayer, Thomas; Lipowsky, Reinhard

    2015-11-01

    The coupling of stochastic growth and shrinkage of one-dimensional structures to random aging of the constituting subunits defines the simple association-dissociation-aging process which captures the essential features of the nonequilibrium assembly of cytoskeletal filaments. Because of correlations, previously employed mean-field methods fail to correctly describe filament growth. We study an alternative formulation of the full master equation of the stochastic process. An ansatz for the steady-state solution leads to a recursion relation which allows for the calculation of all emergent quantities with increasing accuracy and in excellent agreement with stochastic simulations.

  8. Design of a recursive vector processor using polynomial splines

    NASA Technical Reports Server (NTRS)

    Kim, C. S.; Shen, C. N.

    1980-01-01

    The problem of obtaining smoothed estimates of function values, particularly their derivatives, from a finite set of inaccurate measurements is considered. A recursive two-dimensional vector processor is introduced as an approximation to the nonrecursive constrained least-squares estimation. Here, piecewise bicubic Hermite polynomials are extensively used as approximating functions, and the smoothing integral is converted to a discrete quadratic form. This makes it possible to convert the problem of fitting an approximating function to one of estimating the function values and derivatives at the nodes.

  9. Toward a parallel recursive spectral bisection mapping tool

    SciTech Connect

    Leete, C.A.; Peyton, B.W.; Sincovec, R.F.

    1993-01-01

    Recent work has shown that the recursive spectral bisection (RSB) partitioning algorithm results in good load balance and low communication costs for PDE calculations on a distributed- multiprocessor. We have enhanced a recent implementation of the method in the following three ways: we have (1) developed parallel implementation, (2) used simple proportional mapping'' ideas to handle an arbitrary number of processes, and (3) implemented a more flexible load balancing technique. Our tests indicate that the latter two enhancements are quite effective. We also report timings obtained on an Intel iPSC/860.

  10. Description of an exact, recursive method to simplify shading calculations

    NASA Astrophysics Data System (ADS)

    Nawrocki, A. D.; Kammerud, R.

    An exact, recursive method called SHADE is described which attempts to simplify shading calculations as performed by a programmable calculator or microcomputer. Preliminary applications of SHADE using a Hewlett Packard HP-41C programmable calculator are outlined. For a given solar hour, SHADE is used to compute the following quantities for overhang and side fin combinations which shade various openings: the percentage of the total area of the opening which is shaded; the shaded area itself; the cosine of the angle of incidence between the Sun and glazing surface; the direct insolation at this surface, with and without shading; and the direct solar power at this surface, with and without shading.

  11. Random recursive trees and the elephant random walk

    NASA Astrophysics Data System (ADS)

    Kürsten, Rüdiger

    2016-03-01

    One class of random walks with infinite memory, so-called elephant random walks, are simple models describing anomalous diffusion. We present a surprising connection between these models and bond percolation on random recursive trees. We use a coupling between the two models to translate results from elephant random walks to the percolation process. We calculate, besides other quantities, exact expressions for the first and the second moment of the root cluster size and of the number of nodes in child clusters of the first generation. We further introduce another model, the skew elephant random walk, and calculate the first and second moment of this process.

  12. Recursive partitioned inversion of large (1500 x 1500) symmetric matrices

    NASA Technical Reports Server (NTRS)

    Putney, B. H.; Brownd, J. E.; Gomez, R. A.

    1976-01-01

    A recursive algorithm was designed to invert large, dense, symmetric, positive definite matrices using small amounts of computer core, i.e., a small fraction of the core needed to store the complete matrix. The described algorithm is a generalized Gaussian elimination technique. Other algorithms are also discussed for the Cholesky decomposition and step inversion techniques. The purpose of the inversion algorithm is to solve large linear systems of normal equations generated by working geodetic problems. The algorithm was incorporated into a computer program called SOLVE. In the past the SOLVE program has been used in obtaining solutions published as the Goddard earth models.

  13. A recursive method for updating apple firmness prediction models based on spectral scattering images

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Multispectral scattering is effective for nondestructive prediction of fruit firmness. However, the established prediction models for multispectral scattering are variety specific and may not perform appropriately for fruit harvested from different orchards or at different times. In this research, a...

  14. Validation and Simplification of the Radiation Therapy Oncology Group Recursive Partitioning Analysis Classification for Glioblastoma

    SciTech Connect

    Li Jing; Wang Meihua; Won, Minhee; Shaw, Edward G.; Coughlin, Christopher; Curran, Walter J.; Mehta, Minesh P.

    2011-11-01

    Purpose: Previous recursive partitioning analysis (RPA) of patients with malignant glioma (glioblastoma multiforme [GBM] and anaplastic astrocytoma [AA]) produced six prognostic groups (I-VI) classified by six factors. We sought here to determine whether the classification for GBM could be improved by using an updated Radiation Therapy Oncology Group (RTOG) GBM database excluding AA and by considering additional baseline variables. Methods and Materials: The new analysis considered 42 baseline variables and 1,672 GBM patients from the expanded RTOG glioma database. Patients receiving radiation only were excluded such that all patients received radiation+carmustine. 'Radiation dose received' was replaced with 'radiation dose assigned.' The new RPA models were compared with the original model by applying them to a test dataset comprising 488 patients from six other RTOG trials. Fitness of the original and new models was evaluated using explained variation. Results: The original RPA model explained more variations in survival in the test dataset than did the new models (20% vs. 15%) and was therefore chosen for further analysis. It was reduced by combining Classes V and VI to produce three prognostic classes (Classes III, IV, and V+VI), as Classes V and VI had indistinguishable survival in the test dataset. The simplified model did not further improve performance (explained variation 18% vs. 20%) but is easier to apply because it involves only four variables: age, performance status, extent of resection, and neurologic function. Applying this simplified model to the updated GBM database resulted in three distinct classes with median survival times of 17.1, 11.2, and 7.5 months for Classes III, IV, and V+VI, respectively. Conclusions: The final model, the simplified original RPA model combining Classes V and VI, resulted in three distinct prognostic groups defined by age, performance status, extent of resection, and neurologic function. This classification will be used in future RTOG GBM trials.

  15. Value of XML in the implementation of clinical practice guidelines--the issue of content retrieval and presentation.

    PubMed

    Hoelzer, S; Schweiger, R K; Boettcher, H A; Tafazzoli, A G; Dudeck, J

    2001-01-01

    The purpose of guidelines in clinical practice is to improve the effectiveness and efficiency of clinical care. It is known that nationally or internationally produced guidelines which, in particular, do not involve medical processes at the time of consultation, do not take local factors into account, and have no consistent implementation strategy, have limited impact in changing either the behaviour of physicians, or patterns of care. The literature provides evidence for the effectiveness of computerization of CPGs for increasing compliance and improving patient outcomes. Probably the most effective concepts are knowledge-based functions for decision support or monitoring that are integrated in clinical information systems. This approach is mostly restricted by the effort required for development and maintenance of the information systems and the limited number of implemented medical rules. Most of the guidelines are text-based, and are primarily published in medical journals and posted on the internet. However, internet-published guidelines have little impact on the behaviour of physicians. It can be difficult and time-consuming to browse the internet to find (a) the correct guidelines to an existing diagnosis and (b) and adequate recommendation for a specific clinical problem. Our objective is to provide a web-based guideline service that takes as input clinical data on a particular patient and returns as output a customizable set of recommendations regarding diagnosis and treatment. Information in healthcare is to a very large extent transmitted and stored as unstructured or slightly structured text such as discharge letters, reports, forms, etc. The same applies for facilities containing medical information resources for clinical purposes and research such as text books, articles, guidelines, etc. Physicians are used to obtaining information from text-based sources. Since most guidelines are text-based, it would be practical to use a document-based solution that preserves the original cohesiveness. The lack of structure limits the automatic identification and extraction of the information contained in these resources. For this reason, we have chosen a document-based approach using eXtensible Markup Language (XML) with its schema definition and related technologies. XML empowers the applications for in-context searching. In addition it allows the same content to be represented in different ways. Our XML reference clinical data model for guidelines has been realized with the XML schema definition. The schema is used for structuring new text-based guidelines and updating existing documents. It is also used to establish search strategies on the document base. We hypothesize that enabling the physicians to query the available CPGs easily, and to get access to selected and specific information at the point of care will foster increased use. Based on current evidence we are confident that it will have substantial impact on the care provided, and will improve health outcomes. PMID:11560293

  16. Adaptive model reduction for continuous systems via recursive rational interpolation

    NASA Technical Reports Server (NTRS)

    Lilly, John H.

    1994-01-01

    A method for adaptive identification of reduced-order models for continuous stable SISO and MIMO plants is presented. The method recursively finds a model whose transfer function (matrix) matches that of the plant on a set of frequencies chosen by the designer. The algorithm utilizes the Moving Discrete Fourier Transform (MDFT) to continuously monitor the frequency-domain profile of the system input and output signals. The MDFT is an efficient method of monitoring discrete points in the frequency domain of an evolving function of time. The model parameters are estimated from MDFT data using standard recursive parameter estimation techniques. The algorithm has been shown in simulations to be quite robust to additive noise in the inputs and outputs. A significant advantage of the method is that it enables a type of on-line model validation. This is accomplished by simultaneously identifying a number of models and comparing each with the plant in the frequency domain. Simulations of the method applied to an 8th-order SISO plant and a 10-state 2-input 2-output plant are presented. An example of on-line model validation applied to the SISO plant is also presented.

  17. REQUEST: A Recursive QUEST Algorithm for Sequential Attitude Determination

    NASA Technical Reports Server (NTRS)

    Bar-Itzhack, Itzhack Y.

    1996-01-01

    In order to find the attitude of a spacecraft with respect to a reference coordinate system, vector measurements are taken. The vectors are pairs of measurements of the same generalized vector, taken in the spacecraft body coordinates, as well as in the reference coordinate system. We are interested in finding the best estimate of the transformation between these coordinate system.s The algorithm called QUEST yields that estimate where attitude is expressed by a quarternion. Quest is an efficient algorithm which provides a least squares fit of the quaternion of rotation to the vector measurements. Quest however, is a single time point (single frame) batch algorithm, thus measurements that were taken at previous time points are discarded. The algorithm presented in this work provides a recursive routine which considers all past measurements. The algorithm is based on on the fact that the, so called, K matrix, one of whose eigenvectors is the sought quaternion, is linerly related to the measured pairs, and on the ability to propagate K. The extraction of the appropriate eigenvector is done according to the classical QUEST algorithm. This stage, however, can be eliminated, and the computation simplified, if a standard eigenvalue-eigenvector solver algorithm is used. The development of the recursive algorithm is presented and illustrated via a numerical example.

  18. Optimal Recursive Digital Filters for Active Bending Stabilization

    NASA Technical Reports Server (NTRS)

    Orr, Jeb S.

    2013-01-01

    In the design of flight control systems for large flexible boosters, it is common practice to utilize active feedback control of the first lateral structural bending mode so as to suppress transients and reduce gust loading. Typically, active stabilization or phase stabilization is achieved by carefully shaping the loop transfer function in the frequency domain via the use of compensating filters combined with the frequency response characteristics of the nozzle/actuator system. In this paper we present a new approach for parameterizing and determining optimal low-order recursive linear digital filters so as to satisfy phase shaping constraints for bending and sloshing dynamics while simultaneously maximizing attenuation in other frequency bands of interest, e.g. near higher frequency parasitic structural modes. By parameterizing the filter directly in the z-plane with certain restrictions, the search space of candidate filter designs that satisfy the constraints is restricted to stable, minimum phase recursive low-pass filters with well-conditioned coefficients. Combined with optimal output feedback blending from multiple rate gyros, the present approach enables rapid and robust parametrization of autopilot bending filters to attain flight control performance objectives. Numerical results are presented that illustrate the application of the present technique to the development of rate gyro filters for an exploration-class multi-engined space launch vehicle.

  19. Efficient Execution of Recursive Programs on Commodity Vector Hardware

    SciTech Connect

    Ren, Bin; Jo, Youngjoon; Krishnamoorthy, Sriram; Agrawal, Kunal; Kulkarni, Milind

    2015-06-13

    The pursuit of computational efficiency has led to the proliferation of throughput-oriented hardware, from GPUs to increasingly-wide vector units on commodity processors and accelerators. This hardware is designed to efficiently execute data-parallel computations in a vectorized manner. However, many algorithms are more naturally expressed as divide-and-conquer, recursive, task-parallel computations; in the absence of data parallelism, it seems that such algorithms are not well-suited to throughput-oriented architectures. This paper presents a set of novel code transformations that expose the data-parallelism latent in recursive, task-parallel programs. These transformations facilitate straightforward vectorization of task-parallel programs on commodity hardware. We also present scheduling policies that maintain high utilization of vector resources while limiting space usage. Across several task-parallel benchmarks, we demonstrate both efficient vector resource utilization and substantial speedup on chips using Intel's SSE4.2 vector units as well as accelerators using Intel's AVX512 units.

  20. Change-point detection for recursive Bayesian geoacoustic inversions.

    PubMed

    Tan, Bien Aik; Gerstoft, Peter; Yardim, Caglar; Hodgkiss, William S

    2015-04-01

    In order to carry out geoacoustic inversion in low signal-to-noise ratio (SNR) conditions, extended duration observations coupled with source and/or receiver motion may be necessary. As a result, change in the underlying model parameters due to time or space is anticipated. In this paper, an inversion method is proposed for cases when the model parameters change abruptly or slowly. A model parameter change-point detection method is developed to detect the change in the model parameters using the importance samples and corresponding weights that are already available from the recursive Bayesian inversion. If the model parameters change abruptly, a change-point will be detected and the inversion will restart with the pulse measurement after the change-point. If the model parameters change gradually, the inversion (based on constant model parameters) may proceed until the accumulated model parameter mismatch is significant and triggers the detection of a change-point. These change-point detections form the heuristics for controlling the coherent integration time in recursive Bayesian inversion. The method is demonstrated in simulation with parameters corresponding to the low SNR, 100-900 Hz linear frequency modulation pulses observed in the Shallow Water 2006 experiment [Tan, Gerstoft, Yardim, and Hodgkiss, J. Acoust. Soc. Am. 136, 1187-1198 (2014)]. PMID:25920847

  1. Recursive mentalizing and common knowledge in the bystander effect.

    PubMed

    Thomas, Kyle A; De Freitas, Julian; DeScioli, Peter; Pinker, Steven

    2016-05-01

    The more potential helpers there are, the less likely any individual is to help. A traditional explanation for this bystander effect is that responsibility diffuses across the multiple bystanders, diluting the responsibility of each. We investigate an alternative, which combines the volunteer's dilemma (each bystander is best off if another responds) with recursive theory of mind (each infers what the others know about what he knows) to predict that actors will strategically shirk when they think others feel compelled to help. In 3 experiments, participants responded to a (fictional) person who needed help from at least 1 volunteer. Participants were in groups of 2 or 5 and had varying information about whether other group members knew that help was needed. As predicted, people's decision to help zigzagged with the depth of their asymmetric, recursive knowledge (e.g., "John knows that Michael knows that John knows help is needed"), and replicated the classic bystander effect when they had common knowledge (everyone knowing what everyone knows). The results demonstrate that the bystander effect may result not from a mere diffusion of responsibility but specifically from actors' strategic computations. (PsycINFO Database Record PMID:26913616

  2. Quicksort: Combining Concurrency, Recursion, and Mutable Data Structures

    NASA Astrophysics Data System (ADS)

    Kitchin, David; Quark, Adrian; Misra, Jayadev

    Quicksort (Commun. ACM 4(7):321-322, 1961) remains one of the most studied algorithms in computer science. It is important not only as a practical sorting method, but also as a splendid teaching aid for introducing recursion and systematic algorithm development. The algorithm has been studied extensively; so, it is natural to assume that everything that needs to be said about it has already been said. Yet, in attempting to code it using a recent programming language of our design, we discovered that its structure is more clearly expressed as a concurrent program that manipulates a shared mutable store, without any locking or explicit synchronization. In this paper, we describe the essential aspects of our programming language Orc (Proceedings of FMOODS/FORTE, vol. 5522 of LNCS, pp. 1-25. Springer 2009), show a number of examples that combine its features in various forms, and then develop a concise description of Quicksort. We hope to highlight the importance of including concurrency, recursion and mutability within a single theory.

  3. Parsimonious extreme learning machine using recursive orthogonal least squares.

    PubMed

    Wang, Ning; Er, Meng Joo; Han, Min

    2014-10-01

    Novel constructive and destructive parsimonious extreme learning machines (CP- and DP-ELM) are proposed in this paper. By virtue of the proposed ELMs, parsimonious structure and excellent generalization of multiinput-multioutput single hidden-layer feedforward networks (SLFNs) are obtained. The proposed ELMs are developed by innovative decomposition of the recursive orthogonal least squares procedure into sequential partial orthogonalization (SPO). The salient features of the proposed approaches are as follows: 1) Initial hidden nodes are randomly generated by the ELM methodology and recursively orthogonalized into an upper triangular matrix with dramatic reduction in matrix size; 2) the constructive SPO in the CP-ELM focuses on the partial matrix with the subcolumn of the selected regressor including nonzeros as the first column while the destructive SPO in the DP-ELM operates on the partial matrix including elements determined by the removed regressor; 3) termination criteria for CP- and DP-ELM are simplified by the additional residual error reduction method; and 4) the output weights of the SLFN need not be solved in the model selection procedure and is derived from the final upper triangular equation by backward substitution. Both single- and multi-output real-world regression data sets are used to verify the effectiveness and superiority of the CP- and DP-ELM in terms of parsimonious architecture and generalization accuracy. Innovative applications to nonlinear time-series modeling demonstrate superior identification results. PMID:25291736

  4. Exact substructuring in recursive Newton's method for solving transcendental eigenproblems

    NASA Astrophysics Data System (ADS)

    Djoudi, M. S.; Kennedy, D.; Williams, F. W.; Yuan, S.; Ye, K.

    2005-02-01

    The well-established Wittrick-Williams (W-W) algorithm guarantees accurate convergence on natural frequencies or critical buckling loads of structures in the transcendental eigenproblems arising from the use of exact member stiffnesses, i.e. dynamic member stiffnesses for vibration. The associated mode calculations had no such guarantee until they were recently greatly improved by solving the transcendental eigenproblem exactly, by reducing it to a generalised linear eigenproblem which is solved by a guided recursive Newton method involving inverse iteration. The present paper demonstrates the benefits of using frequency squared, rather than frequency, as the eigenparameter. Next, exact substructuring is introduced into the recursive Newton method, with accuracy retained because the inverse iteration includes the substructure nodes. If member fixed end eigenvalues lie close to the sought eigenvalue they can cause inaccuracy, or even wrong results, and so they are removed efficiently by inserting interior nodes to create simple substructures. Numerical results for a moderately large structure show that exact substructuring reduces the transcendental eigensolution time without reducing accuracy. Simpler examples designed to be numerically ill conditioned in the absence of interior nodes show that such ill conditioning is almost removed by inserting interior nodes within substructures. Exact substructuring is also applicable to other inverse iteration and W-W algorithm-based methods.

  5. James Webb Space Telescope XML Database: From the Beginning to Today

    NASA Technical Reports Server (NTRS)

    Gal-Edd, Jonathan; Fatig, Curtis C.

    2005-01-01

    The James Webb Space Telescope (JWST) Project has been defining, developing, and exercising the use of a common eXtensible Markup Language (XML) for the command and telemetry (C&T) database structure. JWST is the first large NASA space mission to use XML for databases. The JWST project started developing the concepts for the C&T database in 2002. The database will need to last at least 20 years since it will be used beginning with flight software development, continuing through Observatory integration and test (I&T) and through operations. Also, a database tool kit has been provided to the 18 various flight software development laboratories located in the United States, Europe, and Canada that allows the local users to create their own databases. Recently the JWST Project has been working with the Jet Propulsion Laboratory (JPL) and Object Management Group (OMG) XML Telemetry and Command Exchange (XTCE) personnel to provide all the information needed by JWST and JPL for exchanging database information using a XML standard structure. The lack of standardization requires custom ingest scripts for each ground system segment, increasing the cost of the total system. Providing a non-proprietary standard of the telemetry and command database definition formation will allow dissimilar systems to communicate without the need for expensive mission specific database tools and testing of the systems after the database translation. The various ground system components that would benefit from a standardized database are the telemetry and command systems, archives, simulators, and trending tools. JWST has exchanged the XML database with the Eclipse, EPOCH, ASIST ground systems, Portable spacecraft simulator (PSS), a front-end system, and Integrated Trending and Plotting System (ITPS) successfully. This paper will discuss how JWST decided to use XML, the barriers to a new concept, experiences utilizing the XML structure, exchanging databases with other users, and issues that have been experienced in creating databases for the C&T system.

  6. Recursive Ant Colony Global Optimization: a new technique for the inversion of geophysical data

    NASA Astrophysics Data System (ADS)

    Gupta, D. K.; Gupta, J. P.; Arora, Y.; Singh, U. K.

    2011-12-01

    We present a new method called Recursive Ant Colony Global Optimization (RACO) technique, a modified form of general ACO, which can be used to find the best solutions to inversion problems in geophysics. RACO simulates the social behaviour of ants to find the best path between the nest and the food source. A new term depth has been introduced, which controls the extent of recursion. A selective number of cities get qualified for the successive depth. The results of one depth are used to construct the models for the next depth and the range of values for each of the parameters is reduced without any change to the number of models. The three additional steps performed after each depth, are the pheromone tracking, pheromone updating and city selection. One of the advantages of RACO over ACO is that if a problem has multiple solutions, then pheromone accumulation will take place at more than one city thereby leading to formation of multiple nested ACO loops within the ACO loop of the previous depth. Also, while the convergence of ACO is almost linear, RACO shows exponential convergence and hence is faster than the ACO. RACO proves better over some other global optimization techniques, as it does not require any initial values to be assigned to the parameters function. The method has been tested on some mathematical functions, synthetic self-potential (SP) and synthetic gravity data. The obtained results reveal the efficiency and practicability of the method. The method is found to be efficient enough to solve the problems of SP and gravity anomalies due to a horizontal cylinder, a sphere, an inclined sheet and multiple idealized bodies buried inside the earth. These anomalies with and without noise were inverted using the RACO algorithm. The obtained results were compared with those obtained from the conventional methods and it was found that RACO results are more accurate. Finally this optimization technique was applied to real field data collected over the Surda Valley region of India and we get encouraging results that is in good agreement with the available results.

  7. Round-off error propagation in four generally applicable, recursive, least-squares-estimation schemes

    NASA Technical Reports Server (NTRS)

    Verhaegen, M. H.

    1987-01-01

    The numerical robustness of four generally applicable, recursive, least-squares-estimation schemes is analyzed by means of a theoretical round-off propagation study. This study highlights a number of practical, interesting insights of widely used recursive least-squares schemes. These insights have been confirmed in an experimental study as well.

  8. Generalized Path Analysis and Generalized Simultaneous Equations Model for Recursive Systems with Responses of Mixed Types

    ERIC Educational Resources Information Center

    Tsai, Tien-Lung; Shau, Wen-Yi; Hu, Fu-Chang

    2006-01-01

    This article generalizes linear path analysis (PA) and simultaneous equations models (SiEM) to deal with mixed responses of different types in a recursive or triangular system. An efficient instrumental variable (IV) method for estimating the structural coefficients of a 2-equation partially recursive generalized path analysis (GPA) model and…

  9. XML at the ADC: Steps to a Next Generation Data Archive

    NASA Astrophysics Data System (ADS)

    Shaya, E.; Blackwell, J.; Gass, J.; Oliversen, N.; Schneider, G.; Thomas, B.; Cheung, C.; White, R. A.

    1999-05-01

    The eXtensible Markup Language (XML) is a document markup language that allows users to specify their own tags, to create hierarchical structures to qualify their data, and to support automatic checking of documents for structural validity. It is being intensively supported by nearly every major corporate software developer. Under the funds of a NASA AISRP proposal, the Astronomical Data Center (ADC, http://adc.gsfc.nasa.gov) is developing an infrastructure for importation, enhancement, and distribution of data and metadata using XML as the document markup language. We discuss the preliminary Document Type Definition (DTD, at http://adc.gsfc.nasa.gov/xml) which specifies the elements and their attributes in our metadata documents. This attempts to define both the metadata of an astronomical catalog and the `header' information of an astronomical table. In addition, we give an overview of the planned flow of data through automated pipelines from authors and journal presses into our XML archive and retrieval through the web via the XML-QL Query Language and eXtensible Style Language (XSL) scripts. When completed, the catalogs and journal tables at the ADC will be tightly hyperlinked to enhance data discovery. In addition one will be able to search on fragmentary information. For instance, one could query for a table by entering that the second author is so-and-so or that the third author is at such-and-such institution.

  10. Alternatives to relational database: comparison of NoSQL and XML approaches for clinical data storage.

    PubMed

    Lee, Ken Ka-Yin; Tang, Wai-Choi; Choi, Kup-Sze

    2013-04-01

    Clinical data are dynamic in nature, often arranged hierarchically and stored as free text and numbers. Effective management of clinical data and the transformation of the data into structured format for data analysis are therefore challenging issues in electronic health records development. Despite the popularity of relational databases, the scalability of the NoSQL database model and the document-centric data structure of XML databases appear to be promising features for effective clinical data management. In this paper, three database approaches--NoSQL, XML-enabled and native XML--are investigated to evaluate their suitability for structured clinical data. The database query performance is reported, together with our experience in the databases development. The results show that NoSQL database is the best choice for query speed, whereas XML databases are advantageous in terms of scalability, flexibility and extensibility, which are essential to cope with the characteristics of clinical data. While NoSQL and XML technologies are relatively new compared to the conventional relational database, both of them demonstrate potential to become a key database technology for clinical data management as the technology further advances. PMID:23177219

  11. Recursive dynamics for flexible multibody systems using spatial operators

    NASA Technical Reports Server (NTRS)

    Jain, A.; Rodriguez, G.

    1990-01-01

    Due to their structural flexibility, spacecraft and space manipulators are multibody systems with complex dynamics and possess a large number of degrees of freedom. Here the spatial operator algebra methodology is used to develop a new dynamics formulation and spatially recursive algorithms for such flexible multibody systems. A key feature of the formulation is that the operator description of the flexible system dynamics is identical in form to the corresponding operator description of the dynamics of rigid multibody systems. A significant advantage of this unifying approach is that it allows ideas and techniques for rigid multibody systems to be easily applied to flexible multibody systems. The algorithms use standard finite-element and assumed modes models for the individual body deformation. A Newton-Euler Operator Factorization of the mass matrix of the multibody system is first developed. It forms the basis for recursive algorithms such as for the inverse dynamics, the computation of the mass matrix, and the composite body forward dynamics for the system. Subsequently, an alternative Innovations Operator Factorization of the mass matrix, each of whose factors is invertible, is developed. It leads to an operator expression for the inverse of the mass matrix, and forms the basis for the recursive articulated body forward dynamics algorithm for the flexible multibody system. For simplicity, most of the development here focuses on serial chain multibody systems. However, extensions of the algorithms to general topology flexible multibody systems are described. While the computational cost of the algorithms depends on factors such as the topology and the amount of flexibility in the multibody system, in general, it appears that in contrast to the rigid multibody case, the articulated body forward dynamics algorithm is the more efficient algorithm for flexible multibody systems containing even a small number of flexible bodies. The variety of algorithms described here permits a user to choose the algorithm which is optimal for the multibody system at hand. The availability of a number of algorithms is even more important for real-time applications, where implementation on parallel processors or custom computing hardware is often necessary to maximize speed.

  12. Recursive Subsystems in Aphasia and Alzheimer's Disease: Case Studies in Syntax and Theory of Mind.

    PubMed

    Bánréti, Zoltán; Hoffmann, Ildikó; Vincze, Veronika

    2016-01-01

    The relationship between recursive sentence embedding and theory-of-mind (ToM) inference is investigated in three persons with Broca's aphasia, two persons with Wernicke's aphasia, and six persons with mild and moderate Alzheimer's disease (AD). We asked questions of four types about photographs of various real-life situations. Type 4 questions asked participants about intentions, thoughts, or utterances of the characters in the pictures ("What may X be thinking/asking Y to do?"). The expected answers typically involved subordinate clauses introduced by conjunctions or direct quotations of the characters' utterances. Broca's aphasics did not produce answers with recursive sentence embedding. Rather, they projected themselves into the characters' mental states and gave direct answers in the first person singular, with relevant ToM content. We call such replies "situative statements." Where the question concerned the mental state of the character but did not require an answer with sentence embedding ("What does X hate?"), aphasics gave descriptive answers rather than situative statements. Most replies given by persons with AD to Type 4 questions were grammatical instances of recursive sentence embedding. They also gave a few situative statements but the ToM content of these was irrelevant. In more than one third of their well-formed sentence embeddings, too, they conveyed irrelevant ToM contents. Persons with moderate AD were unable to pass secondary false belief tests. The results reveal double dissociation: Broca's aphasics are unable to access recursive sentence embedding but they can make appropriate ToM inferences; moderate AD persons make the wrong ToM inferences but they are able to access recursive sentence embedding. The double dissociation may be relevant for the nature of the relationship between the two recursive capacities. Broca's aphasics compensated for the lack of recursive sentence embedding by recursive ToM reasoning represented in very simple syntactic forms: they used one recursive subsystem to stand in for another recursive subsystem. PMID:27064887

  13. Email Updates

    MedlinePlus

    ... unsubscribe. Prevent MedlinePlus emails from being marked as "spam" or "junk" To ensure that MedlinePlus email updates ... com to your email address book, adjust your spam settings, or follow the instructions from your email ...

  14. Adaptive control of large space structures using recursive lattice filters

    NASA Technical Reports Server (NTRS)

    Goglia, G. L.

    1985-01-01

    The use of recursive lattice filters for identification and adaptive control of large space structures was studied. Lattice filters are used widely in the areas of speech and signal processing. Herein, they are used to identify the structural dynamics model of the flexible structures. This identified model is then used for adaptive control. Before the identified model and control laws are integrated, the identified model is passed through a series of validation procedures and only when the model passes these validation procedures control is engaged. This type of validation scheme prevents instability when the overall loop is closed. The results obtained from simulation were compared to those obtained from experiments. In this regard, the flexible beam and grid apparatus at the Aerospace Control Research Lab (ACRL) of NASA Langley Research Center were used as the principal candidates for carrying out the above tasks. Another important area of research, namely that of robust controller synthesis, was investigated using frequency domain multivariable controller synthesis methods.

  15. Structure damage detection based on random forest recursive feature elimination

    NASA Astrophysics Data System (ADS)

    Zhou, Qifeng; Zhou, Hao; Zhou, Qingqing; Yang, Fan; Luo, Linkai

    2014-05-01

    Feature extraction is a key former step in structural damage detection. In this paper, a structural damage detection method based on wavelet packet decomposition (WPD) and random forest recursive feature elimination (RF-RFE) is proposed. In order to gain the most effective feature subset and to improve the identification accuracy a two-stage feature selection method is adopted after WPD. First, the damage features are sorted according to original random forest variable importance analysis. Second, using RF-RFE to eliminate the least important feature and reorder the feature list each time, then get the new feature importance sequence. Finally, k-nearest neighbor (KNN) algorithm, as a benchmark classifier, is used to evaluate the extracted feature subset. A four-storey steel shear building model is chosen as an example in method verification. The experimental results show that using the fewer features got from proposed method can achieve higher identification accuracy and reduce the detection time cost.

  16. Recursive Bayesian estimation of regularized and irregular quasar light curves

    NASA Astrophysics Data System (ADS)

    Hanif, A.; Protopapas, P.

    2015-03-01

    We investigate the efficacy of recursive Bayesian estimation of regularized and irregular astrophysical time series using particle filters to understand latent dynamics. We begin by regularizing a MACHO (massive compact halo object) quasar light curve using linear interpolation techniques. This is subsequently modelled using a variety of autoregressive and autoregressive-integrated moving average models. We find that we can learn regularized astrophysical time series using particle filters. Motivated by this result, we proceed by working on raw, irregular light curves. Accurately modelling the underlying dynamics as a continuous autoregressive stochastic process, calibrated using an MCMC we find that the scale variable, ?, is in fact first-order stable across 55 MACHO quasar light curves and thus not correlated with the black hole mass. We show that particle filters can be used to learn regularized and irregular astrophysical light curves. These results can be used to inform classification systems of stellar type and further study variability characteristics of quasars.

  17. Chaotic spin-spin entanglement on a recursive lattice

    NASA Astrophysics Data System (ADS)

    Chakhmakhchyan, Levon; Guérin, Stéphane; Leroy, Claude

    2015-08-01

    We propose an exactly solvable multisite interaction spin-1/2 Ising-Heisenberg model on a triangulated Husimi lattice for the rigorous studies of chaotic entanglement. By making use of the generalized star-triangle transformation, we map the initial model onto an effective Ising one on a Husimi lattice, which we solve then exactly by applying the recursive method. Expressing the entanglement of the Heisenberg spins, that we quantify by means of the concurrence, in terms of the magnetic quantities of the system, we demonstrate its bifurcation and chaotic behavior. Furthermore, we show that the underlying chaos may slightly enhance the amount of the entanglement and present on the phase diagram the transition lines from the uniform to periodic and from the periodic to chaotic regimes.

  18. Weighting the recursive spectral bisection algorithm for unstructured grids

    SciTech Connect

    Allen, D.; Dube, E.; Rodrigue, G.

    1995-12-01

    When solving partial differential equations numerically on parallel computers it is desirable to decompose the domain on which we are solving the equations in such a way as to equalize the workload among the processors while minimizing the communication between them. This is equivalent to finding a partition of the graph representing the calculation into equal subgraphs cutting as few edges as possible. One such algorithm in use is the recursive spectral bisection algorithm (RSB), (1) generate the Laplace matrix for the graph representing the dependencies between the calculation taking place at each vertex of the graph. (2) Compute the eigenvector corresponding to the smallest non-zero eigen-value, called the Fiedler vector. (3) Sort the vertices according to size of entries in the Fiedler vector. (4) Assign half the vertices into each subgraph.

  19. Source localization using recursively applied and projected (RAP) MUSIC

    SciTech Connect

    Mosher, J.C.; Leahy, R.M.

    1998-03-01

    A new method for source localization is described that is based on a modification of the well known multiple signal classification (MUSIC) algorithm. In classical MUSIC, the array manifold vector is projected onto an estimate of the signal subspace, but errors in the estimate can make location of multiple sources difficult. Recursively applied and projected (RAP) MUSIC uses each successively located source to form an intermediate array gain matrix, and projects both the array manifold and the signal subspace estimate into its orthogonal complement. The MUSIC projection is then performed in this reduced subspace. Using the metric of principal angles, the authors describe a general form of the RAP-MUSIC algorithm for the case of diversely polarized sources. Through a uniform linear array simulation, the authors demonstrate the improved Monte Carlo performance of RAP-MUSIC relative to MUSIC and two other sequential subspace methods, S and IES-MUSIC.

  20. Recursive Newton-Euler formulation of manipulator dynamics

    NASA Technical Reports Server (NTRS)

    Nasser, M. G.

    1989-01-01

    A recursive Newton-Euler procedure is presented for the formulation and solution of manipulator dynamical equations. The procedure includes rotational and translational joints and a topological tree. This model was verified analytically using a planar two-link manipulator. Also, the model was tested numerically against the Walker-Orin model using the Shuttle Remote Manipulator System data. The hinge accelerations obtained from both models were identical. The computational requirements of the model vary linearly with the number of joints. The computational efficiency of this method exceeds that of Walker-Orin methods. This procedure may be viewed as a considerable generalization of Armstrong's method. A six-by-six formulation is adopted which enhances both the computational efficiency and simplicity of the model.

  1. Adaptive control of large space structures using recursive lattice filters

    NASA Technical Reports Server (NTRS)

    Sundararajan, N.; Goglia, G. L.

    1985-01-01

    The use of recursive lattice filters for identification and adaptive control of large space structures is studied. Lattice filters were used to identify the structural dynamics model of the flexible structures. This identification model is then used for adaptive control. Before the identified model and control laws are integrated, the identified model is passed through a series of validation procedures and only when the model passes these validation procedures is control engaged. This type of validation scheme prevents instability when the overall loop is closed. Another important area of research, namely that of robust controller synthesis, was investigated using frequency domain multivariable controller synthesis methods. The method uses the Linear Quadratic Guassian/Loop Transfer Recovery (LQG/LTR) approach to ensure stability against unmodeled higher frequency modes and achieves the desired performance.

  2. Multiangle dynamic light scattering analysis using an improved recursion algorithm

    NASA Astrophysics Data System (ADS)

    Li, Lei; Li, Wei; Wang, Wanyan; Zeng, Xianjiang; Chen, Junyao; Du, Peng; Yang, Kecheng

    2015-10-01

    Multiangle dynamic light scattering (MDLS) compensates for the low information in a single-angle dynamic light scattering (DLS) measurement by combining the light intensity autocorrelation functions from a number of measurement angles. Reliable estimation of PSD from MDLS measurements requires accurate determination of the weighting coefficients and an appropriate inversion method. We propose the Recursion Nonnegative Phillips-Twomey (RNNPT) algorithm, which is insensitive to the noise of correlation function data, for PSD reconstruction from MDLS measurements. The procedure includes two main steps: 1) the calculation of the weighting coefficients by the recursion method, and 2) the PSD estimation through the RNNPT algorithm. And we obtained suitable regularization parameters for the algorithm by using MR-L-curve since the overall computational cost of this method is sensibly less than that of the L-curve for large problems. Furthermore, convergence behavior of the MR-L-curve method is in general superior to that of the L-curve method and the error of MR-L-curve method is monotone decreasing. First, the method was evaluated on simulated unimodal lognormal PSDs and multimodal lognormal PSDs. For comparison, reconstruction results got by a classical regularization method were included. Then, to further study the stability and sensitivity of the proposed method, all examples were analyzed using correlation function data with different levels of noise. The simulated results proved that RNNPT method yields more accurate results in the determination of PSDs from MDLS than those obtained with the classical regulation method for both unimodal and multimodal PSDs.

  3. Learning to play Go using recursive neural networks.

    PubMed

    Wu, Lin; Baldi, Pierre

    2008-11-01

    Go is an ancient board game that poses unique opportunities and challenges for artificial intelligence. Currently, there are no computer Go programs that can play at the level of a good human player. However, the emergence of large repositories of games is opening the door for new machine learning approaches to address this challenge. Here we develop a machine learning approach to Go, and related board games, focusing primarily on the problem of learning a good evaluation function in a scalable way. Scalability is essential at multiple levels, from the library of local tactical patterns, to the integration of patterns across the board, to the size of the board itself. The system we propose is capable of automatically learning the propensity of local patterns from a library of games. Propensity and other local tactical information are fed into recursive neural networks, derived from a probabilistic Bayesian network architecture. The recursive neural networks in turn integrate local information across the board in all four cardinal directions and produce local outputs that represent local territory ownership probabilities. The aggregation of these probabilities provides an effective strategic evaluation function that is an estimate of the expected area at the end, or at various other stages, of the game. Local area targets for training can be derived from datasets of games played by human players. In this approach, while requiring a learning time proportional to N(4), skills learned on a board of size N(2) can easily be transferred to boards of other sizes. A system trained using only 9 x 9 amateur game data performs surprisingly well on a test set derived from 19 x 19 professional game data. Possible directions for further improvements are briefly discussed. PMID:18420381

  4. Recursive least square vehicle mass estimation based on acceleration partition

    NASA Astrophysics Data System (ADS)

    Feng, Yuan; Xiong, Lu; Yu, Zhuoping; Qu, Tong

    2014-05-01

    Vehicle mass is an important parameter in vehicle dynamics control systems. Although many algorithms have been developed for the estimation of mass, none of them have yet taken into account the different types of resistance that occur under different conditions. This paper proposes a vehicle mass estimator. The estimator incorporates road gradient information in the longitudinal accelerometer signal, and it removes the road grade from the longitudinal dynamics of the vehicle. Then, two different recursive least square method (RLSM) schemes are proposed to estimate the driving resistance and the mass independently based on the acceleration partition under different conditions. A 6 DOF dynamic model of four In-wheel Motor Vehicle is built to assist in the design of the algorithm and in the setting of the parameters. The acceleration limits are determined to not only reduce the estimated error but also ensure enough data for the resistance estimation and mass estimation in some critical situations. The modification of the algorithm is also discussed to improve the result of the mass estimation. Experiment data on a sphalt road, plastic runway, and gravel road and on sloping roads are used to validate the estimation algorithm. The adaptability of the algorithm is improved by using data collected under several critical operating conditions. The experimental results show the error of the estimation process to be within 2.6%, which indicates that the algorithm can estimate mass with great accuracy regardless of the road surface and gradient changes and that it may be valuable in engineering applications. This paper proposes a recursive least square vehicle mass estimation method based on acceleration partition.

  5. Interactive, Secure Web-enabled Aircraft Engine Simulation Using XML Databinding Integration

    NASA Technical Reports Server (NTRS)

    Lin, Risheng; Afjeh, Abdollah A.

    2003-01-01

    This paper discusses the detailed design of an XML databinding framework for aircraft engine simulation. The framework provides an object interface to access and use engine data. while at the same time preserving the meaning of the original data. The Language independent representation of engine component data enables users to move around XML data using HTTP through disparate networks. The application of this framework is demonstrated via a web-based turbofan propulsion system simulation using the World Wide Web (WWW). A Java Servlet based web component architecture is used for rendering XML engine data into HTML format and dealing with input events from the user, which allows users to interact with simulation data from a web browser. The simulation data can also be saved to a local disk for archiving or to restart the simulation at a later time.

  6. Study of WebGIS based on XML database and SVG

    NASA Astrophysics Data System (ADS)

    Lian, Jianbo; Sun, Bianfu

    2011-02-01

    With the fast development of WebGIS, data fast transmission and browsing the electronic map by way of vector in WebGIS become the focus of present WebGIS research. This paper proposes a method of establishing a Web-based system that can visualize geographic data. In terms of the graphics format, our proposed system uses Scalable Vector Graphics (SVG) that XML-based 2D vector graphics, and geographic data stored on Web Server with XML database. SVG is a language for describing two-dimensional graphics based on XML, and it can be used to describe the point, line and polygon. The schema has a small amount of data transferred, faster performance, low development costs and so on.

  7. Standardization of XML Database Exchanges and the James Webb Space Telescope Experience

    NASA Technical Reports Server (NTRS)

    Gal-Edd, Jonathan; Detter, Ryan; Jones, Ron; Fatig, Curtis C.

    2007-01-01

    Personnel from the National Aeronautics and Space Administration (NASA) James Webb Space Telescope (JWST) Project have been working with various standard communities such the Object Management Group (OMG) and the Consultative Committee for Space Data Systems (CCSDS) to assist in the definition of a common extensible Markup Language (XML) for database exchange format. The CCSDS and OMG standards are intended for the exchange of core command and telemetry information, not for all database information needed to exercise a NASA space mission. The mission-specific database, containing all the information needed for a space mission, is translated from/to the standard using a translator. The standard is meant to provide a system that encompasses 90% of the information needed for command and telemetry processing. This paper will discuss standardization of the XML database exchange format, tools used, and the JWST experience, as well as future work with XML standard groups both commercial and government.

  8. Research on heterogeneous data query and sharing mode based on XML

    NASA Astrophysics Data System (ADS)

    Sang, Yaqun; Mu, Qi

    2013-07-01

    At present, XML language has been regarded as a standard of data exchange due to such attributes as being unrelevant to platform, auto-description of itself, easy extention, separation of content and form, and so on. Therefore, in the course of study of share of heterogeneous-data in coal enterprises, we design a heterogeneous-data query system based on XML to provide a unified data integration platform and share of information resource with high quality and speed, which will solve a problem with information islands produced in coal enterprises.

  9. Data rate management and real time operation: recursive adaptive frame integration of limited data

    NASA Astrophysics Data System (ADS)

    Rafailov, Michael K.

    2006-08-01

    Recursive Limited Frame Integration was proposed as a way to improve frame integration performance and mitigate issues related to high data rate needed to support conventional frame integration. The technique uses two thresholds -one tuned for optimum probability of detection, the other to manage required false alarm rate, and places integration process between those thresholds. This configuration allows a non-linear integration process that, along with Signal-to-Noise Ratio (SNR) gain, provides system designers more capability where cost, weight, or power considerations limit system data rate, processing, or memory capability. However, Recursive Frame Integration Limited may have performance issues when single-frame SNR is really low. Recursive Adaptive Limited Frame Integration was proposed as a means to improve limited integration performance with really low single-frame SNR. It combines the benefits of nonlinear recursive limited frame integration and adaptive thresholds with a kind of conventional frame integration. Adding the third threshold may help in managing real time operations. In the paper the Recursive Frame Integration is presented in form of multiple parallel recursive integration. Such an approach can help not only in data rate management but in mitigation of low single frame SNR issue for Recursive Integration as well as in real time operations with frame integration.

  10. Theory of Mind Development in Adolescence and Early Adulthood: The Growing Complexity of Recursive Thinking Ability

    PubMed Central

    Valle, Annalisa; Massaro, Davide; Castelli, Ilaria; Marchetti, Antonella

    2015-01-01

    This study explores the development of theory of mind, operationalized as recursive thinking ability, from adolescence to early adulthood (N = 110; young adolescents = 47; adolescents = 43; young adults = 20). The construct of theory of mind has been operationalized in two different ways: as the ability to recognize the correct mental state of a character, and as the ability to attribute the correct mental state in order to predict the character’s behaviour. The Imposing Memory Task, with five recursive thinking levels, and a third-order false-belief task with three recursive thinking levels (devised for this study) have been used. The relationship among working memory, executive functions, and linguistic skills are also analysed. Results show that subjects exhibit less understanding of elevated recursive thinking levels (third, fourth, and fifth) compared to the first and second levels. Working memory is correlated with total recursive thinking, whereas performance on the linguistic comprehension task is related to third level recursive thinking in both theory of mind tasks. An effect of age on third-order false-belief task performance was also found. A key finding of the present study is that the third-order false-belief task shows significant age differences in the application of recursive thinking that involves the prediction of others’ behaviour. In contrast, such an age effect is not observed in the Imposing Memory Task. These results may support the extension of the investigation of the third order false belief after childhood.

  11. A new XML-aware compression technique for improving performance of healthcare information systems over hospital networks.

    PubMed

    Al-Shammary, Dhiah; Khalil, Ibrahim

    2010-01-01

    Most organizations exchange, collect, store and process data over the Internet. Many hospital networks deploy Web services to send and receive patient information. SOAP (Simple Object Access Protocol) is the most usable communication protocol for Web services. XML is the standard encoding language of SOAP messages. However, the major drawback of XML messages is the high network traffic caused by large overheads. In this paper, two XML-aware compressors are suggested to compress patient messages stemming from any data transactions between Web clients and servers. The proposed compression techniques are based on the XML structure concepts and use both fixed-length and Huffman encoding methods for translating the XML message tree. Experiments show that they outperform all the conventional compression methods and can save tremendous amount of network bandwidth. PMID:21095766

  12. Efficient design of two-dimensional recursive digital filters. Final report

    SciTech Connect

    Twogood, R.E.; Mitra, S.K.

    1980-01-01

    This report outlines the research progress during the period August 1978 to July 1979. This work can be divided into seven basic project areas. Project 1 deals with a comparative study of 2-D recursive and nonrecursive digital filters. The second project addresses a new design technique for 2-D half-plane recursive filters, and Projects 3 thru 5 deal with implementation issues. The sixth project presents our recent study of the applicability of array processors to 2-D digital signal processing. The final project involves our investigation into techniques for incorporating symmetry constraints on 2-D recursive filters in order to yield more efficient implementations.

  13. Recursive Principal Components Analysis Using Eigenvector Matrix Perturbation

    NASA Astrophysics Data System (ADS)

    Erdogmus, Deniz; Rao, Yadunandana N.; Peddaneni, Hemanth; Hegde, Anant; Principe, Jose C.

    2004-12-01

    Principal components analysis is an important and well-studied subject in statistics and signal processing. The literature has an abundance of algorithms for solving this problem, where most of these algorithms could be grouped into one of the following three approaches: adaptation based on Hebbian updates and deflation, optimization of a second-order statistical criterion (like reconstruction error or output variance), and fixed point update rules with deflation. In this paper, we take a completely different approach that avoids deflation and the optimization of a cost function using gradients. The proposed method updates the eigenvector and eigenvalue matrices simultaneously with every new sample such that the estimates approximately track their true values as would be calculated from the current sample estimate of the data covariance matrix. The performance of this algorithm is compared with that of traditional methods like Sanger's rule and APEX, as well as a structurally similar matrix perturbation-based method.

  14. Multi-Resolution Seismic Tomography Based on Recursive Tessellation Hierarchy

    SciTech Connect

    Simmons, N A; Myers, S C; Ramirez, A

    2009-07-01

    A 3-D global tomographic model that reconstructs velocity structure at multiple scales and incorporates laterally variable seismic discontinuities is currently being developed. The model parameterization is node-based where nodes are placed along vertices defined by triangular tessellations of a spheroidal surface. The triangular tessellation framework is hierarchical. Starting with a tetrahexahedron representing the whole globe (1st level of the hierarchy, 24 faces), they divide each triangle of the tessellation into daughter triangles. The collection of all daughter triangles comprises the 2nd level of the tessellation hierarchy and further recursion produces an arbitrary number of tessellation levels and arbitrarily fine node-spacing. They have developed an inversion procedure that takes advantage of the recursive properties of the tessellation hierarchies by progressively solving for shorter wavelength heterogeneities. In this procedure, we first perform the tomographic inversion using a tessellation level with coarse node spacing. They find that a coarse node spacing of approximately 8{sup o} is adequate to capture bulk regional properties. They then conduct the tomographic inversion on a 4{sup o} tessellation level using the residuals and inversion results from the 8{sup o} run. In practice they find that the progressive tomography approach is robust, providing an intrinsic regularization for inversion stability and avoids the issue of predefining resolution levels. Further, determining average regional properties with coarser tessellation levels enables long-wavelength heterogeneities to account for sparsely sampled regions (or regions of the mantle where longer wavelength patterns of heterogeneity suffice) while allowing shorter length-scale heterogeneities to emerge where necessary. They demonstrate the inversion approach with a set of synthetic test cases that mimic the complex nature of data arrangements (mixed-determined inversion) common to most tomographic problems. They also apply the progressive inversion approach with Pn waves traveling within the Middle East region and compare the results to simple tomographic inversions. As expected from synthetic testing, the progressive approach results in detailed structure where there is high data density and broader regional anomalies where seismic information is sparse. The ultimate goal is to use these methods to produce a seamless, multi-resolution global tomographic model with local model resolution determined by the constraints afforded by available data. They envisage this new technique as the general approach to be employed for future multi-resolution model development with complex arrangements of regional and teleseismic information.

  15. Strategic Industrial Alliances in Paper Industry: XML- vs Ontology-Based Integration Platforms

    ERIC Educational Resources Information Center

    Naumenko, Anton; Nikitin, Sergiy; Terziyan, Vagan; Zharko, Andriy

    2005-01-01

    Purpose: To identify cases related to design of ICT platforms for industrial alliances, where the use of Ontology-driven architectures based on Semantic web standards is more advantageous than application of conventional modeling together with XML standards. Design/methodology/approach: A comparative analysis of the two latest and the most obvious

  16. A Semantic Analysis of XML Schema Matching for B2B Systems Integration

    ERIC Educational Resources Information Center

    Kim, Jaewook

    2011-01-01

    One of the most critical steps to integrating heterogeneous e-Business applications using different XML schemas is schema matching, which is known to be costly and error-prone. Many automatic schema matching approaches have been proposed, but the challenge is still daunting because of the complexity of schemas and immaturity of technologies in…

  17. Association rule extraction from XML stream data for wireless sensor networks.

    PubMed

    Paik, Juryon; Nam, Junghyun; Kim, Ung Mo; Won, Dongho

    2014-01-01

    With the advances of wireless sensor networks, they yield massive volumes of disparate, dynamic and geographically-distributed and heterogeneous data. The data mining community has attempted to extract knowledge from the huge amount of data that they generate. However, previous mining work in WSNs has focused on supporting simple relational data structures, like one table per network, while there is a need for more complex data structures. This deficiency motivates XML, which is the current de facto format for the data exchange and modeling of a wide variety of data sources over the web, to be used in WSNs in order to encourage the interchangeability of heterogeneous types of sensors and systems. However, mining XML data for WSNs has two challenging issues: one is the endless data flow; and the other is the complex tree structure. In this paper, we present several new definitions and techniques related to association rule mining over XML data streams in WSNs. To the best of our knowledge, this work provides the first approach to mining XML stream data that generates frequent tree items without any redundancy. PMID:25046017

  18. An Electronic Finding Aid Using Extensible Markup Language (XML) and Encoded Archival Description (EAD).

    ERIC Educational Resources Information Center

    Chang, May

    2000-01-01

    Describes the development of electronic finding aids for archives at the University of Illinois, Urbana-Champaign that used XML (extensible markup language) and EAD (encoded archival description) to enable more flexible information management and retrieval than using MARC or a relational database management system. EAD template is appended.

  19. Software to Create and Support XML Version of the Information System HIPNS

    NASA Astrophysics Data System (ADS)

    Duginov, A. V.; Kulikov, I. E.; Litvinenko, E. I.; Pikelner, T. L.; Semenov, R. N.

    The HIPNS project (the information system on neutron sources and scientific research based on these sources) supported by the Russian Foundation of Basic Research (grant 01-07-90347) is under development at the Frank Laboratory of Neutron Physics of JINR. The main distinction of the system is an orientation toward the potential users of neutron instruments, while others are directed to the neutron source specialists. The core of the information system is a relational database, which contains the main parameters of the setups and includes a lot of links to the electronic documents, providing an access to the remote resources and electronic collections. The report presents the current results of XML format application for HIPNS. The work was aimed to the following tasks: XML code generation from the database content, modern design of the resulting pages, tools for HTML code generation from XML/XSL codes and some others. This practical experience could be helpful for other developers because of the fact, that the use of XML is still quite complicated due to the incompleteness of the corresponding software.

  20. 77 FR 46986 - Revisions to Electric Quarterly Report Filing Process; Availability of Draft XML Schema

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-08-07

    ... Reports as proposed in the Commission's June 21, 2012 Notice of Proposed Rulemaking, 77 FR 39447 (July 3, 2012). Please refer to the Supplementary Information Section below for details. DATES: The draft XML....ferc.gov/EventCalendar/EventDetails.aspx?ID=6411&CalType=&CalendarID=116&Date=&View=Listview , and...

  1. SPIRIX: A Peer-to-Peer Search Engine for XML-Retrieval

    NASA Astrophysics Data System (ADS)

    Winter, Judith; Drobnik, Oswald

    At INEX 2008 we presented SPIRIX, a Peer-to-Peer search engine developed to investigate distributed XML-Retrieval. Such investigations have been neglected by INEX so far: while there is a variety of successful and effective XML-Retrieval approaches, all current solutions are centralized search engines. They do not consider distributed scenarios, where it is undesired or impossible to hold the whole collection on one single machine. Such scenarios include search in large-scale collections, where the load of computations and storage consumption is too high for one server. Other systems consist of different owners of heterogeneous collections willing to share their documents without giving up full control over their documents by uploading them on a central server. Currently, there are research solutions for distributed text-retrieval or multimedia-retrieval. With INEX and innovative techniques for exploiting XML-structure, it is now time to extend research to distributed XML-Retrieval. This paper reports on SPIRIX performance at INEX08.

  2. Enhancing the Effectiveness of Description with Innovative Tools--The Example of MidosaXML, Germany

    ERIC Educational Resources Information Center

    Menne-Haritz, Angelika

    2005-01-01

    Online description using new Internet technologies, especially XML, is an interesting way to enhance the effectiveness and productivity of archival processing. Users and archivists benefit from the developments, making the achievement of their aims easier. The provenance-based structured presentation of finding aids can be supported with many new…

  3. Applying XML-Based Technologies to Developing Online Courses: The Case of a Prototype Learning Environment

    ERIC Educational Resources Information Center

    Jedrzejowicz, Joanna; Neumann, Jakub

    2007-01-01

    Purpose: This paper seeks to describe XML technologies and to show how they can be applied for developing web-based courses and supporting authors who do not have much experience with the preparation of web-based courses. Design/methodology/approach: When developing online courses the academic staff has to address the following problem--how to…

  4. Strategic Industrial Alliances in Paper Industry: XML- vs Ontology-Based Integration Platforms

    ERIC Educational Resources Information Center

    Naumenko, Anton; Nikitin, Sergiy; Terziyan, Vagan; Zharko, Andriy

    2005-01-01

    Purpose: To identify cases related to design of ICT platforms for industrial alliances, where the use of Ontology-driven architectures based on Semantic web standards is more advantageous than application of conventional modeling together with XML standards. Design/methodology/approach: A comparative analysis of the two latest and the most obvious…

  5. A Semantic Analysis of XML Schema Matching for B2B Systems Integration

    ERIC Educational Resources Information Center

    Kim, Jaewook

    2011-01-01

    One of the most critical steps to integrating heterogeneous e-Business applications using different XML schemas is schema matching, which is known to be costly and error-prone. Many automatic schema matching approaches have been proposed, but the challenge is still daunting because of the complexity of schemas and immaturity of technologies in

  6. An Electronic Finding Aid Using Extensible Markup Language (XML) and Encoded Archival Description (EAD).

    ERIC Educational Resources Information Center

    Chang, May

    2000-01-01

    Describes the development of electronic finding aids for archives at the University of Illinois, Urbana-Champaign that used XML (extensible markup language) and EAD (encoded archival description) to enable more flexible information management and retrieval than using MARC or a relational database management system. EAD template is appended.…

  7. Association Rule Extraction from XML Stream Data for Wireless Sensor Networks

    PubMed Central

    Paik, Juryon; Nam, Junghyun; Kim, Ung Mo; Won, Dongho

    2014-01-01

    With the advances of wireless sensor networks, they yield massive volumes of disparate, dynamic and geographically-distributed and heterogeneous data. The data mining community has attempted to extract knowledge from the huge amount of data that they generate. However, previous mining work in WSNs has focused on supporting simple relational data structures, like one table per network, while there is a need for more complex data structures. This deficiency motivates XML, which is the current de facto format for the data exchange and modeling of a wide variety of data sources over the web, to be used in WSNs in order to encourage the interchangeability of heterogeneous types of sensors and systems. However, mining XML data for WSNs has two challenging issues: one is the endless data flow; and the other is the complex tree structure. In this paper, we present several new definitions and techniques related to association rule mining over XML data streams in WSNs. To the best of our knowledge, this work provides the first approach to mining XML stream data that generates frequent tree items without any redundancy. PMID:25046017

  8. Update '98.

    ERIC Educational Resources Information Center

    Mock, Karen R.

    1998-01-01

    Updates cases and issues previously discussed in this regular column on human rights in Canada, including racism and anti-Semitism, laws on hate crimes, hate sites on the World Wide Web, the use of the "free speech" defense by hate groups, and legal challenges to antiracist groups by individuals criticized by them. (DSK)

  9. A Parallel Implementation of Multilevel Recursive Spectral Bisection for Application to Adaptive Unstructured Meshes. Chapter 1

    NASA Technical Reports Server (NTRS)

    Barnard, Stephen T.; Simon, Horst; Lasinski, T. A. (Technical Monitor)

    1994-01-01

    The design of a parallel implementation of multilevel recursive spectral bisection is described. The goal is to implement a code that is fast enough to enable dynamic repartitioning of adaptive meshes.

  10. A parallel implementation of multilevel recursive spectral bisection for application to adaptive unstructured meshes

    SciTech Connect

    Barnard, S.T.; Simon, H.

    1995-12-01

    The design of a parallel implementation of multilevel recursive spectral bisection is described. The goal is to implement a code that is fast enough to enable dynamic repartitioning of adaptive meshes.

  11. Modifications to recursion unfolding algorithms to find more appropriate neutron spectra.

    PubMed

    Lowry, K A; Johnson, T L

    1984-10-01

    Modifications were made to recursion unfolding algorithms which allow the incorporation of prior knowledge, such as cutoff energy or preferred spectral shape, into the solution. Using these modifications, more appropriate neutron spectra are obtained. PMID:6548987

  12. Fast and Efficient XML Data Access for Next-Generation Mass Spectrometry

    PubMed Central

    Röst, Hannes L.; Schmitt, Uwe; Aebersold, Ruedi; Malmström, Lars

    2015-01-01

    Motivation In mass spectrometry-based proteomics, XML formats such as mzML and mzXML provide an open and standardized way to store and exchange the raw data (spectra and chromatograms) of mass spectrometric experiments. These file formats are being used by a multitude of open-source and cross-platform tools which allow the proteomics community to access algorithms in a vendor-independent fashion and perform transparent and reproducible data analysis. Recent improvements in mass spectrometry instrumentation have increased the data size produced in a single LC-MS/MS measurement and put substantial strain on open-source tools, particularly those that are not equipped to deal with XML data files that reach dozens of gigabytes in size. Results Here we present a fast and versatile parsing library for mass spectrometric XML formats available in C++ and Python, based on the mature OpenMS software framework. Our library implements an API for obtaining spectra and chromatograms under memory constraints using random access or sequential access functions, allowing users to process datasets that are much larger than system memory. For fast access to the raw data structures, small XML files can also be completely loaded into memory. In addition, we have improved the parsing speed of the core mzML module by over 4-fold (compared to OpenMS 1.11), making our library suitable for a wide variety of algorithms that need fast access to dozens of gigabytes of raw mass spectrometric data. Availability Our C++ and Python implementations are available for the Linux, Mac, and Windows operating systems. All proposed modifications to the OpenMS code have been merged into the OpenMS mainline codebase and are available to the community at https://github.com/OpenMS/OpenMS. PMID:25927999

  13. XML-Based Generator of C++ Code for Integration With GUIs

    NASA Technical Reports Server (NTRS)

    Hua, Hook; Oyafuso, Fabiano; Klimeck, Gerhard

    2003-01-01

    An open source computer program has been developed to satisfy a need for simplified organization of structured input data for scientific simulation programs. Typically, such input data are parsed in from a flat American Standard Code for Information Interchange (ASCII) text file into computational data structures. Also typically, when a graphical user interface (GUI) is used, there is a need to completely duplicate the input information while providing it to a user in a more structured form. Heretofore, the duplication of the input information has entailed duplication of software efforts and increases in susceptibility to software errors because of the concomitant need to maintain two independent input-handling mechanisms. The present program implements a method in which the input data for a simulation program are completely specified in an Extensible Markup Language (XML)-based text file. The key benefit for XML is storing input data in a structured manner. More importantly, XML allows not just storing of data but also describing what each of the data items are. That XML file contains information useful for rendering the data by other applications. It also then generates data structures in the C++ language that are to be used in the simulation program. In this method, all input data are specified in one place only, and it is easy to integrate the data structures into both the simulation program and the GUI. XML-to-C is useful in two ways: 1. As an executable, it generates the corresponding C++ classes and 2. As a library, it automatically fills the objects with the input data values.

  14. Recursive dynamic programming for adaptive sequence and structure alignment.

    PubMed

    Thiele, R; Zimmer, R; Lengauer, T

    1995-01-01

    We propose a new alignment procedure that is capable of aligning protein sequences and structures in a unified manner. Recursive dynamic programming (RDP) is a hierarchical method which, on each level of the hierarchy, identifies locally optimal solutions and assembles them into partial alignments of sequences and/or structures. In contrast to classical dynamic programming, RDP can also handle alignment problems that use objective functions not obeying the principle of prefix optimality, e.g. scoring schemes derived from energy potentials of mean force. For such alignment problems, RDP aims at computing solutions that are near-optimal with respect to the involved cost function and biologically meaningful at the same time. Towards this goal, RDP maintains a dynamic balance between different factors governing alignment fitness such as evolutionary relationships and structural preferences. As in the RDP method gaps are not scored explicitly, the problematic assignment of gap cost parameters is circumvented. In order to evaluate the RDP approach we analyse whether known and accepted multiple alignments based on structural information can be reproduced with the RDP method. For this purpose, we consider the family of ferredoxins as our prime example. Our experiments show that, if properly tuned, the RDP method can outperform methods based on classical sequence alignment algorithms as well as methods that take purely structural information into account. PMID:7584462

  15. Implicit Learning of Recursive Context-Free Grammars

    PubMed Central

    Rohrmeier, Martin; Fu, Qiufang; Dienes, Zoltan

    2012-01-01

    Context-free grammars are fundamental for the description of linguistic syntax. However, most artificial grammar learning experiments have explored learning of simpler finite-state grammars, while studies exploring context-free grammars have not assessed awareness and implicitness. This paper explores the implicit learning of context-free grammars employing features of hierarchical organization, recursive embedding and long-distance dependencies. The grammars also featured the distinction between left- and right-branching structures, as well as between centre- and tail-embedding, both distinctions found in natural languages. People acquired unconscious knowledge of relations between grammatical classes even for dependencies over long distances, in ways that went beyond learning simpler relations (e.g. n-grams) between individual words. The structural distinctions drawn from linguistics also proved important as performance was greater for tail-embedding than centre-embedding structures. The results suggest the plausibility of implicit learning of complex context-free structures, which model some features of natural languages. They support the relevance of artificial grammar learning for probing mechanisms of language learning and challenge existing theories and computational models of implicit learning. PMID:23094021

  16. Recursive recovery of Markov transition probabilities from boundary value data

    SciTech Connect

    Patch, S.K.

    1994-04-01

    In an effort to mathematically describe the anisotropic diffusion of infrared radiation in biological tissue Gruenbaum posed an anisotropic diffusion boundary value problem in 1989. In order to accommodate anisotropy, he discretized the temporal as well as the spatial domain. The probabilistic interpretation of the diffusion equation is retained; radiation is assumed to travel according to a random walk (of sorts). In this random walk the probabilities with which photons change direction depend upon their previous as well as present location. The forward problem gives boundary value data as a function of the Markov transition probabilities. The inverse problem requires finding the transition probabilities from boundary value data. Problems in the plane are studied carefully in this thesis. Consistency conditions amongst the data are derived. These conditions have two effects: they prohibit inversion of the forward map but permit smoothing of noisy data. Next, a recursive algorithm which yields a family of solutions to the inverse problem is detailed. This algorithm takes advantage of all independent data and generates a system of highly nonlinear algebraic equations. Pluecker-Grassmann relations are instrumental in simplifying the equations. The algorithm is used to solve the 4 {times} 4 problem. Finally, the smallest nontrivial problem in three dimensions, the 2 {times} 2 {times} 2 problem, is solved.

  17. Random forest methodology for model-based recursive partitioning: the mobForest package for R

    PubMed Central

    2013-01-01

    Background Recursive partitioning is a non-parametric modeling technique, widely used in regression and classification problems. Model-based recursive partitioning is used to identify groups of observations with similar values of parameters of the model of interest. The mob() function in the party package in R implements model-based recursive partitioning method. This method produces predictions based on single tree models. Predictions obtained through single tree models are very sensitive to small changes to the learning sample. We extend the model-based recursive partition method to produce predictions based on multiple tree models constructed on random samples achieved either through bootstrapping (random sampling with replacement) or subsampling (random sampling without replacement) on learning data. Results Here we present an R package called “mobForest” that implements bagging and random forests methodology for model-based recursive partitioning. The mobForest package constructs large number of model-based trees and the predictions are aggregated across these trees resulting in more stable predictions. The package also includes functions for computing predictive accuracy estimates and plots, residuals plot, and variable importance plot. Conclusion The mobForest package implements a random forest type approach for model-based recursive partitioning. The R package along with it source code is available at http://CRAN.R-project.org/package=mobForest. PMID:23577585

  18. Recursive Subsystems in Aphasia and Alzheimer's Disease: Case Studies in Syntax and Theory of Mind

    PubMed Central

    Bánréti, Zoltán; Hoffmann, Ildikó; Vincze, Veronika

    2016-01-01

    The relationship between recursive sentence embedding and theory-of-mind (ToM) inference is investigated in three persons with Broca's aphasia, two persons with Wernicke's aphasia, and six persons with mild and moderate Alzheimer's disease (AD). We asked questions of four types about photographs of various real-life situations. Type 4 questions asked participants about intentions, thoughts, or utterances of the characters in the pictures (“What may X be thinking/asking Y to do?”). The expected answers typically involved subordinate clauses introduced by conjunctions or direct quotations of the characters' utterances. Broca's aphasics did not produce answers with recursive sentence embedding. Rather, they projected themselves into the characters' mental states and gave direct answers in the first person singular, with relevant ToM content. We call such replies “situative statements.” Where the question concerned the mental state of the character but did not require an answer with sentence embedding (“What does X hate?”), aphasics gave descriptive answers rather than situative statements. Most replies given by persons with AD to Type 4 questions were grammatical instances of recursive sentence embedding. They also gave a few situative statements but the ToM content of these was irrelevant. In more than one third of their well-formed sentence embeddings, too, they conveyed irrelevant ToM contents. Persons with moderate AD were unable to pass secondary false belief tests. The results reveal double dissociation: Broca's aphasics are unable to access recursive sentence embedding but they can make appropriate ToM inferences; moderate AD persons make the wrong ToM inferences but they are able to access recursive sentence embedding. The double dissociation may be relevant for the nature of the relationship between the two recursive capacities. Broca's aphasics compensated for the lack of recursive sentence embedding by recursive ToM reasoning represented in very simple syntactic forms: they used one recursive subsystem to stand in for another recursive subsystem. PMID:27064887

  19. Recursive stochastic deconvolution in the estimation of earthquake source parameters: Synthetic waveforms

    NASA Astrophysics Data System (ADS)

    Qu, J.; Teng, T. L.

    1994-11-01

    In this paper, a method of the linear minimum mean-squares error (LMMSE) solution for source inversion is presented in terms of a recursive algorithm. A covariance matrix of estimation error, as well as a resolution matrix are also computed through recursion. It is shown that this recursive solution corresponds to a stationary Kalman filtering estimation for a linear dynamic system, which makes it possible to perform satisfactorily in an environment where complete knowledge of the relevant signal characteristics is not available. In a stationary environment, our recursive solution converges to the optimum Wiener solution. In a rather straightforward manner, the multichannel deconvolution problem is translated into a set of recursive expressions. The procedures have been tested using a number of synthetic data sets, including a point and a complex source, with satisfactory results. It is found that the solution is improved recursively with each addition of new data. We have found further that it is the error-covariance matrix, not the resolution matrix, that gives a measurement of the recursive performance. Since the recursive scheme of LMMSE runs in a manner based on either block-by-block or sample-by-sample operation, the memory requirement can be quite small. For problems involving sparse matrices, the recursive algorithm leads to fast and efficient computation. This method is tested by examining the Sierra Madre earthquake M(sub s) of 28 June 1991, California. This event is well-recorded by the broad-band TERRAscope array. The moment tensor inversion through the presented method indicates that the solution is improved recursively when new data become available. It was found that the later arrivals on the observed seismograms have very little influence on the solution while the inclusion of new data from different stations yields substantial improvement on the mechanism to a certain point where further addition of data will not make much difference to the resulting best double-couple decomposition. However, the content of double-couple components shows a striking increase from approximately 50 to 90% with the inclusion of more data from other stations. This result demonstrates clearly the robustness of our approach since the inadequacy of the source representation and the earth model in the moment-tensor inversion may be remedied by the inclusion of more data.

  20. Creating an XML representation for science data formats with XDF: cases, design and implementation.

    NASA Astrophysics Data System (ADS)

    Thomas, B.; Shaya, E.; Huang, Z.; Cheung, C. Y.

    2001-12-01

    XDF (the eXtensible Data Format) is an XML-based general science data format designed and maintained by the ADC (Astronomical Data Center). XDF may be used to describe both binary and ASCII scientific data and may alternatively serve as an XML wrapper for legacy science data files. The XML heritage of XDF lends it many desirable features including the ability to describe both science data files locally or remotely, to be manipulated by and viewable with a wide variety of XML aware software (including browsers), to have universal validation (via its DTD or schema) and semantic mark-up of data products at all points in the science data cycle. In this poster we will focus on a major aspect of XDF: its ability to describe existing legacy data formats. XDF may be achieve this by either directly describing the legacy data files and serving as a wrapper or, if greater sophistication and a higher level API is needed, by being extended in an object-oriented fashion to create a new hybrid science data format with the features of the legacy science data format. Either of these approaches can be used with minimal effort to create an XML-based representation for existing science data formats. We anticipate that this ability will allow for greater cost effectiveness in science data processing environments and science data archives because XDF can serve as a standard, extensible data format that can be used and re-used for any type of data in any mission and may serve as a transitional bridge between existing dataformats. We will present how XDF may generally be used to describe existing legacy data and cover a general outline of the steps needed to extend XDF to easily create your own XML/XDF-based data format. We will present several test cases including FITS and CDF and use them illustrate the problems and abilities of both approaches and to give examples of how to solve specific common problems.

  1. Development of the Plate Tectonics and Seismology markup languages with XML

    NASA Astrophysics Data System (ADS)

    Babaie, H.; Babaei, A.

    2003-04-01

    The Extensible Markup Language (XML) and its specifications such as the XSD Schema, allow geologists to design discipline-specific vocabularies such as Seismology Markup Language (SeismML) or Plate Tectonics Markup Language (TectML). These languages make it possible to store and interchange structured geological information over the Web. Development of a geological markup language requires mapping geological concepts, such as "Earthquake" or "Plate" into a UML object model, applying a modeling and design environment. We have selected four inter-related geological concepts: earthquake, fault, plate, and orogeny, and developed four XML Schema Definitions (XSD), that define the relationships, cardinalities, hierarchies, and semantics of these concepts. In such a geological concept model, the UML object "Earthquake" is related to one or more "Wave" objects, each arriving to a seismic station at a specific "DateTime", and relating to a specific "Epicenter" object that lies at a unique "Location". The "Earthquake" object occurs along a "Segment" of a "Fault" object, which is related to a specific "Plate" object. The "Fault" has its own associations with such things as "Bend", "Step", and "Segment", and could be of any kind (e.g., "Thrust", "Transform'). The "Plate" is related to many other objects such as "MOR", "Subduction", and "Forearc", and is associated with an "Orogeny" object that relates to "Deformation" and "Strain" and several other objects. These UML objects were mapped into XML Metadata Interchange (XMI) formats, which were then converted into four XSD Schemas. The schemas were used to create and validate the XML instance documents, and to create a relational database hosting the plate tectonics and seismological data in the Microsoft Access format. The SeismML and TectML allow seismologists and structural geologists, among others, to submit and retrieve structured geological data on the Internet. A seismologist, for example, can submit peer-reviewed and reliable data about a specific earthquake to a Java Server Page on our web site hosting the XML application. Other geologists can readily retrieve the submitted data, saved in files or special tables of the designed database, through a search engine designed with J2EE (JSP, servlet, Java Bean) and XML specifications such as XPath, XPointer, and XSLT. When extended to include all the important concepts of seismology and plate tectonics, the two markup languages will make global interchange of geological data a reality.

  2. Enhancement of the Work in Scia Engineer's Environment by Employment of XML Programming Language

    NASA Astrophysics Data System (ADS)

    Kortiš, Ján

    2015-12-01

    The productivity of the work of engineers in the design of building structures by applying the rules of technical standards [1] has been increasing by using different software products for recent years. The software products offer engineers new possibilities to design different structures. However, there are problems especially for design of structures with similar static schemes as it is needed to follow the same work-steps. This can be more effective if the steps are done automatically by using a programming language for leading the processes that are done by software. The design process of timber structure which is done in the environment of Scia Engineer software is presented in the article. XML Programming Language is used for automatization of the design and the XML code is modified in the Excel environment by using VBA Programming language [2], [3].

  3. Using XML Configuration-Driven Development to Create a Customizable Ground Data System

    NASA Technical Reports Server (NTRS)

    Nash, Brent; DeMore, Martha

    2009-01-01

    The Mission data Processing and Control Subsystem (MPCS) is being developed as a multi-mission Ground Data System with the Mars Science Laboratory (MSL) as the first fully supported mission. MPCS is a fully featured, Java-based Ground Data System (GDS) for telecommand and telemetry processing based on Configuration-Driven Development (CDD). The eXtensible Markup Language (XML) is the ideal language for CDD because it is easily readable and editable by all levels of users and is also backed by a World Wide Web Consortium (W3C) standard and numerous powerful processing tools that make it uniquely flexible. The CDD approach adopted by MPCS minimizes changes to compiled code by using XML to create a series of configuration files that provide both coarse and fine grained control over all aspects of GDS operation.

  4. The Knowledge Sharing Based on PLIB Ontology and XML for Collaborative Product Commerce

    NASA Astrophysics Data System (ADS)

    Ma, Jun; Luo, Guofu; Li, Hao; Xiao, Yanqiu

    Collaborative Product Commerce (CPC) has become a brand-new commerce mode for manufacture. In order to promote information communication with each other more efficiently in CPC, a knowledge-sharing framework based on PLIB (ISO 13584) ontology and XML was presented, and its implementation method was studied. At first, according to the methodology of PLIB (ISO 13584), a common ontology—PLIB ontology was put forward which provide a coherent conceptual meaning within the context of CPC domain. Meanwhile, for the sake of knowledge intercommunion via internet, the PLIB ontology formalization description by EXPRESS mode was converted into XML Schema, and two mapping methods were presented: direct mapping approach and meta-levels mapping approach, while the latter was adopted. Based on above work, a parts resource knowledge-sharing framework (CPC-KSF) was put forward and realized, which has been applied in the process of automotive component manufacturing collaborative product commerce.

  5. Light at Night Markup Language (LANML): XML Technology for Light at Night Monitoring Data

    NASA Astrophysics Data System (ADS)

    Craine, B. L.; Craine, E. R.; Craine, E. M.; Crawford, D. L.

    2013-05-01

    Light at Night Markup Language (LANML) is a standard, based upon XML, useful in acquiring, validating, transporting, archiving and analyzing multi-dimensional light at night (LAN) datasets of any size. The LANML standard can accommodate a variety of measurement scenarios including single spot measures, static time-series, web based monitoring networks, mobile measurements, and airborne measurements. LANML is human-readable, machine-readable, and does not require a dedicated parser. In addition LANML is flexible; ensuring future extensions of the format will remain backward compatible with analysis software. The XML technology is at the heart of communicating over the internet and can be equally useful at the desktop level, making this standard particularly attractive for web based applications, educational outreach and efficient collaboration between research groups.

  6. Transforming XML-based electronic patient records for use in medical case based reasoning systems.

    PubMed

    Abidi, S S; Manickam, S

    2000-01-01

    Electronic patient records (EPR) can be regarded as an implicit source of clinical behaviour and problem-solving knowledge, systematically compiled by clinicians. We present an approach, together with its computational implementation, to pro-actively transform XML-based EPR into specialised Clinical Cases (CC) in the realm of Medical Case Base Systems. The 'correct' transformation of EPR to CC involves structural, terminological and conceptual standardisation, which is achieved by a confluence of techniques and resources, such as XML, UMLS (meta-thesaurus) and medical knowledge ontologies. We present below the functional architecture of a Medical Case-Base Reasoning Info-Structure (MCRIS) that features two distinct, yet related, functionalities: (1) a generic medical case-based reasoning system for decision-support activities; and (2) an EPR-CC transformation system to transform typical EPR's to CC. PMID:11187645

  7. Transparent XML Binding using the ALMA Common Software (ACS) Container/Component Framework

    NASA Astrophysics Data System (ADS)

    Sommer, H.; Chiozzi, G.; Fugate, D.; Sekoranja, M.

    2004-07-01

    ALMA software, from high-level data flow applications down to instrument control, is built using the ACS framework. The common architecture and infrastructure used for the whole ALMA software is presented at this conference in (Schwarz, Farris, & Sommer 2004). ACS offers a CORBA-based container/component model and supports the exchange and persistence of XML data. For the Java programming language, the container integrates transparently the use of type-safe Java binding classes to let applications conveniently work with XML transfer objects without having to parse or serialize them. This paper will show how the ACS container/component architecture serves to pass complex data structures, such as observation meta-data, between heterogeneous applications.

  8. [Pharmacovigilance update].

    PubMed

    Fogarasi Szabo, Nathalie; Diezi, Léonore; Delenclos, Laurie; Renard, Delphine; Chtioui, Haithem; Rothuizen, Laura E; Buclin, Thierry; Livio, Françoise

    2015-01-14

    The main pharmacovigilance updates in 2014 are reviewed. Ivabradine: increased risk of cardiovascular death and myocardial infarction in patients with symptomatic angina treated with high dosages. Clopidogrel: rare observations of acquired hemophilia. Orlistat: may reduce the absorption of HIV antiretrovirals. Ponatinib: increased risk of arteriopathy and thrombosis. Axitinib: significant risk of heart failure (class effect). Tocilizumab: possible causal relationship with the emergence or aggravation of psoriasis. Lithium: hypercalcemia and hyperparathyroidism commonly observed. Sildenalfil: suspected causal association with melanoma, so far not proven, Methylphenidate: rare observations of priapism. St John's wort (Hypericum): reduced effectiveness of hormonal contraceptives, including implants. PMID:25799668

  9. NAMAS update

    NASA Astrophysics Data System (ADS)

    Goodier, I. W.

    1991-09-01

    An outline of the 'NAMAS update' talk is given. Details of calibrated and testing laboratories that have been accredited by NAMAS in the radiological field are given. The customers of 'NAMAS ' accredited laboratories can be assured of the following: the competence of the laboratory has been assessed by independent, expert assessors; all the measurements performed are traceable to the national standards of measurements at NPL; the laboratories have performed their measurements in a technically sound way; the laboratories operate an approved quality system and therefore quality measurements are made at all times and not just on special occasions; the uncertainties of measurements quoted by the laboratories have been assessed and can be substantiated.

  10. XTCE and XML Database Evolution and Lessons from JWST, LandSat, and Constellation

    NASA Technical Reports Server (NTRS)

    Gal-Edd, Jonathan; Kreistle, Steven; Fatig. Cirtos; Jones, Ronald

    2008-01-01

    The database organizations within three different NASA projects have advanced current practices by creating database synergy between the various spacecraft life cycle stakeholders and educating users in the benefits of the Consultative Committee for Space Data Systems (CCSDS) XML Telemetry and Command Exchange (XTCE) format. The combination of XML for managing program data and CCSDS XTCE for exchange is a robust approach that will meet all user requirements using Standards and Non proprietary tools. COTS tools for XTCEKML are very wide and varied. To combine together various low cost and free tools can be more expensive in the long run than choosing a more expensive COTS tool that meets all the needs. This was especially important when deploying in 32 remote sites with no need for licenses. A common mission XTCEKML format between dissimilar systems is possible and is not difficult. Command XMLKTCE is more complex than telemetry and the use of XTCEKML metadata to describe pages and scripts is needed due to the proprietary nature of most current ground systems. Other mission and science products such as spacecraft loads, science image catalogs, and mission operation procedures can all be described with XML as well to increase there flexibility as systems evolve and change. Figure 10 is an example of a spacecraft table load. The word is out and the XTCE community is growing, The f sXt TCE user group was held in October and in addition to ESAESOC, SC02000, and CNES identified several systems based on XTCE. The second XTCE user group is scheduled for March 10, 2008 with LDMC and others joining. As the experience with XTCE grows and the user community receives the promised benefits of using XTCE and XML the interest is growing fast.

  11. Comparing Emerging XML Based Formats from a Multi-discipline Perspective

    NASA Astrophysics Data System (ADS)

    Sawyer, D. M.; Reich, L. I.; Nikhinson, S.

    2002-12-01

    This paper analyzes the similarity and differences among several examples of an emerging generation of Scientific Data Formats that are based on XML technologies. Some of the factors evaluated include the goals of these efforts, the data models, and XML technologies used, and the maturity of currently available software. This paper then investigates the practicality of developing a single set of structural data objects and basic scientific concepts, such as units, that could be used across discipline boundaries and extended by disciplines and missions to create Scientific Data Formats for their communities. This analysis is partly based on an effort sponsored by the ESDIS office at GSFC to compare the Earth Science Markup Language (ESML) and the eXtensible Data Format( XDF), two members of this new generation of XML based Data Description Languages that have been developed by NASA funded efforts in recent years. This paper adds FITSML and potentially CDFML to the list of XML based Scientific Data Formats discussed. This paper draws heavily a Formats Evolution Process Committee (http://ssdoo.gsfc.nasa.gov/nost/fep/) draft white paper primarily developed by Lou Reich, Mike Folk and Don Sawyer to assist the Space Science community in understanding Scientific Data Formats. One of primary conclusions of that paper is that a scientific data format object model should be examined along two basic axes. The first is the complexity of the computer/mathematical data types supported and the second is the level of scientific domain specialization incorporated. This paper also discusses several of the issues that affect the decision on whether to implement a discipline or project specific Scientific Data Format as a formal extension of a general purpose Scientific Data Format or to implement the APIs independently.

  12. DICOM-compatible format for analytical cytology data that can be expressed in XML

    NASA Astrophysics Data System (ADS)

    Leif, Robert C.; Leif, Suzanne B.

    2001-05-01

    Flow cytometry data can be directly mapped to the Digital Imaging and Communications in Medicine, DICOM standard. A preliminary mapping of list-mode data to the DICOM Waveform information Object will be presented. This mapping encompasses both flow and image list-mode data. Since list- mode data is also produced by digital slide microscopy, which has already been standardized under DICOM, both branches of Analytical Cytology can be united under the DICOM standard. This will result in the functionality of the present International Society for Analytical Cytology Flow Cytometry Standard, FCS, being significantly extended and the elimination of the previously reported FCS design deficiencies. Thus, the present Flow Cytometry Standard can and should be replaced by a Digital Imaging and Communications in Medicine, DICOM, standard. Expression of Analytical Cytology data in any other format, such as XML, can be made interoperable with DICOM by employing the DICOM data types. A fragment of an XML Schema has been created, which demonstrates the feasibility of expressing DICOM data types in XML syntax. The extension of DICOM to include Flow Cytometry will have the benefits of 1) retiring the present FCS, 2) providing a standard that is ubiquitous, internationally accepted, and backed by the medical profession, and 3) inter-operating with the existing medical informatics infrastructure.

  13. XML-based 3D model visualization and simulation framework for dynamic models

    NASA Astrophysics Data System (ADS)

    Kim, Taewoo; Fishwick, Paul A.

    2002-07-01

    Relatively recent advances in computer technology enable us to create three-dimensional (3D) dynamic models and simulate them within a 3D web environment. The use of such models is especially valuable when teaching simulation, and the concepts behind dynamic models, since the models are made more accessible to the students. Students tend to enjoy a construction process in which they are able to employ their own cultural and aesthetic forms. The challenge is to create a language that allows for a grammar for modeling, while simultaneously permitting arbitrary presentation styles. For further flexibility, we need an effective way to represent and simulate dynamic models that can be shared by modelers over the Internet. We present an Extensible Markup Language (XML)-based framework that will guide a modeler in creating personalized 3D models, visualizing its dynamic behaviors, and simulating the created models. A model author will use XML files to represent geometries and topology of a dynamic model. Model Fusion Engine, written in Extensible Stylesheet Language Transformation (XSLT), expedites the modeling process by automating the creation of dynamic models with the user-defined XML files. Modelers can also link simulation programs with a created model to analyze the characteristics of the model. The advantages of this system lie in the education of modeling and simulating dynamic models, and in the exploitation of visualizing the dynamic model behaviors.

  14. Scalable 3D GIS environment managed by 3D-XML-based modeling

    NASA Astrophysics Data System (ADS)

    Shi, Beiqi; Rui, Jianxun; Chen, Neng

    2008-10-01

    Nowadays, the namely 3D GIS technologies become a key factor in establishing and maintaining large-scale 3D geoinformation services. However, with the rapidly increasing size and complexity of the 3D models being acquired, a pressing needed for suitable data management solutions has become apparent. This paper outlines that storage and exchange of geospatial data between databases and different front ends like 3D models, GIS or internet browsers require a standardized format which is capable to represent instances of 3D GIS models, to minimize loss of information during data transfer and to reduce interface development efforts. After a review of previous methods for spatial 3D data management, a universal lightweight XML-based format for quick and easy sharing of 3D GIS data is presented. 3D data management based on XML is a solution meeting the requirements as stated, which can provide an efficient means for opening a new standard way to create an arbitrary data structure and share it over the Internet. To manage reality-based 3D models, this paper uses 3DXML produced by Dassault Systemes. 3DXML uses opening XML schemas to communicate product geometry, structure and graphical display properties. It can be read, written and enriched by standard tools; and allows users to add extensions based on their own specific requirements. The paper concludes with the presentation of projects from application areas which will benefit from the functionality presented above.

  15. A novel XML-based document format with printing quality for web publishing

    NASA Astrophysics Data System (ADS)

    Qiu, Ruiheng; Tang, Zhi; Gao, Liangcai; Yu, Yinyan

    2010-02-01

    Although many XML-based document formats are available for printing or publishing on the Internet, none of them is well designed to support both high quality printing and web publishing. Therefore, we propose a novel XML-based document format for web publishing, called CEBX, in this paper. The proposed format is a fixed-layout document supporting high quality printing, which has optimized document content organization, physical structure and protection scheme to support web publishing. There are four noteworthy features of CEBX documents: (1) CEBX provides original fixed layout by graphic units for printing quality. (2) The content in CEBX document can be reflowed to fit the display device basing on the content blocks and additional fluid information. (3) XML Document Archiving model (XDA), the packaging model used in CEBX, supports document linearization and incremental edit well. (4) By introducing a segment-based content protection scheme into CEBX, some part of a document can be previewed directly while the remaining part is protected effectively such that readers only need to purchase partial content of a book that they are interested in. This will be very helpful to document distribution and support flexible business models such as try-beforebuy, on-demand reading, superdistribution, etc.

  16. The language faculty that wasn't: a usage-based account of natural language recursion

    PubMed Central

    Christiansen, Morten H.; Chater, Nick

    2015-01-01

    In the generative tradition, the language faculty has been shrinking—perhaps to include only the mechanism of recursion. This paper argues that even this view of the language faculty is too expansive. We first argue that a language faculty is difficult to reconcile with evolutionary considerations. We then focus on recursion as a detailed case study, arguing that our ability to process recursive structure does not rely on recursion as a property of the grammar, but instead emerges gradually by piggybacking on domain-general sequence learning abilities. Evidence from genetics, comparative work on non-human primates, and cognitive neuroscience suggests that humans have evolved complex sequence learning skills, which were subsequently pressed into service to accommodate language. Constraints on sequence learning therefore have played an important role in shaping the cultural evolution of linguistic structure, including our limited abilities for processing recursive structure. Finally, we re-evaluate some of the key considerations that have often been taken to require the postulation of a language faculty. PMID:26379567

  17. The language faculty that wasn't: a usage-based account of natural language recursion.

    PubMed

    Christiansen, Morten H; Chater, Nick

    2015-01-01

    In the generative tradition, the language faculty has been shrinking-perhaps to include only the mechanism of recursion. This paper argues that even this view of the language faculty is too expansive. We first argue that a language faculty is difficult to reconcile with evolutionary considerations. We then focus on recursion as a detailed case study, arguing that our ability to process recursive structure does not rely on recursion as a property of the grammar, but instead emerges gradually by piggybacking on domain-general sequence learning abilities. Evidence from genetics, comparative work on non-human primates, and cognitive neuroscience suggests that humans have evolved complex sequence learning skills, which were subsequently pressed into service to accommodate language. Constraints on sequence learning therefore have played an important role in shaping the cultural evolution of linguistic structure, including our limited abilities for processing recursive structure. Finally, we re-evaluate some of the key considerations that have often been taken to require the postulation of a language faculty. PMID:26379567

  18. Controlling friction-induced instability by recursive time-delayed acceleration feedback

    NASA Astrophysics Data System (ADS)

    Chatterjee, S.; Mahata, P.

    2009-11-01

    A novel time-delayed acceleration feedback method of controlling friction-induced instability is proposed. A single degree-of-freedom mechanical oscillator on a moving belt represents the basic friction-driven system. The control force is synthesized based on an infinite weighted sum of the acceleration of the vibrating mass measured at regular intervals in the past. Such a control force can be effectively produced by the recursive summation of the time-delayed acceleration and the time-delayed control signal, and hence the technique is termed as the recursive time-delayed acceleration feedback control. The local stability analysis of the equilibrium reveals nontrivial and beneficial influences of the recursive gain on the system performance. Robustness of the control is shown to improve with the increasing value of the recursive gain. A multiple time scale based analysis of the system elucidates the role of the recursive gain in enhancing the amount of dissipation produced by the control action. The influences of the time-delay and the control gain on the optimized performance of the system are also discussed. Numerical simulations of the system equations corroborate the analytical results. The present method is believed to be applicable to any self-excited system having a large degree of instability that is not removable by an ordinary time-delayed feedback.

  19. Recursive algorithms for two-dimensional smoothing using bicubic hermite polynomial

    NASA Technical Reports Server (NTRS)

    Kim, C. S.; Shen, C. N.

    1981-01-01

    It is noted that in the past, smoothing splines originated from approximation theory have been successfully applied to data filtering and image smoothing problems. Even though the nonrecursive technique of smoothing splines gives an optimal solution, the amount of computation increases rapidly with the size of the two-dimensional data. A derivation is presented here of quarter-plane filtering algorithms that provide smoothed estimates of function values and their derivatives by fitting two-dimensional smoothing splines in a recursive manner. The derivation procedure sheds light on specific problems encountered in two-dimensional filtering problems. What is more, the amount of computation for this recursive processor increases only linearly with the size of the two-dimensional data. Because of certain approximations introduced in its derivation, this recursive processor becomes suboptimal.

  20. Extended recursion in operator space (EROS), a new impurity solver for the single impurity Anderson model

    SciTech Connect

    Albers, Robert C; Julien, Jean P

    2008-01-01

    We have developed a new efficient and accurate impurity solver for the single impurity Anderson model (SIAM), which is based on a non-perturbative recursion technique in a space of operators and involves expanding the self-energy as a continued fraction. The method has no special occupation number or temperature restrictions; the only approximation is the number of levels of the continued fraction retained in the expansion. We also show how this approach can be used as a new approach to Dynamical Mean Field Theory (DMTF) and illustrate this with the Hubbard model. The three lowest orders of recursion give the Hartree-Fock, Hubbard I, and Hubbard III approximations. A higher level of recursion is able to reproduce the expected 3-peak structure in the spectral function and Fermi liquid behavior.

  1. Methods for assessing movement path recursion with application to African buffalo in South Africa

    USGS Publications Warehouse

    Bar-David, S.; Bar-David, I.; Cross, P.C.; Ryan, S.J.; Knechtel, C.U.; Getz, W.M.

    2009-01-01

    Recent developments of automated methods for monitoring animal movement, e.g., global positioning systems (GPS) technology, yield high-resolution spatiotemporal data. To gain insights into the processes creating movement patterns, we present two new techniques for extracting information from these data on repeated visits to a particular site or patch ("recursions"). Identification of such patches and quantification of recursion pathways, when combined with patch-related ecological data, should contribute to our understanding of the habitat requirements of large herbivores, of factors governing their space-use patterns, and their interactions with the ecosystem. We begin by presenting output from a simple spatial model that simulates movements of large-herbivore groups based on minimal parameters: resource availability and rates of resource recovery after a local depletion. We then present the details of our new techniques of analyses (recursion analysis and circle analysis) and apply them to data generated by our model, as well as two sets of empirical data on movements of African buffalo (Syncerus coffer): the first collected in Klaserie Private Nature Reserve and the second in Kruger National Park, South Africa. Our recursion analyses of model outputs provide us with a basis for inferring aspects of the processes governing the production of buffalo recursion patterns, particularly the potential influence of resource recovery rate. Although the focus of our simulations was a comparison of movement patterns produced by different resource recovery rates, we conclude our paper with a comprehensive discussion of how recursion analyses can be used when appropriate ecological data are available to elucidate various factors influencing movement. Inter alia, these include the various limiting and preferred resources, parasites, and topographical and landscape factors. ?? 2009 by the Ecological Society of America.

  2. The early acquisition of viable knowledge: A use of recursive model as an analytical devise (methodolosocial).

    PubMed

    Karstadt, Lyn; Thomas, Keith Robert; Abed, Shaymaa N

    2016-01-01

    In nurse education typically, information is presented to students within the classroom and then applied within a clinical situation. Acquisition of the knowledge required to inform the student's early practice is the focus of this research. This paper centres upon the construction of a cognitive model that is recursive in nature, and forms an integral part of a qualitative research study. The primary study investigated how first year student nurses use the information received in the classroom to underpin their early practice. Data were collected from 10 students and 4 of their lecturers, via blogs and interviews and used iteratively to create a model that is recursive in nature. Recursion is a process of repeatedly revisiting the same thing, in this case the data, which are considered in an iterative or progressive way. Recursion thus facilitated the development of a model, which was seen to change and develop in sophistication as more data were considered and evaluated. Visual devices were used throughout to bring clarity during the construction of the model. This visual process was pivotal to the analysis. This paper chronicles the development of an analytical device through the medium of the study presented. Viable knowledge is represented as the synthesis of concepts, as presented in the classroom, and practice, as experienced within the clinical area. It illustrates how conceptual material delivered within the classroom has become embedded within an individual student's consciousness and is used during a clinical placement to make sense of a specific situation. The study identifies how students use information and makes recommendations as to how appropriate curricula integrate all the facets of the recursive model. The process of recursive modelling is thus offered as an analytical devise, which may be applied by researchers to other qualitative data. PMID:26586185

  3. Recursive geometric integrators for wave propagation in a functionally graded multilayered elastic medium

    NASA Astrophysics Data System (ADS)

    Wang, Lugen; Rokhlin, S. I.

    2004-11-01

    The differential equations governing transfer and stiffness matrices and acoustic impedance for a functionally graded generally anisotropic magneto-electro-elastic medium have been obtained. It is shown that the transfer matrix satisfies a linear 1st order matrix differential equation, while the stiffness matrix satisfies a nonlinear Riccati equation. For a thin nonhomogeneous layer, approximate solutions with different levels of accuracy have been formulated in the form of a transfer matrix using a geometrical integration in the form of a Magnus expansion. This integration method preserves qualitative features of the exact solution of the differential equation, in particular energy conservation. The wave propagation solution for a thick layer or a multilayered structure of inhomogeneous layers is obtained recursively from the thin layer solutions. Since the transfer matrix solution becomes computationally unstable with increase of frequency or layer thickness, we reformulate the solution in the form of a stable stiffness-matrix solution which is obtained from the relation of the stiffness matrices to the transfer matrices. Using an efficient recursive algorithm, the stiffness matrices of the thin nonhomogeneous layer are combined to obtain the total stiffness matrix for an arbitrary functionally graded multilayered system. It is shown that the round-off error for the stiffness-matrix recursive algorithm is higher than that for the transfer matrices. To optimize the recursive procedure, a computationally stable hybrid method is proposed which first starts the recursive computation with the transfer matrices and then, as the thickness increases, transits to the stiffness matrix recursive algorithm. Numerical results show this solution to be stable and efficient. As an application example, we calculate the surface wave velocity dispersion for a functionally graded coating on a semispace.

  4. Modification of an amplification reaction in recursively dynamic compartments driven by stirring.

    PubMed

    Ichii, Tetsuo; Tanahashi, Genya; Suzuki, Hiroaki; Yomo, Tetsuya

    2013-12-17

    In living systems, biochemical reactions are confined to cellular or subcellular compartments, such as the plasma membrane and the organelles within a cell. These biological compartments are usually subjected to recursive changes, such as combinations of growth, fusion, and division, to constitute repeating cell cycles. In such recursively dynamic compartments, the encapsulated biochemical reaction may exhibit dynamics that differ from those of the static compartment (i.e., test tubes) used in conventional biochemistry experiments. To test this hypothesis in a simplified model, we mechanically stirred femtoliter-sized water-in-oil emulsion droplets so that individual droplets were subjected to repeated coalescence and breakage. We show that recursive dynamics appeared in the emulsion, which were measured by the exponential propagation of a water-soluble dye. The rate of the propagation, μ, was controlled by modulating the pulse-width of stirring in an electromagnetic stirrer. Within this system, we studied the dynamics of an RNA-amplification reaction in recursively increasing reaction compartments at various values of μ. We showed that there was an optimal value of μ that maximized RNA amplification. This effect was explained by the balance between the opposing effects of supply of substrate and the dilution of amplified RNA both resulting from coalescence. Moreover, when we mixed two RNA species with different kinetic properties, we found a preferential amplification for one of the species only in the recursively dynamic emulsion. This effect was partly explained by a separation effect which preferentially amplifies the number of compartments for the molecular specie that can better follow the breakage dynamics of the compartments. The present work demonstrated how the recursive dynamics of compartments modifies the internal biochemical reaction. PMID:24219119

  5. Methods for assessing movement path recursion with application to African buffalo in South Africa.

    PubMed

    Bar-David, Shirli; Bar-David, Israel; Cross, Paul C; Ryan, Sadie J; Knechtel, Christiane U; Getz, Wayne M

    2009-09-01

    Recent developments of automated methods for monitoring animal movement, e.g., global positioning systems (GPS) technology, yield high-resolution spatiotemporal data. To gain insights into the processes creating movement patterns, we present two new techniques for extracting information from these data on repeated visits to a particular site or patch ("recursions"). Identification of such patches and quantification of recursion pathways, when combined with patch-related ecological data, should contribute to our understanding of the habitat requirements of large herbivores, of factors governing their space-use patterns, and their interactions with the ecosystem. We begin by presenting output from a simple spatial model that simulates movements of large-herbivore groups based on minimal parameters: resource availability and rates of resource recovery after a local depletion. We then present the details of our new techniques of analyses (recursion analysis and circle analysis) and apply them to data generated by our model, as well as two sets of empirical data on movements of African buffalo (Syncerus caffer): the first collected in Klaserie Private Nature Reserve and the second in Kruger National Park, South Africa. Our recursion analyses of model outputs provide us with a basis for inferring aspects of the processes governing the production of buffalo recursion patterns, particularly the potential influence of resource recovery rate. Although the focus of our simulations was a comparison of movement patterns produced by different resource recovery rates, we conclude our paper with a comprehensive discussion of how recursion analyses can be used when appropriate ecological data are available to elucidate various factors influencing movement. Inter alia, these include the various limiting and preferred resources, parasites, and topographical and landscape factors. PMID:19769125

  6. [Pharmacovigilance update].

    PubMed

    Livio, F; Biollaz, J; Buclin, T

    2011-01-12

    Main pharmacovigilance updates are reviewed. Rosiglitazone and sibutramine have been suspended due to cardiovascular risks. The safety profile of H1N1 vaccines is similar to the established profile of seasonal influenza vaccines. Paroxetine reduces the benefit of tamoxifen. The use of serotoninergic antidepressants in pregnancy is still disputed. The risk of venous thromboembolism could be higher with oral combined contraceptives containing drospirenone compared to those containing levonorgestrel. Prolonged QT and PR intervals have been observed with saquinavir. The correct use of transdermal patches is reviewed with the example of rivastigmine. Aseptic meningitis is a rare adverse reaction of lamotrigine. An increased risk of fractures after long term use of proton pump inhibitors is suspected. PMID:21309181

  7. Polarization demultiplexing by recursive least squares constant modulus algorithm based on QR decomposition

    NASA Astrophysics Data System (ADS)

    Ling, Zhao; Yeling, Wang; Guijun, Hu; Yunpeng, Cui; Jian, Shi; Li, Li

    2013-07-01

    Recursive least squares constant modulus algorithm based on QR decomposition (QR-RLS-CMA) is first proposed as the polarization demultiplexing method. We compare its performance with the stochastic gradient descent constant modulus algorithm (SGD-CMA) and the recursive least squares constant modulus algorithm (RLS-CMA) in a polarization-division-multiplexing system with coherent detection. It is demonstrated that QR-RLS-CMA is an efficient demultiplexing algorithm which can avoid the problem of step-length choice in SGD-CMA. Meanwhile, it also has better symbol error rate (SER) performance and more stable convergence property.

  8. Fast and stable recursive algorithms for continuous-time and discrete-time model conversions

    NASA Technical Reports Server (NTRS)

    Shieh, L. S.; Lian, S. R.; Park, C. B.; Coleman, N. P.

    1987-01-01

    Based on the Newton-Raphson method, this paper presents recursive algorithms that are rapidly convergent and more stable for modeling the equivalent continuous-time (discrete-time) model from the available discrete-time (continuous-time) model for a fixed sampling period. The newly developed recursive algorithms relax the constraints imposed upon the existing model conversion algorithms, and, thus, enhance the applications of microprocessors and associated microelectronics to digital control systems. A practical example is presented to demonstrate the effectiveness of the proposed procedures.

  9. Recursive method for computing matrix elements for two-body interactions

    NASA Astrophysics Data System (ADS)

    Hyvärinen, Juhani; Suhonen, Jouni

    2015-05-01

    A recursive method for the efficient computation of two-body matrix elements is presented. The method consists of a set of recursion relations for the computationally demanding radial integral and adds one more tool to the set of computational methods introduced by Horie and Sasaki [H. Horie and K. Sasaki, Prog. Theor. Phys. 25, 475 (1961), 10.1143/PTP.25.475]. The neutrinoless double-β decay will serve as the primary application and example, but the method is general and can be applied equally well to other kinds of nuclear structure calculations involving matrix elements of two-body interactions.

  10. Recursive analytical solution describing artificial satellite motion perturbed by an arbitrary number of zonal terms

    NASA Technical Reports Server (NTRS)

    Mueller, A. C.

    1977-01-01

    An analytical first order solution has been developed which describes the motion of an artificial satellite perturbed by an arbitrary number of zonal harmonics of the geopotential. A set of recursive relations for the solution, which was deduced from recursive relations of the geopotential, was derived. The method of solution is based on Von-Zeipel's technique applied to a canonical set of two-body elements in the extended phase space which incorporates the true anomaly as a canonical element. The elements are of Poincare type, that is, they are regular for vanishing eccentricities and inclinations. Numerical results show that this solution is accurate to within a few meters after 500 revolutions.

  11. A new design for SLAM front-end based on recursive SOM

    NASA Astrophysics Data System (ADS)

    Yang, Xuesi; Xia, Shengping

    2015-12-01

    Aiming at the graph optimization-based monocular SLAM, a novel design for the front-end in single camera SLAM is proposed, based on the recursive SOM. Pixel intensities are directly used to achieve image registration and motion estimation, which can save time compared with the current appearance-based frameworks, usually including feature extraction and matching. Once a key-frame is identified, a recursive SOM is used to actualize loop-closure detecting, resulting a more precise location. The experiment on a public dataset validates our method on a computer with a quicker and effective result.

  12. Student Monks--Teaching Recursion in an IS or CS Programming Course Using the Towers of Hanoi

    ERIC Educational Resources Information Center

    Benander, Alan C.; Benander, Barbara A.

    2008-01-01

    Educators have been using the Towers of Hanoi problem for many years as an example of a problem that has a very elegant recursive solution. However, the elegance and conciseness of this solution can make it difficult for students to understand the amount of computer time required in the execution of this solution. And, like many recursive computer…

  13. Lapin Data Interchange Among Database, Analysis and Display Programs Using XML-Based Text Files

    NASA Technical Reports Server (NTRS)

    2005-01-01

    The purpose of grant NCC3-966 was to investigate and evaluate the interchange of application-specific data among multiple programs each carrying out part of the analysis and design task. This has been carried out previously by creating a custom program to read data produced by one application and then write that data to a file whose format is specific to the second application that needs all or part of that data. In this investigation, data of interest is described using the XML markup language that allows the data to be stored in a text-string. Software to transform output data of a task into an XML-string and software to read an XML string and extract all or a portion of the data needed for another application is used to link two independent applications together as part of an overall design effort. This approach was initially used with a standard analysis program, Lapin, along with standard applications a standard spreadsheet program, a relational database program, and a conventional dialog and display program to demonstrate the successful sharing of data among independent programs. Most of the effort beyond that demonstration has been concentrated on the inclusion of more complex display programs. Specifically, a custom-written windowing program organized around dialogs to control the interactions have been combined with an independent CAD program (Open Cascade) that supports sophisticated display of CAD elements such as lines, spline curves, and surfaces and turbine-blade data produced by an independent blade design program (UD0300).

  14. A structured interface to the object-oriented genomics unified schema for XML-formatted data.

    PubMed

    Clark, Terry; Jurek, Josef; Kettler, Gregory; Preuss, Daphe

    2005-01-01

    Data management systems are fast becoming required components in many biology laboratories as the role of computer-based information grows. Although the need for data management systems is on the rise, their inherent complexities can deter the full and routine use of their computational capabilities. The significant undertaking to implement a capable production system can be reduced in part by adapting an established data management system. In such a way, we are leveraging the Genomics Unified Schema (GUS) developed at the Computational Biology and Informatics Laboratory at the University of Pennsylvania as a foundation for managing and analysing DNA sequence data in centromere research projects around Arabidopsis thaliana and related species. Because GUS provides a core schema that includes support for genome sequences, mRNA and its expression, and annotated chromosomes, it is ideal for synthesising a variety of parameters to analyse these repetitive and highly dynamic portions of the genome. Despite this, production-strength data management frameworks are complex, requiring dedicated efforts to adapt and maintain. The work reported in this article addresses one component of such an effort, namely the pivotal task of marshalling data from various sources into GUS. In order to harness GUS for our project, and motivated by efficiency needs, we developed a structured framework for transferring data into GUS from outside sources. This technology is embodied in a GUS object-layer processor, XMLGUS. XMLGUS facilitates incorporating data into GUS by (i) formulating an XML interface that includes relational database key constraint definitions, (ii) regularising traversal through that XML, (iii) realising automatic processing of the XML with database key constraints and (iv) allowing for special processing of input data within the framework for automated processing. The application of XMLGUS to production pipeline processing for a sequencing project and inputting the Arabidopsis genome into GUS is discussed. XMLGUS is available from the Flora website (http://flora.ittc.ku.edu/). PMID:16000009

  15. XML&CORBA-based distributed system for information integration and dissemination in multimedia neuroimaging applications

    NASA Astrophysics Data System (ADS)

    Cao, Xinhua; Wong, Stephen T. C.; Hoo, Kent S., Jr.; Ching, Wan; Hawkins, Randall A.; Dillon, William P.

    2002-05-01

    Medical information systems based in different hospital departments face tremendous difficulty in information exchange and dissemination due to the multitudes of hardware and software platforms running these systems. In this paper, we describe a distributed information system for integrating various hospital systems in supporting clinical neuroimaging research and epilepsy surgical planning. Our distributed information system uses a three-tiered architecture consists of a user-interface tier, application logic tier and data store tier. Two system implementations based on this software architecture but using different integration technologies were developed and are discussed in this paper: the XML (extensible Mark Up Language)-based implementation and the CORBA (Common Object Request Broker Architecture)-based implementation. In the XML-based implementation, application logic tier communicates with user-interface tier and data store tier using HTTP and XML for data exchange. For the data exchange in CORBA-based implementation, the middleware uses IIOP (Internet inter-ORB protocol) to call CORBA objects in the data store tier then to return the results to the user-interface tier. For the user-interface tier of either implementation, the Web browsers are served as clients to invoke application components or agents in the middleware. The application of the proposed distributed system allows clinical users to access, search and retrieve the multimedia information in any underlying computer systems with commonly used Web browsers. Preliminary results show that the system is effective for information integration and data sharing among the different departmental systems in the hospital for neuroimaging applications.

  16. Lapin Data Interchange Among Database, Analysis and Display Programs Using XML-Based Text Files

    NASA Technical Reports Server (NTRS)

    2004-01-01

    The purpose was to investigate and evaluate the interchange of application- specific data among multiple programs each carrying out part of the analysis and design task. This has been carried out previously by creating a custom program to read data produced by one application and then write that data to a file whose format is specific to the second application that needs all or part of that data. In this investigation, data of interest is described using the XML markup language that allows the data to be stored in a text-string. Software to transform output data of a task into an XML-string and software to read an XML string and extract all or a portion of the data needed for another application is used to link two independent applications together as part of an overall design effort. This approach was initially used with a standard analysis program, Lapin, along with standard applications a standard spreadsheet program, a relational database program, and a conventional dialog and display program to demonstrate the successful sharing of data among independent programs. See Engineering Analysis Using a Web-Based Protocol by J.D. Schoeffler and R.W. Claus, NASA TM-2002-211981, October 2002. Most of the effort beyond that demonstration has been concentrated on the inclusion of more complex display programs. Specifically, a custom-written windowing program organized around dialogs to control the interactions have been combined with an independent CAD program (Open Cascade) that supports sophisticated display of CAD elements such as lines, spline curves, and surfaces and turbine-blade data produced by an independent blade design program (UD0300).

  17. The recursive maximum likelihood proportion estimator: User's guide and test results

    NASA Technical Reports Server (NTRS)

    Vanrooy, D. L.

    1976-01-01

    Implementation of the recursive maximum likelihood proportion estimator is described. A user's guide to programs as they currently exist on the IBM 360/67 at LARS, Purdue is included, and test results on LANDSAT data are described. On Hill County data, the algorithm yields results comparable to the standard maximum likelihood proportion estimator.

  18. Closed-form recursive formula for an optimal tracker with terminal constraints

    NASA Technical Reports Server (NTRS)

    Juang, J.-N.; Turner, J. D.; Chun, H. M.

    1984-01-01

    Feedback control laws are derived for a class of optimal finite time tracking problems with terminal constraints. Analytical solutions are obtained for the feedback gain and the closed-loop response trajectory. Such formulations are expressed in recursive forms so that a real-time computer implementation becomes feasible. Two examples are given to illustrate the validity and usefulness of the formulations.

  19. Replicating Freeman's Recursive Adjustment Model of Demand for Higher Education. AIR Forum 1979 Paper.

    ERIC Educational Resources Information Center

    Wish, John R.; Hamilton, William D.

    The viability of Freeman's Recursive Adjustment Model as a management tool for explaining and predicting enrollments and the job market is examined using macro and micro cases. To establish a relationship between college attendance and the college job market, Freeman employed a 3-equation model to analyze the causative supply and demand…

  20. User's Guide for the Precision Recursive Estimator for Ephemeris Refinement (PREFER)

    NASA Technical Reports Server (NTRS)

    Gibbs, B. P.

    1982-01-01

    PREFER is a recursive orbit determination program which is used to refine the ephemerides produced by a batch least squares program (e.g., GTDS). It is intended to be used primarily with GTDS and, thus, is compatible with some of the GTDS input/output files.

  1. Cross-Validation of Survival Bump Hunting by Recursive Peeling Methods

    PubMed Central

    Dazard, Jean-Eudes; Choe, Michael; LeBlanc, Michael; Rao, J. Sunil

    2015-01-01

    We introduce a survival/risk bump hunting framework to build a bump hunting model with a possibly censored time-to-event type of response and to validate model estimates. First, we describe the use of adequate survival peeling criteria to build a survival/risk bump hunting model based on recursive peeling methods. Our method called “Patient Recursive Survival Peeling” is a rule-induction method that makes use of specific peeling criteria such as hazard ratio or log-rank statistics. Second, to validate our model estimates and improve survival prediction accuracy, we describe a resampling-based validation technique specifically designed for the joint task of decision rule making by recursive peeling (i.e. decision-box) and survival estimation. This alternative technique, called “combined” cross-validation is done by combining test samples over the cross-validation loops, a design allowing for bump hunting by recursive peeling in a survival setting. We provide empirical results showing the importance of cross-validation and replication. PMID:26997922

  2. X-ray resonant reflection from magnetic multilayers: Recursion matrix algorithm

    SciTech Connect

    Stepanov, S. A.; Sinha, S. K.

    2000-06-01

    Recursion equations for 2x2 scattering matrices have been derived to calculate resonant x-ray reflection from magnetic multilayers. The solution has been basically reduced to that found in Stepanov et al, Phys. Rev. B 57, 4829 (1998) for grazing incidence x-ray diffraction from crystalline multilayers. (c) 2000 The American Physical Society.

  3. Quantitative genetic models for describing simultaneous and recursive relationships between phenotypes.

    PubMed Central

    Gianola, Daniel; Sorensen, Daniel

    2004-01-01

    Multivariate models are of great importance in theoretical and applied quantitative genetics. We extend quantitative genetic theory to accommodate situations in which there is linear feedback or recursiveness between the phenotypes involved in a multivariate system, assuming an infinitesimal, additive, model of inheritance. It is shown that structural parameters defining a simultaneous or recursive system have a bearing on the interpretation of quantitative genetic parameter estimates (e.g., heritability, offspring-parent regression, genetic correlation) when such features are ignored. Matrix representations are given for treating a plethora of feedback-recursive situations. The likelihood function is derived, assuming multivariate normality, and results from econometric theory for parameter identification are adapted to a quantitative genetic setting. A Bayesian treatment with a Markov chain Monte Carlo implementation is suggested for inference and developed. When the system is fully recursive, all conditional posterior distributions are in closed form, so Gibbs sampling is straightforward. If there is feedback, a Metropolis step may be embedded for sampling the structural parameters, since their conditional distributions are unknown. Extensions of the model to discrete random variables and to nonlinear relationships between phenotypes are discussed. PMID:15280252

  4. A Recursive Theory for the Mathematical Understanding--Some Elements and Implications.

    ERIC Educational Resources Information Center

    Pirie, Susan; Kieren, Thomas

    There has been considerable interest in mathematical understanding. Both those attempting to build, and those questioning the possibility of building intelligent artificial tutoring systems, struggle with the notions of mathematical understanding. The purpose of this essay is to show a transcendently recursive theory of mathematical understanding…

  5. Recursive inverse kinematics for robot arms via Kalman filtering and Bryson-Frazier smoothing

    NASA Technical Reports Server (NTRS)

    Rodriguez, G.; Scheid, R. E., Jr.

    1987-01-01

    This paper applies linear filtering and smoothing theory to solve recursively the inverse kinematics problem for serial multilink manipulators. This problem is to find a set of joint angles that achieve a prescribed tip position and/or orientation. A widely applicable numerical search solution is presented. The approach finds the minimum of a generalized distance between the desired and the actual manipulator tip position and/or orientation. Both a first-order steepest-descent gradient search and a second-order Newton-Raphson search are developed. The optimal relaxation factor required for the steepest descent method is computed recursively using an outward/inward procedure similar to those used typically for recursive inverse dynamics calculations. The second-order search requires evaluation of a gradient and an approximate Hessian. A Gauss-Markov approach is used to approximate the Hessian matrix in terms of products of first-order derivatives. This matrix is inverted recursively using a two-stage process of inward Kalman filtering followed by outward smoothing. This two-stage process is analogous to that recently developed by the author to solve by means of spatial filtering and smoothing the forward dynamics problem for serial manipulators.

  6. Recursive Frame Analysis: Reflections on the Development of a Qualitative Research Method

    ERIC Educational Resources Information Center

    Keeney, Hillary; Keeney, Bradford

    2012-01-01

    The origin of recursive frame analysis (RFA) is revisited and discussed as a postmodern alternative to modernist therapeutic models and research methods that foster hegemony of a preferred therapeutic metaphor, narrative, or strategy. It encourages improvisational performance while enabling a means of scoring the change and movement of the…

  7. Split-remerge method for eliminating processing window artifacts in recursive hierarchical segmentation

    NASA Technical Reports Server (NTRS)

    Tilton, James C. (Inventor)

    2010-01-01

    A method, computer readable storage, and apparatus for implementing recursive segmentation of data with spatial characteristics into regions including splitting-remerging of pixels with contagious region designations and a user controlled parameter for providing a preference for merging adjacent regions to eliminate window artifacts.

  8. Mean-Field Analysis of Recursive Entropic Segmentation of Biological Sequences

    NASA Astrophysics Data System (ADS)

    Cheong, Siew-Ann; Stodghill, Paul; Schneider, David; Myers, Christopher

    2007-03-01

    Horizontal gene transfer in bacteria results in genomic sequences which are mosaic in nature. An important first step in the analysis of a bacterial genome would thus be to model the statistically nonstationary nucleotide or protein sequence with a collection of P stationary Markov chains, and partition the sequence of length N into M statistically stationary segments/domains. This can be done for Markov chains of order K = 0 using a recursive segmentation scheme based on the Jensen-Shannon divergence, where the unknown parameters P and M are estimated from a hypothesis testing/model selection process. In this talk, we describe how the Jensen-Shannon divergence can be generalized to Markov chains of order K > 0, as well as an algorithm optimizing the positions of a fixed number of domain walls. We then describe a mean field analysis of the generalized recursive Jensen-Shannon segmentation scheme, and show how most domain walls appear as local maxima in the divergence spectrum of the sequence, before highlighting the main problem associated with the recursive segmentation scheme, i.e. the strengths of the domain walls selected recursively do not decrease monotonically. This problem is especially severe in repetitive sequences, whose statistical signatures we will also discuss.

  9. Evapotranspiration: Measured with a lysimeter vs. calculated with a recursive method

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Recently, a recursive combination method (RCM) to calculate potential and crop evapotranspiration (ET) was given by Lascano and Van Bavel (Agron. J. 2007, 99:585-590) that differs from the Penman-Monteith (PM) method. The main difference between the two methods is that the assumptions made regarding...

  10. A General Recursion Method for Calculating Diffracted Intensities from Crystals Containing Planar Faults

    NASA Astrophysics Data System (ADS)

    Treacy, M. M. J.; Newsam, J. M.; Deem, M. W.

    1991-06-01

    A general recursion algorithm is described for calculating kinematical diffraction intensities from crystals containing coherent planar faults. The method exploits the self-similar stacking sequences that occur when layers stack non-deterministically. Recursion gives a set of simple relations between average interference terms from a statistical crystal, which can be solved as a set of simultaneous equations. The diffracted intensity for a polycrystalline sample is given by the incoherent sum of scattered intensities over an ensemble of crystallites. The relations between this and previous approaches, namely the Hendricks-Teller matrix formulation, the difference equation method, the summed series formula of Cowley, and Michalski's recurrence relations between average phase factors, are discussed. Although formally identical to these previous methods, the present recursive description has an intuitive appeal and proves easier to apply to complex crystal structure types. The method is valid for all types of planar faults, can accommodate long-range stacking correlations, and is applicable to crystals that contain only a finite number of layers. A FORTRAN program DIFFaX, based on this recursion algorithm, has been written and used to simulate powder X-ray (and neutron) diffraction patterns and single crystal electron (kinematical) diffraction patterns. Calculations for diamond-lonsdaleite and for several synthetic zeolite systems that contain high densities of stacking faults are presented as examples.

  11. Web-Based Delivery System for Disaster Prevention Information Using a New Jma Dpi Xml Format and Amedas Data

    NASA Astrophysics Data System (ADS)

    Nishio, M.; Mori, M.

    2012-07-01

    The Automated Meteorological Data Acquisition System (AMeDAS) Data is used along with compound disaster information for a geographic information system (GIS) by integration into the Japan Meteorological Agency (JMA) disaster prevention information XML data. A JMA XML format is a next generation format that contains weather warnings, tsunami warnings, and earthquake information, etc. However, it is not possible to process it by reading disaster prevention information XML Data and AMeDAS Data directly to the GIS system. Therefore, development of a program that converts the data structure is important to consolidate a variety of disaster prevention information on the GIS system. Information on escape routes and evacuation sites, etc. were given as points for regional meteorological observation forecasts using AMeDAS Data by disaster prevention information XML data and integrating it where the disaster was generated, giving a range of expansion of damage and a damage level. There are two main aims; the first is to deliver these compound data of disaster prevention information XML data and AMeDAS Data via the Internet. The second aim is to provide GIS files (shapefile format) of these data to such as local governments for their individual analysis. This was furthermore confirmed to enable the construction of a system using WebGIS (Google Maps) and Open Source Software GIS to monitor disaster information at low cost.

  12. A Transcription System from MusicXML Format to Braille Music Notation

    NASA Astrophysics Data System (ADS)

    Goto, D.; Gotoh, T.; Minamikawa-Tachino, R.; Tamura, N.

    2006-12-01

    The Internet enables us to freely access music as recorded sound and even music scores. For the visually impaired, music scores must be transcribed from computer-based musical formats to Braille music notation. This paper proposes a transcription system from the MusicXML format to Braille music notation using a structural model of Braille music notation. The resultant Braille scores inspected by volunteer transcribers are up to the international standard. Using this simple and efficient transcription system, it should be possible to provide Braille music scores via the Internet to the visually impaired.

  13. Some applications of recursive functionals to the foundations of mathematics and physics

    NASA Astrophysics Data System (ADS)

    Szudzik, Matthew P.

    We consider two applications of recursive functionals. The first application concerns Godel's theory T , which provides a rudimentary foundation for the formalization of mathematics. T can be understood as a theory of the simply-typed lambda calculus that is extended to include the constant 0N, the successor function S, and the operator Rtau for primitive recursion on objects of type tau. It is known that the functions from non-negative integers to non-negative integers that can be defined in this theory are exactly the < epsilon 0-recursive functions of non-negative integers. But it is not well-known which functionals of arbitrary type can be defined in T . We show that when the domain and codomain are restricted to pure closed normal forms, the functionals of arbitrary type that are definable in T are exactly those functionals that can be encoded as < epsilon 0-recursive functions of non-negative integers. This result has many interesting consequences, including a new characterization of T . The second application is concerned with the question: "When can a model of a physical system be regarded as computable?" We provide the definition of a computable physical model to answer this question. The connection between our definition and Kreisel's notion of a mechanistic theory is discussed, and several examples of computable physical models are given, including models which feature discrete motion, a model which features non-discrete continuous motion, and non-deterministic models such as radioactive decay. We show how computable physical models on effective topological spaces can be formulated with recursive functionals in the theory of type-two effectivity (TTE). Various common operations on computable physical models are described, such as the operation of coarse-graining and the formation of statistical ensembles. The definition of a computable physical model also allows for a precise formalization of the computable universe hypothesis---the claim that all the laws of physics are computable.

  14. Developing the earthquake markup language and database with UML and XML schema

    NASA Astrophysics Data System (ADS)

    Babaie, Hassan A.; Babaei, Abbed

    2005-11-01

    Seismic information is commonly shared across networks, and stored in heterogeneous databases that are maintained by different and autonomous groups. Because the metadata of the seismic information is generally not structured, information loss commonly occurs when data is transmitted from one source to another. Minimizing such information loss and enhancing information interchange over the Web require construction of specific markup languages for each part of seismology. The sub-disciplinary data syntax and vocabularies allow construction of corresponding relational databases to store and interchange the structured data. Developing modular data structures and databases for each sub-discipline allows composition of new languages through reuse, and enables future extension of the vocabulary and maintenance of the knowledge base. In this paper, we show how to design the conceptual and logical models for a small set of seismology that deals with earthquake ( EarthquakeML). We discuss the application of the UML to the design and mapping of the object model into W3 XML Schema and database. The emphasis is on designing components that can be reused across discipline, and reduce redundancies by developing type libraries. We show how to map the XML schema into a relational data model; and implement the EarthquakeDB with the Microsoft Access DBMS.

  15. Use of XML and Java for collaborative petroleum reservoir modeling on the Internet

    USGS Publications Warehouse

    Victorine, J.; Watney, W.L.; Bhattacharya, S.

    2005-01-01

    The GEMINI (Geo-Engineering Modeling through INternet Informatics) is a public-domain, web-based freeware that is made up of an integrated suite of 14 Java-based software tools to accomplish on-line, real-time geologic and engineering reservoir modeling. GEMINI facilitates distant collaborations for small company and academic clients, negotiating analyses of both single and multiple wells. The system operates on a single server and an enterprise database. External data sets must be uploaded into this database. Feedback from GEMINI users provided the impetus to develop Stand Alone Web Start Applications of GEMINI modules that reside in and operate from the user's PC. In this version, the GEMINI modules run as applets, which may reside in local user PCs, on the server, or Java Web Start. In this enhanced version, XML-based data handling procedures are used to access data from remote and local databases and save results for later access and analyses. The XML data handling process also integrates different stand-alone GEMINI modules enabling the user(s) to access multiple databases. It provides flexibility to the user to customize analytical approach, database location, and level of collaboration. An example integrated field-study using GEMINI modules and Stand Alone Web Start Applications is provided to demonstrate the versatile applicability of this freeware for cost-effective reservoir modeling. ?? 2005 Elsevier Ltd. All rights reserved.

  16. Presentation of problem-specific, text-based medical knowledge: XML and related technologies.

    PubMed

    Hoelzer, S; Boettcher, H; Schweiger, R K; Konetschny, J; Dudeck, J

    2001-01-01

    Systematic reviews of the impact of clinical decision support systems on provider behavior and patient outcome have shown evidence of benefit. Knowledge-based functions for decision support or monitoring that are integrated in clinical information systems are a potentially effective way. But these concepts are restricted by the efforts required for development and maintenance of the information systems and the limited number of implemented medical rules. Physicians are familiar to get their information from text-based sources. It seems to be straight-forward to rely on a document-based solution in order to present problem-specific information at the point of care. We have developed a concept for context-sensitive retrieving and presentation of text-based medical knowledge (textbook of internal medicine) using the eXtensible Markup Language (XML) and related technologies. This concept can facilitate the electronic query and presentation of this resource. XML may replace narrative text as a storage format and allows to structure the data in a stepwise fashion. On the basis of structured data we are able to improve the search quality for clinical information and its presentation which forms a crucial pre-requisite for the use of the information and the implementation of evidence-based care in the clinical routine. PMID:11825191

  17. The tissue microarray data exchange specification: A document type definition to validate and enhance XML data

    PubMed Central

    Nohle, David G; Ayers, Leona W

    2005-01-01

    Background The Association for Pathology Informatics (API) Extensible Mark-up Language (XML) TMA Data Exchange Specification (TMA DES) proposed in April 2003 provides a community-based, open source tool for sharing tissue microarray (TMA) data in a common format. Each tissue core within an array has separate data including digital images; therefore an organized, common approach to produce, navigate and publish such data facilitates viewing, sharing and merging TMA data from different laboratories. The AIDS and Cancer Specimen Resource (ACSR) is a HIV/AIDS tissue bank consortium sponsored by the National Cancer Institute (NCI) Division of Cancer Treatment and Diagnosis (DCTD). The ACSR offers HIV-related malignancies and uninfected control tissues in microarrays (TMA) accompanied by de-identified clinical data to approved researchers. Exporting our TMA data into the proposed API specified format offers an opportunity to evaluate the API specification in an applied setting and to explore its usefulness. Results A document type definition (DTD) that governs the allowed common data elements (CDE) in TMA DES export XML files was written, tested and evolved and is in routine use by the ACSR. This DTD defines TMA DES CDEs which are implemented in an external file that can be supplemented by internal DTD extensions for locally defined TMA data elements (LDE). Conclusion ACSR implementation of the TMA DES demonstrated the utility of the specification and allowed application of a DTD to validate the language of the API specified XML elements and to identify possible enhancements within our TMA data management application. Improvements to the specification have additionally been suggested by our experience in importing other institution's exported TMA data. Enhancements to TMA DES to remove ambiguous situations and clarify the data should be considered. Better specified identifiers and hierarchical relationships will make automatic use of the data possible. Our tool can be used to reorder data and add identifiers; upgrading data for changes in the specification can be automatically accomplished. Using a DTD (optionally reflecting our proposed enhancements) can provide stronger validation of exported TMA data. PMID:15871741

  18. Berends-Giele recursions and the BCJ duality in superspace and components

    NASA Astrophysics Data System (ADS)

    Mafra, Carlos R.; Schlotterer, Oliver

    2016-03-01

    The recursive method of Berends and Giele to compute tree-level gluon amplitudes is revisited using the framework of ten-dimensional super Yang-Mills. First, we prove that the pure spinor formula to compute SYM tree amplitudes derived in 2010 reduces to the standard Berends-Giele formula from the 80s when restricted to gluon amplitudes and additionally determine the fermionic completion. Second, using BRST cohomology manipulations in superspace, alternative representations of the component amplitudes are explored and the Bern-Carrasco-Johansson relations among partial tree amplitudes are derived in a novel way. Finally, it is shown how the supersymmetric components of manifestly local BCJ-satisfying tree-level numerators can be computed in a recursive fashion.

  19. A recursive vesicle-based model protocell with a primitive model cell cycle

    NASA Astrophysics Data System (ADS)

    Kurihara, Kensuke; Okura, Yusaku; Matsuo, Muneyuki; Toyota, Taro; Suzuki, Kentaro; Sugawara, Tadashi

    2015-09-01

    Self-organized lipid structures (protocells) have been proposed as an intermediate between nonliving material and cellular life. Synthetic production of model protocells can demonstrate the potential processes by which living cells first arose. While we have previously described a giant vesicle (GV)-based model protocell in which amplification of DNA was linked to self-reproduction, the ability of a protocell to recursively self-proliferate for multiple generations has not been demonstrated. Here we show that newborn daughter GVs can be restored to the status of their parental GVs by pH-induced vesicular fusion of daughter GVs with conveyer GVs filled with depleted substrates. We describe a primitive model cell cycle comprising four discrete phases (ingestion, replication, maturity and division), each of which is selectively activated by a specific external stimulus. The production of recursive self-proliferating model protocells represents a step towards eventual production of model protocells that are able to mimic evolution.

  20. Recursive dynamics of topological trees of rigid bodies via Kalman filtering and Bryson-Frazier smoothing

    NASA Technical Reports Server (NTRS)

    Rodriguez, G.

    1988-01-01

    The inverse and forward dynamics problems for a set of rigid bodies connected by hinges to form a topological tree are solved by using recursive techniques from linear filtering and smoothing theory. An inward filtering sequence computes a set of constraint moments and forces. This is followed by an outward sequence to determine a corresponding set of angular and linear accelerations. An inward sequence begins at the tips of all of the terminal bodies of the tree and proceeds inwardly through all of the branches until it reaches the root. Similarly, an outward sequence begins at the root and propagates to all of the tree branches until it reaches the tips of the terminal bodies. The paper also provides an approach to evaluate recursively the composite multibody system inertia matrix and its inverse.

  1. Non-recursive augmented Lagrangian algorithms for the forward and inverse dynamics of constrained flexible multibodies

    NASA Technical Reports Server (NTRS)

    Bayo, Eduardo; Ledesma, Ragnar

    1993-01-01

    A technique is presented for solving the inverse dynamics of flexible planar multibody systems. This technique yields the non-causal joint efforts (inverse dynamics) as well as the internal states (inverse kinematics) that produce a prescribed nominal trajectory of the end effector. A non-recursive global Lagrangian approach is used in formulating the equations for motion as well as in solving the inverse dynamics equations. Contrary to the recursive method previously presented, the proposed method solves the inverse problem in a systematic and direct manner for both open-chain as well as closed-chain configurations. Numerical simulation shows that the proposed procedure provides an excellent tracking of the desired end effector trajectory.

  2. Recursive Double-Size Modular Multiplications from Euclidean and Montgomery Multipliers

    NASA Astrophysics Data System (ADS)

    Yoshino, Masayuki; Okeya, Katsuyuki; Vuillaume, Camille

    A technique for computing the quotient (⌊ ab/n ⌋) of Euclidean divisions from the difference of two remainders (ab (mod n) - ab (mod n+1)) was proposed by Fischer and Seifert. The technique allows a 2l-bit modular multiplication to work on most l-bit modular multipliers. However, the cost of the quotient computation rises sharply when computing modular multiplications larger than 2l bits with a recursive approach. This paper addresses the computation cost and improves on previous 2l-bit modular multiplication algorithms to return not only the remainder but also the quotient, resulting in an higher performance in the recursive approach, which becomes twice faster in the quadrupling case and four times faster in the octupling case. In addition to Euclidean multiplication, this paper proposes a new 2l-bit Montgomery multiplication algorithm to return both of the remainder and the quotient.

  3. A recursive vesicle-based model protocell with a primitive model cell cycle

    PubMed Central

    Kurihara, Kensuke; Okura, Yusaku; Matsuo, Muneyuki; Toyota, Taro; Suzuki, Kentaro; Sugawara, Tadashi

    2015-01-01

    Self-organized lipid structures (protocells) have been proposed as an intermediate between nonliving material and cellular life. Synthetic production of model protocells can demonstrate the potential processes by which living cells first arose. While we have previously described a giant vesicle (GV)-based model protocell in which amplification of DNA was linked to self-reproduction, the ability of a protocell to recursively self-proliferate for multiple generations has not been demonstrated. Here we show that newborn daughter GVs can be restored to the status of their parental GVs by pH-induced vesicular fusion of daughter GVs with conveyer GVs filled with depleted substrates. We describe a primitive model cell cycle comprising four discrete phases (ingestion, replication, maturity and division), each of which is selectively activated by a specific external stimulus. The production of recursive self-proliferating model protocells represents a step towards eventual production of model protocells that are able to mimic evolution. PMID:26418735

  4. A Note on Local Stability Conditions for Two Types of Monetary Models with Recursive Utility

    NASA Astrophysics Data System (ADS)

    Miyazaki, Kenji; Utsunomiya, Hitoshi

    2009-09-01

    This note explores local stability conditions for money-in-utility-function (MIUF) and transaction-costs (TC) models with recursive utility. Although Chen et al. [Chen, B.-L., M. Hsu, and C.-H. Lin, 2008, Inflation and growth: impatience and a qualitative equivalent, Journal of Money, Credit, and Banking, Vol. 40, No. 6, 1310-1323] investigated the relationship between inflation and growth in MIUF and TC models with recursive utility, they conducted only a comparative static analysis in a steady state. By establishing sufficient conditions for local stability, this note proves that impatience should be increasing in consumption and real balances. Increasing impatience, although less plausible from an empirical point of view, receives more support from a theoretical viewpoint.

  5. Modifications to iterative recursion unfolding algorithms and computer codes to find more appropriate neutron spectra. Final report

    SciTech Connect

    Lowry, K.A.; Johnson, T.L.

    1984-06-06

    Modifications were made to iterative recursion unfolding algorithms and computer codes which allow the incorporation of prior knowledge, such as cutoff energy or preferred spectral shape, into the solution. Using these modifications, more appropriate neutron spectra are obtained.

  6. Closed-form recursive formula for an optimal tracker with terminal constraints

    NASA Technical Reports Server (NTRS)

    Juang, J. N.; Turner, J. D.; Chun, H. M.

    1986-01-01

    Feedback control laws are derived for a class of optimal finite time tracking problems with terminal constraints. Analytical solutions are obtained for the feedback gain and the closed-loop response trajectory. Such formulations are expressed in recursive forms so that a real-time computer implementation becomes feasible. An example involving the feedback slewing of a flexible spacecraft is given to illustrate the validity and usefulness of the formulations.

  7. Systematic generation of multibody equations of motion suitable for recursive and parallel manipulation

    NASA Technical Reports Server (NTRS)

    Nikravesh, Parviz E.; Gim, Gwanghum; Arabyan, Ara; Rein, Udo

    1989-01-01

    The formulation of a method known as the joint coordinate method for automatic generation of the equations of motion for multibody systems is summarized. For systems containing open or closed kinematic loops, the equations of motion can be reduced systematically to a minimum number of second order differential equations. The application of recursive and nonrecursive algorithms to this formulation, computational considerations and the feasibility of implementing this formulation on multiprocessor computers are discussed.

  8. Recursive Estimation of the Stein Center of SPD Matrices & its Applications*

    PubMed Central

    Salehian, Hesamoddin; Cheng, Guang; Ho, Jeffrey

    2014-01-01

    Symmetric positive-definite (SPD) matrices are ubiquitous in Computer Vision, Machine Learning and Medical Image Analysis. Finding the center/average of a population of such matrices is a common theme in many algorithms such as clustering, segmentation, principal geodesic analysis, etc. The center of a population of such matrices can be defined using a variety of distance/divergence measures as the minimizer of the sum of squared distances/divergences from the unknown center to the members of the population. It is well known that the computation of the Karcher mean for the space of SPD matrices which is a negatively-curved Riemannian manifold is computationally expensive. Recently, the LogDet divergence-based center was shown to be a computationally attractive alternative. However, the LogDet-based mean of more than two matrices can not be computed in closed form, which makes it computationally less attractive for large populations. In this paper we present a novel recursive estimator for center based on the Stein distance – which is the square root of the LogDet divergence – that is significantly faster than the batch mode computation of this center. The key theoretical contribution is a closed-form solution for the weighted Stein center of two SPD matrices, which is used in the recursive computation of the Stein center for a population of SPD matrices. Additionally, we show experimental evidence of the convergence of our recursive Stein center estimator to the batch mode Stein center. We present applications of our recursive estimator to K-means clustering and image indexing depicting significant time gains over corresponding algorithms that use the batch mode computations. For the latter application, we develop novel hashing functions using the Stein distance and apply it to publicly available data sets, and experimental results have shown favorable comparisons to other competing methods. PMID:25350135

  9. Recursive Estimation of the Stein Center of SPD Matrices & its Applications.

    PubMed

    Salehian, Hesamoddin; Cheng, Guang; Vemuri, Baba C; Ho, Jeffrey

    2013-12-01

    Symmetric positive-definite (SPD) matrices are ubiquitous in Computer Vision, Machine Learning and Medical Image Analysis. Finding the center/average of a population of such matrices is a common theme in many algorithms such as clustering, segmentation, principal geodesic analysis, etc. The center of a population of such matrices can be defined using a variety of distance/divergence measures as the minimizer of the sum of squared distances/divergences from the unknown center to the members of the population. It is well known that the computation of the Karcher mean for the space of SPD matrices which is a negatively-curved Riemannian manifold is computationally expensive. Recently, the LogDet divergence-based center was shown to be a computationally attractive alternative. However, the LogDet-based mean of more than two matrices can not be computed in closed form, which makes it computationally less attractive for large populations. In this paper we present a novel recursive estimator for center based on the Stein distance - which is the square root of the LogDet divergence - that is significantly faster than the batch mode computation of this center. The key theoretical contribution is a closed-form solution for the weighted Stein center of two SPD matrices, which is used in the recursive computation of the Stein center for a population of SPD matrices. Additionally, we show experimental evidence of the convergence of our recursive Stein center estimator to the batch mode Stein center. We present applications of our recursive estimator to K-means clustering and image indexing depicting significant time gains over corresponding algorithms that use the batch mode computations. For the latter application, we develop novel hashing functions using the Stein distance and apply it to publicly available data sets, and experimental results have shown favorable comparisons to other competing methods. PMID:25350135

  10. Updating Situation Models

    ERIC Educational Resources Information Center

    Zwaan, Rolf A.; Madden, Carol J.

    2004-01-01

    The authors examined how situation models are updated during text comprehension. If comprehenders keep track of the evolving situation, they should update their models such that the most current information, the here and now, is more available than outdated information. Contrary to this updating hypothesis, E. J. O'Brien, M. L. Rizzella, J. E.…

  11. 77 FR 28541 - Request for Comments on the Recommendation for the Disclosure of Sequence Listings Using XML...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-05-15

    ... setting forth the controlled vocabularies for use with the sequence part of the standard, and a second... for the disclosure of sequence listings using XML (Proposed ST.26); Annex B.1. Controlled vocabularies...) Feature Keys and Qualifiers. ST.25 uses a controlled vocabulary of feature keys to describe nucleic...

  12. Using Extensible Markup Language (XML) for the Single Source Delivery of Educational Resources by Print and Online: A Case Study

    ERIC Educational Resources Information Center

    Walsh, Lucas

    2007-01-01

    This article seeks to provide an introduction to Extensible Markup Language (XML) by looking at its use in a single source publishing approach to the provision of teaching resources in both hardcopy and online. Using the development of the International Baccalaureate Organisation's online Economics Subject Guide as a practical example, this…

  13. Integrating XQuery-Enabled SCORM XML Metadata Repositories into an RDF-Based E-Learning P2P Network

    ERIC Educational Resources Information Center

    Qu, Changtao; Nejdl, Wolfgang

    2004-01-01

    Edutella is an RDF-based E-Learning P2P network that is aimed to accommodate heterogeneous learning resource metadata repositories in a P2P manner and further facilitate the exchange of metadata between these repositories based on RDF. Whereas Edutella provides RDF metadata repositories with a quite natural integration approach, XML metadata…

  14. Haydock’s recursive solution of self-adjoint problems. Discrete spectrum

    SciTech Connect

    Moroz, Alexander

    2014-12-15

    Haydock’s recursive solution is shown to underline a number of different concepts such as (i) quasi-exactly solvable models, (ii) exactly solvable models, (iii) three-term recurrence solutions based on Schweber’s quantization criterion in Hilbert spaces of entire analytic functions, and (iv) a discrete quantum mechanics of Odake and Sasaki. A recurrent theme of Haydock’s recursive solution is that the spectral properties of any self-adjoint problem can be mapped onto a corresponding sequence of polynomials (p{sub n}(E)) in energy variable E. The polynomials (p{sub n}(E)) are orthonormal with respect to the density of states n{sub 0}(E) and energy eigenstate |E〉 is the generating function of (p{sub n}(E)). The generality of Haydock’s recursive solution enables one to see the different concepts from a unified perspective and mutually benefiting from each other. Some results obtained within the particular framework of any of (i) to (iv) may have much broader significance.

  15. Transitory minimal solutions of hypergeometric recursions and pseudoconvergence of associated continued fractions

    NASA Astrophysics Data System (ADS)

    Deano, Alfredo; Segura, Javier

    2007-06-01

    Three term recurrence relations y_{n+1}+b_n y_n+a_n y_{n-1}D0 can be used for computing recursively a great number of special functions. Depending on the asymptotic nature of the function to be computed, different recursion directions need to be considered: backward for minimal solutions and forward for dominant solutions. However, some solutions interchange their role for finite values of n with respect to their asymptotic behaviour and certain dominant solutions may transitorily behave as minimal. This phenomenon, related to Gautschi's anomalous convergence of the continued fraction for ratios of confluent hypergeometric functions, is shown to be a general situation which takes place for recurrences with a_n negative and b_n changing sign once. We analyze the anomalous convergence of the associated continued fractions for a number of different recurrence relations (modified Bessel functions, confluent and Gauss hypergeometric functions) and discuss the implication of such transitory behaviour on the numerical stability of recursion.

  16. Commutative Recursive Filters for Explicit-Filter Large-Eddy Simulation of Turbulent Flows

    NASA Astrophysics Data System (ADS)

    Kim, Myeongkyun; Yoon, Daegeun; You, Donghyun

    2012-11-01

    One of the most notable drawbacks associated with the implicit-filter LES is that the simulation result is dependent on the numerical grid employed due to the inherent dependence of the filtering operation on the numerical discretization. Alternatively, commutative explicit filters can be applied to distinguish the filtering operation from the underlying mesh distribution, thereby eliminating grid sensitivities. The efficacy of explicit-filtering to obtain grid-independent solutions of turbulent flows has been successfully demonstrated in the previous research (Bose, Moin & You, Phys. Fluids, 2010; Singh, You & Bose, Phys. Fluids, 2012). However, the use of broad-width filters accompanies significant increase in computational cost in terms of memory space and communication load for a distributed memory (MPI-based) parallel computation. To overcome the difficulty, a recursive filtering algorithm which can effectively replace a broad-width commutative filter with a series of narrow-width filters. The efficacy of the commutative recursive filtering method is evaluated in explicit-filter LES of turbulent channel flow, with particular attention to the performance of commutative recursive filters in terms of computational cost and memory requirement for a parallel computation. Supported by the National Research Foundation of Korea and the Brain Korea 21 Program of the Korea Research Foundation.

  17. Recursive Hierarchical Image Segmentation by Region Growing and Constrained Spectral Clustering

    NASA Technical Reports Server (NTRS)

    Tilton, James C.

    2002-01-01

    This paper describes an algorithm for hierarchical image segmentation (referred to as HSEG) and its recursive formulation (referred to as RHSEG). The HSEG algorithm is a hybrid of region growing and constrained spectral clustering that produces a hierarchical set of image segmentations based on detected convergence points. In the main, HSEG employs the hierarchical stepwise optimization (HS WO) approach to region growing, which seeks to produce segmentations that are more optimized than those produced by more classic approaches to region growing. In addition, HSEG optionally interjects between HSWO region growing iterations merges between spatially non-adjacent regions (i.e., spectrally based merging or clustering) constrained by a threshold derived from the previous HSWO region growing iteration. While the addition of constrained spectral clustering improves the segmentation results, especially for larger images, it also significantly increases HSEG's computational requirements. To counteract this, a computationally efficient recursive, divide-and-conquer, implementation of HSEG (RHSEG) has been devised and is described herein. Included in this description is special code that is required to avoid processing artifacts caused by RHSEG s recursive subdivision of the image data. Implementations for single processor and for multiple processor computer systems are described. Results with Landsat TM data are included comparing HSEG with classic region growing. Finally, an application to image information mining and knowledge discovery is discussed.

  18. Evaluation of ISO EN 13606 as a result of its implementation in XML

    PubMed Central

    Sun, Shanghua; Hassan, Taher; Kalra, Dipak

    2013-01-01

    The five parts of the ISO EN 13606 standard define a means by which health-care records can be exchanged between computer systems. Starting within the European standardisation process, it has now become internationally ratified in ISO. However, ISO standards do not require that a reference implementation be provided, and in order for ISO EN 13606 to deliver the expected benefits, it must be provided not as a document, but as an operational system that is not vendor specific. This article describes the evolution of an Extensible Markup Language (XML) Schema through three iterations, each of which emphasised one particular approach to delivering an executable equivalent to the printed standard. Developing these operational versions and incorporating feedback from users of these demonstrated where implementation compromises were needed and exposed defects in the standard. These are discussed herein. They may require a future technical revision to ISO EN 13606 to resolve the issues identified. PMID:23995217

  19. An Infrastructure for Integrated Electronic Health Record Services: The Role of XML (Extensible Markup Language)

    PubMed Central

    Sfakianakis, Stelios; Tsiknakis, Manolis; Orphanoudakis, Stelios C

    2001-01-01

    Background The sharing of information resources is generally accepted as the key to substantial improvements in productivity and better quality of care. In addition, due to the greater mobility of the population, national and international healthcare networks are increasingly used to facilitate the sharing of healthcare-related information among the various actors of the field. In the context of HYGEIAnet, the regional health telematics network of Crete, an Integrated Electronic Health Record environment has been developed to provide integrated access to online clinical information, accessible throughout the island. Objectives To make available comprehensive medical information about a patient by means of incorporating all the distributed and heterogeneous health record segments into an Integrated Electronic Health Record that can be viewed on-line through a unified user interface and visualization environment. Methods The technological approach for implementing this Integrated Electronic Health Record environment is based on the HYGEIAnet Reference Architecture, which provides the necessary framework for the reuse of services, components, and interfaces. Seamless presentation of information is achieved by means of the Extensible Markup Language (XML), while its underlying capabilities allow for dynamic navigation according to personalized end-user preferences and authorities. Results The Integrated Electronic Health Record environment developed in HYGEIAnet provides the basis for consistent and authenticated access to primary information over the Internet in order to support decision-making. Primary information is always kept at the place where it has been produced, and is maintained by the most appropriate clinical information system, contrasting traditional store and forward techniques, or centralized clinical data repositories. Conclusions Since documents are much more easily accessible rather than data inside a database, Extensible Markup Language has the potential of becoming a very cheap technology provided, of course, that the underlying Healthcare Information Infrastructure exists. XML can be introduced incrementally and its implementation is completely transparent to the end user. PMID:11720949

  20. MeMo: a hybrid SQL/XML approach to metabolomic data management for functional genomics

    PubMed Central

    Spasić, Irena; Dunn, Warwick B; Velarde, Giles; Tseng, Andy; Jenkins, Helen; Hardy, Nigel; Oliver, Stephen G; Kell, Douglas B

    2006-01-01

    Background The genome sequencing projects have shown our limited knowledge regarding gene function, e.g. S. cerevisiae has 5–6,000 genes of which nearly 1,000 have an uncertain function. Their gross influence on the behaviour of the cell can be observed using large-scale metabolomic studies. The metabolomic data produced need to be structured and annotated in a machine-usable form to facilitate the exploration of the hidden links between the genes and their functions. Description MeMo is a formal model for representing metabolomic data and the associated metadata. Two predominant platforms (SQL and XML) are used to encode the model. MeMo has been implemented as a relational database using a hybrid approach combining the advantages of the two technologies. It represents a practical solution for handling the sheer volume and complexity of the metabolomic data effectively and efficiently. The MeMo model and the associated software are available at . Conclusion The maturity of relational database technology is used to support efficient data processing. The scalability and self-descriptiveness of XML are used to simplify the relational schema and facilitate the extensibility of the model necessitated by the creation of new experimental techniques. Special consideration is given to data integration issues as part of the systems biology agenda. MeMo has been physically integrated and cross-linked to related metabolomic and genomic databases. Semantic integration with other relevant databases has been supported through ontological annotation. Compatibility with other data formats is supported by automatic conversion. PMID:16753052

  1. HepML, an XML-based format for describing simulated data in high energy physics

    NASA Astrophysics Data System (ADS)

    Belov, S.; Dudko, L.; Kekelidze, D.; Sherstnev, A.

    2010-10-01

    In this paper we describe a HepML format and a corresponding C++ library developed for keeping complete description of parton level events in a unified and flexible form. HepML tags contain enough information to understand what kind of physics the simulated events describe and how the events have been prepared. A HepML block can be included into event files in the LHEF format. The structure of the HepML block is described by means of several XML Schemas. The Schemas define necessary information for the HepML block and how this information should be located within the block. The library libhepml is a C++ library intended for parsing and serialization of HepML tags, and representing the HepML block in computer memory. The library is an API for external software. For example, Matrix Element Monte Carlo event generators can use the library for preparing and writing a header of an LHEF file in the form of HepML tags. In turn, Showering and Hadronization event generators can parse the HepML header and get the information in the form of C++ classes. libhepml can be used in C++, C, and Fortran programs. All necessary parts of HepML have been prepared and we present the project to the HEP community. Program summaryProgram title: libhepml Catalogue identifier: AEGL_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEGL_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: GNU GPLv3 No. of lines in distributed program, including test data, etc.: 138 866 No. of bytes in distributed program, including test data, etc.: 613 122 Distribution format: tar.gz Programming language: C++, C Computer: PCs and workstations Operating system: Scientific Linux CERN 4/5, Ubuntu 9.10 RAM: 1 073 741 824 bytes (1 Gb) Classification: 6.2, 11.1, 11.2 External routines: Xerces XML library ( http://xerces.apache.org/xerces-c/), Expat XML Parser ( http://expat.sourceforge.net/) Nature of problem: Monte Carlo simulation in high energy physics is divided into several stages. Various programs exist for these stages. In this article we are interested in interfacing different Monte Carlo event generators via data files, in particular, Matrix Element (ME) generators and Showering and Hadronization (SH) generators. There is a widely accepted format for data files for such interfaces - Les Houches Event Format (LHEF). Although information kept in an LHEF file is enough for proper working of SH generators, it is insufficient for understanding how events in the LHEF file have been prepared and which physical model has been applied. In this paper we propose an extension of the format for keeping additional information available in generators. We propose to add a new information block, marked up with XML tags, to the LHEF file. This block describes events in the file in more detail. In particular, it stores information about a physical model, kinematical cuts, generator, etc. This helps to make LHEF files self-documented. Certainly, HepML can be applied in more general context, not in LHEF files only. Solution method: In order to overcome drawbacks of the original LHEF accord we propose to add a new information block of HepML tags. HepML is an XML-based markup language. We designed several XML Schemas for all tags in the language. Any HepML document should follow rules of the Schemas. The language is equipped with a library for operation with HepML tags and documents. This C++ library, called libhepml, consists of classes for HepML objects, which represent a HepML document in computer memory, parsing classes, serializating classes, and some auxiliary classes. Restrictions: The software is adapted for solving problems, described in the article. There are no additional restrictions. Running time: Tests have been done on a computer with Intel(R) Core(TM)2 Solo, 1.4 GHz. Parsing of a HepML file: 6 ms (size of the HepML files is 12.5 Kb) Writing of a HepML block to file: 14 ms (file size 12.5 Kb) Merging of two HepML blocks and writing to file: 18 ms (file size - 25.0 Kb).

  2. Development of Updated ABsorption SIMulation Software (ABSIM)

    SciTech Connect

    Yang, Zhiyao; Tang, Xin; Qu, Ming; Abdelaziz, Omar; Gluesenkamp, Kyle R

    2014-01-01

    ABsorption SIMulation, ABSIM, was developed for the simulation of absorption systems by The Oak Ridge National Laboratory during 1980s and 1990s. ABSIM provides a platform for users to investigate various cycle configurations and working fluids, to calculate their operating parameters, to predict their performance, and to compare them with each other on a uniform basis. ABSIM is indeed a very useful and accurate tool for researchers to investigate various absorption systems. However, it has not been well maintained: it is incompatible with recent operating systems; the interface needs improved user-friendliness, and the system needs better parameter setting and debugging tools to help achieve convergence. Therefore, it is highly needed to update and improve ABSIM. The paper presents recent efforts to improve ABSIM s compatibility with current operating systems, user interface, and analysis capabilities. The paper details the features and functions of the newly updated ABSIM software. The new ABSIM still uses the previously validated calculation engine of the old ABSIM. The new graphic user interfaces (GUI) were developed in Qt, which is an open source license GUI software based on C++. XML was used as the database for data storage in the new ABSIM. The new ABSIM has been designed to be easily learned and used. It has enhanced editing and construction functions, plus enhanced analysis features including parametric tables, plotting, property plots, and master panels for debugging. A single effect water/LiBr absorption system is used as a case study in this paper to illustrate the features, capabilities, and functions of the new ABSIM. This case study was actually an example system available in the old ABSIM. The new version of ABSIM will be continuously developed to include additional subroutines for the components in liquid desiccant systems. The new ABSIM will be available to public for free. The ultimate goal of the new ABSIM is to allow it to become a simulation platform for both absorption heat pump and liquid desiccant systems.

  3. Providing access to risk prediction tools via the HL7 XML-formatted risk web service.

    PubMed

    Chipman, Jonathan; Drohan, Brian; Blackford, Amanda; Parmigiani, Giovanni; Hughes, Kevin; Bosinoff, Phil

    2013-07-01

    Cancer risk prediction tools provide valuable information to clinicians but remain computationally challenging. Many clinics find that CaGene or HughesRiskApps fit their needs for easy- and ready-to-use software to obtain cancer risks; however, these resources may not fit all clinics' needs. The HughesRiskApps Group and BayesMendel Lab therefore developed a web service, called "Risk Service", which may be integrated into any client software to quickly obtain standardized and up-to-date risk predictions for BayesMendel tools (BRCAPRO, MMRpro, PancPRO, and MelaPRO), the Tyrer-Cuzick IBIS Breast Cancer Risk Evaluation Tool, and the Colorectal Cancer Risk Assessment Tool. Software clients that can convert their local structured data into the HL7 XML-formatted family and clinical patient history (Pedigree model) may integrate with the Risk Service. The Risk Service uses Apache Tomcat and Apache Axis2 technologies to provide an all Java web service. The software client sends HL7 XML information containing anonymized family and clinical history to a Dana-Farber Cancer Institute (DFCI) server, where it is parsed, interpreted, and processed by multiple risk tools. The Risk Service then formats the results into an HL7 style message and returns the risk predictions to the originating software client. Upon consent, users may allow DFCI to maintain the data for future research. The Risk Service implementation is exemplified through HughesRiskApps. The Risk Service broadens the availability of valuable, up-to-date cancer risk tools and allows clinics and researchers to integrate risk prediction tools into their own software interface designed for their needs. Each software package can collect risk data using its own interface, and display the results using its own interface, while using a central, up-to-date risk calculator. This allows users to choose from multiple interfaces while always getting the latest risk calculations. Consenting users contribute their data for future research, thus building a rich multicenter resource. PMID:23793601

  4. Recursive Focal Plane Wavefront and Bias Estimation for the Direct Imaging of Exoplanets

    NASA Astrophysics Data System (ADS)

    Eldorado Riggs, A. J.; Kasdin, N. Jeremy; Groff, Tyler Dean

    2016-01-01

    To image the reflected light from exoplanets and disks, an instrument must suppress diffracted starlight by about nine orders of magnitude. A coronagraph alters the stellar PSF to create regions of high contrast, but it is extremely sensitive to wavefront aberrations. Deformable mirrors (DMs) are necessary to mitigate these quasi-static aberrations and recover high-contrast. To avoid non-common path aberrations, the science camera must be used as the primary wavefront sensor. Focal plane wavefront correction is an iterative process, and obtaining sufficient signal in the dark holes requires long exposure times. The fastest coronagraphic wavefront correction techniques require estimates of the stellar electric field. The main challenge of coronagraphy is thus to perform complex wavefront estimation quickly and efficiently using intensity images from the camera. The most widely applicable and tested technique is DM Diversity, in which a DM modulates the focal plane intensity and several images are used to reconstruct the stellar electric field in a batch process. At the High Contrast Imaging Lab (HCIL) at Princeton, we have developed an iterative extended Kalman filter (IEKF) to improve upon this technique. The IEKF enables recursive starlight estimation and can utilize fewer images per iteration, thereby speeding up wavefront correction. This IEKF formulation also estimates the bias in the images recursively. Since exoplanets and disks are embedded in the incoherent bias signal, the IEKF enables detection of science targets during wavefront correction. Here we present simulated and experimental results from Princeton's HCIL demonstrating the effectiveness of the IEKF for recursive electric field estimation and exoplanet detection.

  5. Cyclic period-3 window in antiferromagnetic potts and Ising models on recursive lattices

    NASA Astrophysics Data System (ADS)

    Ananikian, N. S.; Ananikyan, L. N.; Chakhmakhchyan, L. A.

    2011-09-01

    The magnetic properties of the antiferromagnetic Potts model with two-site interaction and the antiferromagnetic Ising model with three-site interaction on recursive lattices have been studied. A cyclic period-3 window has been revealed by the recurrence relation method in the antiferromagnetic Q-state Potts model on the Bethe lattice (at Q < 2) and in the antiferromagnetic Ising model with three-site interaction on the Husimi cactus. The Lyapunov exponents have been calculated, modulated phases and a chaotic regime in the cyclic period-3 window have been found for one-dimensional rational mappings determined the properties of these systems.

  6. Recursive identification and tracking of parameters for linear and non-linear multivariable systems. [aeronautical applications

    NASA Technical Reports Server (NTRS)

    Sidar, M.

    1976-01-01

    The problem of identifying constant and variable parameters in multi-input, multi-output, linear and nonlinear systems is considered, using the maximum likelihood approach. An iterative algorithm, leading to recursive identification and tracking of the unknown parameters and the noise covariance matrix, is developed. Agile tracking and accurate and unbiased identified parameters are obtained. Necessary conditions for a globally asymptotically stable identification process are provided; the conditions proved to be useful and efficient. Among different cases studied, the stability derivatives of an aircraft were identified and some of the results are shown as examples.

  7. Recursive polarization of nuclear spins in diamond at arbitrary magnetic fields

    SciTech Connect

    Pagliero, Daniela; Laraoui, Abdelghani; Henshaw, Jacob D.; Meriles, Carlos A.

    2014-12-15

    We introduce an alternate route to dynamically polarize the nuclear spin host of nitrogen-vacancy (NV) centers in diamond. Our approach articulates optical, microwave, and radio-frequency pulses to recursively transfer spin polarization from the NV electronic spin. Using two complementary variants of the same underlying principle, we demonstrate nitrogen nuclear spin initialization approaching 80% at room temperature both in ensemble and single NV centers. Unlike existing schemes, our approach does not rely on level anti-crossings and is thus applicable at arbitrary magnetic fields. This versatility should prove useful in applications ranging from nanoscale metrology to sensitivity-enhanced NMR.

  8. A convolutional recursive modified Self Organizing Map for handwritten digits recognition.

    PubMed

    Mohebi, Ehsan; Bagirov, Adil

    2014-12-01

    It is well known that the handwritten digits recognition is a challenging problem. Different classification algorithms have been applied to solve it. Among them, the Self Organizing Maps (SOM) produced promising results. In this paper, first we introduce a Modified SOM for the vector quantization problem with improved initialization process and topology preservation. Then we develop a Convolutional Recursive Modified SOM and apply it to the problem of handwritten digits recognition. The computational results obtained using the well known MNIST dataset demonstrate the superiority of the proposed algorithm over the existing SOM-based algorithms. PMID:25190490

  9. Chandrasekhar-type algorithms for fast recursive estimation in linear systems with constant parameters

    NASA Technical Reports Server (NTRS)

    Choudhury, A. K.; Djalali, M.

    1975-01-01

    In this recursive method proposed, the gain matrix for the Kalman filter and the convariance of the state vector are computed not via the Riccati equation, but from certain other equations. These differential equations are of Chandrasekhar-type. The 'invariant imbedding' idea resulted in the reduction of the basic boundary value problem of transport theory to an equivalent initial value system, a significant computational advance. Initial value experience showed that there is some computational savings in the method and the loss of positive definiteness of the covariance matrix is less vulnerable.

  10. A recursive solution for a fading memory filter derived from Kalman filter theory

    NASA Technical Reports Server (NTRS)

    Statman, J. I.

    1986-01-01

    A simple recursive solution for a class of fading memory tracking filters is presented. A fading memory filter provides estimates of filter states based on past measurements, similar to a traditional Kalman filter. Unlike a Kalman filter, an exponentially decaying weight is applied to older measurements, discounting their effect on present state estimates. It is shown that Kalman filters and fading memory filters are closely related solutions to a general least squares estimator problem. Closed form filter transfer functions are derived for a time invariant, steady state, fading memory filter. These can be applied in loop filter implementation of the Deep Space Network (DSN) Advanced Receiver carrier phase locked loop (PLL).

  11. Fast Multilevel Implementation of Recursive Spectral Bisection for Partitioning Unstructured Problems

    NASA Technical Reports Server (NTRS)

    Barnard, Stephen T.; Simon, Horst D.

    1993-01-01

    If problems involving unstructured meshes are to be solved efficiently on distributed-memory parallel computers, the meshes must be partitioned and distributed across processors in a way that balances tile computational load and minimizes communication. The recursive spectral bisection method (RSB) has been shown to be very effective for such partitioning problems compared to alternative methods, but RSB in its simplest form is expensive. Here a multilevel version of RSB is introduced that attains about an order-of-magnitude improvement in run time on typical examples.

  12. Generalized Recursion Relations for Correlators in the Gauge-Gravity Correspondence

    SciTech Connect

    Raju, Suvrat

    2011-03-04

    We show that a generalization of the Britto-Cachazo-Feng-Witten recursion relations gives a new and efficient method of computing correlation functions of the stress tensor or conserved currents in conformal field theories with an (d+1)-dimensional anti-de Sitter space dual, for d{>=}4, in the limit where the bulk theory is approximated by tree-level Yang-Mills theory or gravity. In supersymmetric theories, additional correlators of operators that live in the same multiplet as a conserved current or stress tensor can be computed by these means.

  13. Generalized recursion relations for correlators in the gauge-gravity correspondence.

    PubMed

    Raju, Suvrat

    2011-03-01

    We show that a generalization of the Britto-Cachazo-Feng-Witten recursion relations gives a new and efficient method of computing correlation functions of the stress tensor or conserved currents in conformal field theories with an (d+1)-dimensional anti-de Sitter space dual, for d?4, in the limit where the bulk theory is approximated by tree-level Yang-Mills theory or gravity. In supersymmetric theories, additional correlators of operators that live in the same multiplet as a conserved current or stress tensor can be computed by these means. PMID:21405614

  14. Recursive identification and tracking of parameters for linear and nonlinear multivariable systems

    NASA Technical Reports Server (NTRS)

    Sidar, M.

    1975-01-01

    The problem of identifying constant and variable parameters in multi-input, multi-output, linear and nonlinear systems is considered, using the maximum likelihood approach. An iterative algorithm, leading to recursive identification and tracking of the unknown parameters and the noise covariance matrix, is developed. Agile tracking, and accurate and unbiased identified parameters are obtained. Necessary conditions for a globally, asymptotically stable identification process are provided; the conditions proved to be useful and efficient. Among different cases studied, the stability derivatives of an aircraft were identified and some of the results are shown as examples.

  15. Recursive encoding and decoding of the noiseless subsystem and decoherence-free subspace

    SciTech Connect

    Li, Chi-Kwong; Nakahara, Mikio; Poon, Yiu-Tung; Sze, Nung-Sing; Tomita, Hiroyuki

    2011-10-15

    When an environmental disturbance to a quantum system has a wavelength much larger than the system size, all qubits in the system are under the action of the same error operator. The noiseless subsystem and decoherence-free subspace are immune to such collective noise. We construct simple quantum circuits that implement these error-avoiding codes for a small number n of physical qubits. A single logical qubit is encoded with n=3 and 4, while two and three logical qubits are encoded with n=5 and 7, respectively. Recursive relations among subspaces employed in these codes play essential roles in our implementation.

  16. System Simulation by Recursive Feedback: Coupling A Set of Stand-Alone Subsystem Simulations

    NASA Technical Reports Server (NTRS)

    Nixon, Douglas D.; Hanson, John M. (Technical Monitor)

    2002-01-01

    Recursive feedback is defined and discussed as a framework for development of specific algorithms and procedures that propagate the time-domain solution for a dynamical system simulation consisting of multiple numerically coupled self-contained stand-alone subsystem simulations. A satellite motion example containing three subsystems (other dynamics, attitude dynamics, and aerodynamics) has been defined and constructed using this approach. Conventional solution methods are used in the subsystem simulations. Centralized and distributed versions of coupling structure have been addressed. Numerical results are evaluated by direct comparison with a standard total-system simultaneous-solution approach.

  17. System Simulation by Recursive Feedback: Coupling a Set of Stand-Alone Subsystem Simulations

    NASA Technical Reports Server (NTRS)

    Nixon, Douglas D.; Ryan, Stephen G. (Technical Monitor)

    2002-01-01

    Recursive feedback is defined and discussed as a framework for development of specific algorithms and procedures that propagate the time-domain solution for a dynamical system simulation consisting of multiple numerically coupled, self-contained, stand-alone subsystem simulations. A satellite motion example containing three subsystems (orbit dynamics, attitude dynamics, and aerodynamics) has been defined and constructed using this approach. Conventional solution methods are used in the subsystem simulations. Centralized and distributed versions of coupling structure have been addressed. Numerical results are evaluated by direct comparison with a standard total-system, simultaneous-solution approach.

  18. Non-sequential Recursive Pair Substitutions and Numerical Entropy Estimates in Symbolic Dynamical Systems

    NASA Astrophysics Data System (ADS)

    Calcagnile, Lucio M.; Galatolo, Stefano; Menconi, Giulia

    2010-12-01

    We numerically test the method of non-sequential recursive pair substitutions to estimate the entropy of an ergodic source. We compare its performance with other classical methods to estimate the entropy (empirical frequencies, return times, and Lyapunov exponent). We have considered as a benchmark for the methods several systems with different statistical properties: renewal processes, dynamical systems provided and not provided with a Markov partition, and slow or fast decay of correlations. Most experiments are supported by rigorous mathematical results, which are explained in the paper.

  19. ART-ML - a novel XML format for the biological procedures modeling and the representation of blood flow simulation.

    PubMed

    Karvounis, E C; Tsakanikas, V D; Fotiou, E; Fotiadis, D I

    2010-01-01

    The paper proposes a novel Extensible Markup Language (XML) based format called ART-ML that aims at supporting the interoperability and the reuse of models of blood flow, mass transport and plaque formation, exported by ARTool. ARTool is a platform for the automatic processing of various image modalities of coronary and carotid arteries. The images and their content are fused to develop morphological models of the arteries in easy to handle 3D representations. The platform incorporates efficient algorithms which are able to perform blood flow simulation. In addition atherosclerotic plaque development is estimated taking into account morphological, flow and genetic factors. ART-ML provides a XML format that enables the representation and management of embedded models within the ARTool platform and the storage and interchange of well-defined information. This approach influences in the model creation, model exchange, model reuse and result evaluation. PMID:21096364

  20. Recursive Averaging

    ERIC Educational Resources Information Center

    Smith, Scott G.

    2015-01-01

    In this article, Scott Smith presents an innocent problem (Problem 12 of the May 2001 Calendar from "Mathematics Teacher" ("MT" May 2001, vol. 94, no. 5, p. 384) that was transformed by several timely "what if?" questions into a rewarding investigation of some interesting mathematics. These investigations led to two…

  1. An Efficient Bottom-up Filtering of XML Messages by Exploiting the Postfix Commonality of XPath Queries

    NASA Astrophysics Data System (ADS)

    Kim, Jaehoon; Kim, Youngsoo; Park, Seog

    Recently, for more efficient filtering of XML data, YFilter system has been suggested to exploit the prefix commonalities that exist among path expressions. Sharing the prefix commonality gives the benefit of improving filtering performance through the tremendous reduction in filtering machine size. However, exploiting the postfix commonality can also be useful for an XML filtering situation. For example, when a stream of XML messages does not have any defined schema, or users cannot remember the defined schema exactly, users often use the partial matching path queries which begins with the descendant axis (“//”), e. g., ‘//science/article/title’, ‘//entertainment/article/title’, and ‘//title’. If so, the registered XPath queries are most likely to have the postfix commonality, e. g., the sample queries share the partial path expressions ‘article/title’ and ‘title’. Therefore, in this paper, we introduce a bottom-up filtering approach exploiting the postfix commonality against the top-down approach of YFilter exploiting the prefix commonality. Some experimental results show that our method has better filtering performance when registered XPath queries mainly consist of the partial matching path queries with the postfix commonality.

  2. Automatic Indexing for Content Analysis of Whale Recordings and XML Representation

    NASA Astrophysics Data System (ADS)

    Bénard, Frédéric; Glotin, Hervé

    2010-12-01

    This paper focuses on the robust indexing of sperm whale hydrophone recordings based on a set of features extracted from a real-time passive underwater acoustic tracking algorithm for multiple whales using four hydrophones. Acoustic localization permits the study of whale behavior in deep water without interfering with the environment. Given the position coordinates, we are able to generate different features such as the speed, energy of the clicks, Inter-Click-Interval (ICI), and so on. These features allow to construct different markers which allow us to index and structure the audio files. Thus, the behavior study is facilitated by choosing and accessing the corresponding index in the audio file. The complete indexing algorithm is processed on real data from the NUWC (Naval Undersea Warfare Center of the US Navy) and the AUTEC (Atlantic Undersea Test & Evaluation Center-Bahamas). Our model is validated by similar results from the US Navy (NUWC) and SOEST (School of Ocean and Earth Science and Technology) Hawaii university labs in a single whale case. Finally, as an illustration, we index a single whale sound file using the extracted whale's features provided by the tracking, and we present an example of an XML script structuring it.

  3. Adding Hierarchical Objects to Relational Database General-Purpose XML-Based Information Managements

    NASA Technical Reports Server (NTRS)

    Lin, Shu-Chun; Knight, Chris; La, Tracy; Maluf, David; Bell, David; Tran, Khai Peter; Gawdiak, Yuri

    2006-01-01

    NETMARK is a flexible, high-throughput software system for managing, storing, and rapid searching of unstructured and semi-structured documents. NETMARK transforms such documents from their original highly complex, constantly changing, heterogeneous data formats into well-structured, common data formats in using Hypertext Markup Language (HTML) and/or Extensible Markup Language (XML). The software implements an object-relational database system that combines the best practices of the relational model utilizing Structured Query Language (SQL) with those of the object-oriented, semantic database model for creating complex data. In particular, NETMARK takes advantage of the Oracle 8i object-relational database model using physical-address data types for very efficient keyword searches of records across both context and content. NETMARK also supports multiple international standards such as WEBDAV for drag-and-drop file management and SOAP for integrated information management using Web services. The document-organization and -searching capabilities afforded by NETMARK are likely to make this software attractive for use in disciplines as diverse as science, auditing, and law enforcement.

  4. XML-based clinical data standardisation in the National Health Service Scotland.

    PubMed

    Bunduchi, Raluca; Williams, Robin; Graham, Ian; Smart, Alison

    2006-01-01

    The objective of this paper is to clarify the role that socio-economic factors played in shaping the development of XML-based clinical data standards in the National Health Service in Scotland from 2000 to 2004. The paper discusses the NHS Scotland approach to clinical data standardisation, emphasising the actors involved, their choices during the standard development process and the factors that have shaped these choices. The case suggests that the NHS Scotland approach to clinical data standardisation is shaped by strong political pressures for fast development of an integrated electronic patient care system, economic pressures for high efficiency and cost reductions, and organisational requirements for strong clinical support. Such economic, political and organisational pressures explain the informal approach to standard development, the emphasis on fast system development and strong clinical involvement. At the same time, market factors explain the low commitment of the IT vendors, which might have otherwise put significant pressure on NHS Scotland to pursue a more formalised standardisation approach within an internationally recognised standard-setting body. PMID:17504574

  5. Country Update: Israel 2005

    ERIC Educational Resources Information Center

    Marar, Marianne Maurice

    2005-01-01

    Country Updates is a new section of "Intercultural Education." Starting in "Intercultural Education," Volume 16 No. 5, this column will focus on recent developments during the last two to three years in the field of intercultural education in one particular country. These updates can include recent policy decisions, the main results of important…

  6. AQMEII Status Update

    EPA Science Inventory

    AQMEII Status UpdateThis presentation provided an overview and status update of the Air Quality Model Evaluation International Initative (AQMEII) to participants of a workshop of the Task Force on Hemispheric Transport of Air Pollution (TF-HTAP) . In addition, the p...

  7. The template update problem.

    PubMed

    Matthews, Iain; Ishikawa, Takahiro; Baker, Simon

    2004-06-01

    Template tracking dates back to the 1981 Lucas-Kanade algorithm. One question that has received very little attention, however, is how to update the template so that it remains a good model of the tracked object. We propose a template update algorithm that avoids the "drifting" inherent in the naive algorithm. PMID:18579941

  8. A recursive field-normalized bibliometric performance indicator: an application to the field of library and information science.

    PubMed

    Waltman, Ludo; Yan, Erjia; van Eck, Nees Jan

    2011-10-01

    Two commonly used ideas in the development of citation-based research performance indicators are the idea of normalizing citation counts based on a field classification scheme and the idea of recursive citation weighing (like in PageRank-inspired indicators). We combine these two ideas in a single indicator, referred to as the recursive mean normalized citation score indicator, and we study the validity of this indicator. Our empirical analysis shows that the proposed indicator is highly sensitive to the field classification scheme that is used. The indicator also has a strong tendency to reinforce biases caused by the classification scheme. Based on these observations, we advise against the use of indicators in which the idea of normalization based on a field classification scheme and the idea of recursive citation weighing are combined. PMID:21957321

  9. Development of Fast Algorithms Using Recursion, Nesting and Iterations for Computational Electromagnetics

    NASA Technical Reports Server (NTRS)

    Chew, W. C.; Song, J. M.; Lu, C. C.; Weedon, W. H.

    1995-01-01

    In the first phase of our work, we have concentrated on laying the foundation to develop fast algorithms, including the use of recursive structure like the recursive aggregate interaction matrix algorithm (RAIMA), the nested equivalence principle algorithm (NEPAL), the ray-propagation fast multipole algorithm (RPFMA), and the multi-level fast multipole algorithm (MLFMA). We have also investigated the use of curvilinear patches to build a basic method of moments code where these acceleration techniques can be used later. In the second phase, which is mainly reported on here, we have concentrated on implementing three-dimensional NEPAL on a massively parallel machine, the Connection Machine CM-5, and have been able to obtain some 3D scattering results. In order to understand the parallelization of codes on the Connection Machine, we have also studied the parallelization of 3D finite-difference time-domain (FDTD) code with PML material absorbing boundary condition (ABC). We found that simple algorithms like the FDTD with material ABC can be parallelized very well allowing us to solve within a minute a problem of over a million nodes. In addition, we have studied the use of the fast multipole method and the ray-propagation fast multipole algorithm to expedite matrix-vector multiplication in a conjugate-gradient solution to integral equations of scattering. We find that these methods are faster than LU decomposition for one incident angle, but are slower than LU decomposition when many incident angles are needed as in the monostatic RCS calculations.

  10. A small-world network derived from the deterministic uniform recursive tree by line graph operation

    NASA Astrophysics Data System (ADS)

    Hou, Pengfeng; Zhao, Haixing; Mao, Yaping; Wang, Zhao

    2016-03-01

    The deterministic uniform recursive tree ({DURT}) is one of the deterministic versions of the uniform recursive tree ({URT}). Zhang et al (2008 Eur. Phys. J. B 63 507-13) studied the properties of DURT, including its topological characteristics and spectral properties. Although DURT shows a logarithmic scaling with the size of the network, DURT is not a small-world network since its clustering coefficient is zero. Lu et al (2012 Physica A 391 87-92) proposed a deterministic small-world network by adding some edges with a simple rule in each DURT iteration. In this paper, we intoduce a method for constructing a new deterministic small-world network by the line graph operation in each DURT iteration. The line graph operation brings about cliques at each node of the previous given graph, and the resulting line graph possesses larger clustering coefficients. On the other hand, this operation can decrease the diameter at almost one, then giving the analytic solutions to several topological characteristics of the model proposed. Supported by The Ministry of Science and Technology 973 project (No. 2010C B334708); National Science Foundation of China (Nos. 61164005, 11161037, 11101232, 11461054, 11551001); The Ministry of education scholars and innovation team support plan of Yangtze River (No. IRT1068); Qinghai Province Nature Science Foundation Project (Nos. 2012-Z-943, 2014-ZJ-907).

  11. Recursive Random Lasso (RRLasso) for Identifying Anti-Cancer Drug Targets

    PubMed Central

    Park, Heewon; Imoto, Seiya; Miyano, Satoru

    2015-01-01

    Uncovering driver genes is crucial for understanding heterogeneity in cancer. L1-type regularization approaches have been widely used for uncovering cancer driver genes based on genome-scale data. Although the existing methods have been widely applied in the field of bioinformatics, they possess several drawbacks: subset size limitations, erroneous estimation results, multicollinearity, and heavy time consumption. We introduce a novel statistical strategy, called a Recursive Random Lasso (RRLasso), for high dimensional genomic data analysis and investigation of driver genes. For time-effective analysis, we consider a recursive bootstrap procedure in line with the random lasso. Furthermore, we introduce a parametric statistical test for driver gene selection based on bootstrap regression modeling results. The proposed RRLasso is not only rapid but performs well for high dimensional genomic data analysis. Monte Carlo simulations and analysis of the “Sanger Genomics of Drug Sensitivity in Cancer dataset from the Cancer Genome Project” show that the proposed RRLasso is an effective tool for high dimensional genomic data analysis. The proposed methods provide reliable and biologically relevant results for cancer driver gene selection. PMID:26544691

  12. A generalized recursive convolution method for time-domain propagation in porous media.

    PubMed

    Dragna, Didier; Pineau, Pierre; Blanc-Benon, Philippe

    2015-08-01

    An efficient numerical method, referred to as the auxiliary differential equation (ADE) method, is proposed to compute convolutions between relaxation functions and acoustic variables arising in sound propagation equations in porous media. For this purpose, the relaxation functions are approximated in the frequency domain by rational functions. The time variation of the convolution is thus governed by first-order differential equations which can be straightforwardly solved. The accuracy of the method is first investigated and compared to that of recursive convolution methods. It is shown that, while recursive convolution methods are first or second-order accurate in time, the ADE method does not introduce any additional error. The ADE method is then applied for outdoor sound propagation using the equations proposed by Wilson et al. in the ground [(2007). Appl. Acoust. 68, 173-200]. A first one-dimensional case is performed showing that only five poles are necessary to accurately approximate the relaxation functions for typical applications. Finally, the ADE method is used to compute sound propagation in a three-dimensional geometry over an absorbing ground. Results obtained with Wilson's equations are compared to those obtained with Zwikker and Kosten's equations and with an impedance surface for different flow resistivities. PMID:26328719

  13. Real-Time Adaptive EEG Source Separation Using Online Recursive Independent Component Analysis.

    PubMed

    Hsu, Sheng-Hsiou; Mullen, Tim R; Jung, Tzyy-Ping; Cauwenberghs, Gert

    2016-03-01

    Independent component analysis (ICA) has been widely applied to electroencephalographic (EEG) biosignal processing and brain-computer interfaces. The practical use of ICA, however, is limited by its computational complexity, data requirements for convergence, and assumption of data stationarity, especially for high-density data. Here we study and validate an optimized online recursive ICA algorithm (ORICA) with online recursive least squares (RLS) whitening for blind source separation of high-density EEG data, which offers instantaneous incremental convergence upon presentation of new data. Empirical results of this study demonstrate the algorithm's: 1) suitability for accurate and efficient source identification in high-density (64-channel) realistically-simulated EEG data; 2) capability to detect and adapt to nonstationarity in 64-ch simulated EEG data; and 3) utility for rapidly extracting principal brain and artifact sources in real 61-channel EEG data recorded by a dry and wearable EEG system in a cognitive experiment. ORICA was implemented as functions in BCILAB and EEGLAB and was integrated in an open-source Real-time EEG Source-mapping Toolbox (REST), supporting applications in ICA-based online artifact rejection, feature extraction for real-time biosignal monitoring in clinical environments, and adaptable classifications in brain-computer interfaces. PMID:26685257

  14. Block recursive LU preconditioners for the thermally coupled incompressible inductionless MHD problem

    NASA Astrophysics Data System (ADS)

    Badia, Santiago; Martín, Alberto F.; Planas, Ramon

    2014-10-01

    The thermally coupled incompressible inductionless magnetohydrodynamics (MHD) problem models the flow of an electrically charged fluid under the influence of an external electromagnetic field with thermal coupling. This system of partial differential equations is strongly coupled and highly nonlinear for real cases of interest. Therefore, fully implicit time integration schemes are very desirable in order to capture the different physical scales of the problem at hand. However, solving the multiphysics linear systems of equations resulting from such algorithms is a very challenging task which requires efficient and scalable preconditioners. In this work, a new family of recursive block LU preconditioners is designed and tested for solving the thermally coupled inductionless MHD equations. These preconditioners are obtained after splitting the fully coupled matrix into one-physics problems for every variable (velocity, pressure, current density, electric potential and temperature) that can be optimally solved, e.g., using preconditioned domain decomposition algorithms. The main idea is to arrange the original matrix into an (arbitrary) 2×2 block matrix, and consider an LU preconditioner obtained by approximating the corresponding Schur complement. For every one of the diagonal blocks in the LU preconditioner, if it involves more than one type of unknowns, we proceed the same way in a recursive fashion. This approach is stated in an abstract way, and can be straightforwardly applied to other multiphysics problems. Further, we precisely explain a flexible and general software design for the code implementation of this type of preconditioners.

  15. Recursive forward dynamics for multiple robot arms moving a common task object

    NASA Technical Reports Server (NTRS)

    Rodriguez, G.

    1988-01-01

    Recursive forward dynamics algorithms are developed for an arbitrary number of robot arms moving a commonly held object. The multiarm forward dynamics problem is to find the angular accelerations at the joints and the contact forces that the arms impart to the task object. The problem also involves finding the acceleration of this object. The multiarm forward dynamics solutions provide a thorough physical and mathematical understanding of the way several arms behave in response to a set of applied joint moments. Such an understanding simplifies and guides the subsequent control design and experimentation process. The forward dynamics algorithms also provide the necessary analytical foundation for conducting analysis and simulation studies. The multiarm algorithms are based on the filtering and smoothing approach recently advanced for single-arm dynamics, and they can be built up modularly from the single-arm algorithms. The algorithms compute recursively the joint-angle accelerations, the contact forces, and the task-object accelerations. Algorithms are also developed to evaluate in closed form the linear transformations from the active joint moments to the joint-angle accelerations, to the task-object accelerations., and to the task-object contact forces. A possible computing architecture is presented as a precursor to a more complete investigation of the computational performance of the dynamics algorithms.

  16. Recursive Sparse Point Process Regression With Application to Spectrotemporal ReceptiveField Plasticity Analysis

    NASA Astrophysics Data System (ADS)

    Sheikhattar, Alireza; Fritz, Jonathan B.; Shamma, Shihab A.; Babadi, Behtash

    2016-04-01

    We consider the problem of estimating the sparse time-varying parameter vectors of a point process model in an online fashion, where the observations and inputs respectively consist of binary and continuous time series. We construct a novel objective function by incorporating a forgetting factor mechanism into the point process log-likelihood to enforce adaptivity and employ $\\ell_1$-regularization to capture the sparsity. We provide a rigorous analysis of the maximizers of the objective function, which extends the guarantees of compressed sensing to our setting. We construct two recursive filters for online estimation of the parameter vectors based on proximal optimization techniques, as well as a novel filter for recursive computation of statistical confidence regions. Simulation studies reveal that our algorithms outperform several existing point process filters in terms of trackability, goodness-of-fit and mean square error. We finally apply our filtering algorithms to experimentally recorded spiking data from the ferret primary auditory cortex during attentive behavior in a click rate discrimination task. Our analysis provides new insights into the time-course of the spectrotemporal receptive field plasticity of the auditory neurons.

  17. Recursive Organizer (ROR): an Analytic Framework for Sequence-Based Association Analysis

    PubMed Central

    Zhao, Lue Ping; Huang, Xin

    2013-01-01

    The advent of next-generation sequencing technologies affords the ability to sequence thousands of subjects cost-effectively, and is revolutionizing the landscape of genetic research. With the evolving genotyping/sequencing technologies, it is not unrealistic to expect that we will soon obtain a pair of diploidic fully-phased genome sequences from each subject in the near future. Here, in light of this potential, we propose an analytic framework called, recursive organizer (ROR), which recursively groups sequence variants based upon sequence similarities and their empirical disease associations, into fewer and potentially more interpretable super sequence variants (SSV). As an illustration, we applied ROR to assess an association between HLA-DRB1 and type 1 diabetes (T1D), discovering SSVs of HLA-DRB1 with sequence data from the Wellcome Trust Case Control Consortium (WTCCC). Specifically, ROR reduces 36 observed unique HLA-DRB1 sequences into 8 SSVs that empirically associate with T1D, a four-fold reduction of sequence complexity. Using HLA-DRB1 data from Type 1 Diabetes Genetics Consortium (T1DGC) as cases and data from Fred Hutchinson Cancer Research Center as controls, we are able to validate associations of these SSVs with T1D. Further, SSVs consist of nine nucleotides, and each associates with its corresponding amino acids. Detailed examination of these selected amino acids reveals their potential functional roles in protein structures and possible implication to the mechanism of T1D. PMID:23494241

  18. Recursive subspace identification for on-line tracking of structural modal parameter

    NASA Astrophysics Data System (ADS)

    Weng, Jian-Huang; Loh, Chin-Hsiung

    2011-11-01

    The objective of this paper is to develop an on-line tracking of system parameter estimation and damage detection techniques using response measurements. To avoid the singular-value-decomposition in data Hankel matrix, a new subspace identification algorithm was developed. Seismic response data of a 3-story steel frame with abrupt change of inter-story stiffness from the shaking table test was used to verify the proposed recursive subspace identification (RSI) method by using both input and output measurements. With the implementation of forgetting factor in RSI method the ability of on-line damage detection of the abrupt change of structural stiffness can be enhanced. Then, the recursive stochastic subspace identification (RSSI) algorithm is also developed for continuous structural health monitor of structure by using the output-only measurements. Verification of the proposed RSSI method by using the white noise response data of a 2-story reinforced concrete frame from its low level white noise excitation was used. Discussion of the subspace identification model parameters is also investigated.

  19. Exploring Biological Relationships Between Calving Traits in Primiparous Cattle with a Bayesian Recursive Model

    PubMed Central

    de Maturana, Evangelina López; Wu, Xiao-Lin; Gianola, Daniel; Weigel, Kent A.; Rosa, Guilherme J. M.

    2009-01-01

    Structural equation models (SEMs) of a recursive type with heterogeneous structural coefficients were used to explore biological relationships between gestation length (GL), calving difficulty (CD), and perinatal mortality, also known as stillbirth (SB), in cattle, with the last two traits having categorical expression. An acyclic model was assumed, where recursive effects existed from the GL phenotype to the liabilities (latent variables) to CD and SB and from the liability to CD to that of SB considering four periods regarding GL. The data contained GL, CD, and SB records from 90,393 primiparous cows, sired by 1122 bulls, distributed over 935 herd-calving year classes. Low genetic correlations between GL and the other calving traits were found, whereas the liabilities to CD and SB were high and positively correlated, genetically. The model indicated that gestations of ∼274 days of length (3 days shorter than the average) would lead to the lowest CD and SB and confirmed the existence of an intermediate optimum of GL with respect to these traits. PMID:18984571

  20. SU-E-T-48: Automated Quality Assurance for XML Controlled Linacs

    SciTech Connect

    Valdes, G; Morin, O; Pouliot, J; Chuang, C

    2014-06-01

    Purpose: To automate routine imaging QA procedures so that complying with TG 142 and TG 179 can be efficient and reliable. Methods: Two QA tests for a True Beam Linac were automatized. A Winston Lutz test as described by Lutz et al{sup 1} using the Winston Lutz test kit from BrainLab, Germany and a CBCT Image Quality test as described in TG 179 using the EMMA phantom, Siemens Medical Physics, Germany were performed in our True Beam. For each QA procedure tested, a 3 step paradigm was used. First, the data was automatically acquired using True Beam Developer Mode and XML scripting. Second, the data acquired in the first step was automatically processed using in-home grown Matlab GUIs. Third, Machine Learning algorithms were used to automatically classify the processed data and reports generated. Results: The Winston Luzt test could be performed by an experienced medical physicist in 29.0 ± 8.0 min. The same test, if automated using our paradigm, could be performed in 3.0 ± 0.1 min. In the same lieu, time could be substantially saved for image quality tests. In this case, the amount of time saved will depend on the phantoms used and the initial localization method. Additionally, machine learning algorithms could automatically identify the roots of the problems if any and possibly help reduce machine down time. Conclusion: Modern linear accelerators are equipped with advanced 2D and 3D imaging that are used for patient alignment substantially improving IGRT protocols. However, this extra complexity exponentially increases the number of QA tests needed. Using the new paradigm described above, not only bare minimum but best practice QA programs could be implemented with the same manpower. This work is supported by Varian, Palo Alto, CA.

  1. XML Web Services and 3D Orbit Viewer Application of SSCWeb

    NASA Astrophysics Data System (ADS)

    Candey, R. M.; Chimiak, R. A.; Harris, B. T.; Kovalick, T. J.; McGuire, R. E.

    2002-12-01

    The Satellite Situation Center Web (SSCWeb) is a browser-based service to provide geocentric spacecraft location information and cast it into a framework of (empirical) geophysical regions and mappings of spacecraft locations along lines of the Earth's magnetic field. While backed by a substantial and maintained database of spacecraft orbit information and extensive software logic, several shortcomings in the present service are the inability of the architecture to support externally-developed clients and limitation of the orbit graphics to static, 2-D plots. This talk introduces a new distributed programming interface to the SSCWeb software based on the SOAP (Simple Object Access Protocol), a modern, industry standard technology. This interface, SSC XML Web Services, allows systems to communicate with SSC over the open protocols of the Internet. This flexible architecture will enable new SPDF developed client applications, as well as externally developed clients, to access the SSCWeb data and logic to bring new services and capabilities to the SEC community. The first such client application is TIPSOD (Tool for Interactive Plotting, Sonification and 3-D Orbit Display). Implemented in Java 3D, TIPSOD extends the existing SSCWeb 2-D static orbit graphics with 3-D interactive and animated displays linking set of spacecraft positions as a function of time. Additional capability and functional enhancements to SSCWeb services and TIPSOD, as well as the extension of this technology to the CDAWeb service, are being considered to further the relevance and usefulness of this work to the science community. SSCWeb is a joint effort of the NASA GSFC Space Physics Data Facility (SPDF) and the National Space Science Data Center (NSSDC).

  2. The recursive combination filter approach of pre-processing for the estimation of standard deviation of RR series.

    PubMed

    Mishra, Alok; Swati, D

    2015-09-01

    Variation in the interval between the R-R peaks of the electrocardiogram represents the modulation of the cardiac oscillations by the autonomic nervous system. This variation is contaminated by anomalous signals called ectopic beats, artefacts or noise which mask the true behaviour of heart rate variability. In this paper, we have proposed a combination filter of recursive impulse rejection filter and recursive 20% filter, with recursive application and preference of replacement over removal of abnormal beats to improve the pre-processing of the inter-beat intervals. We have tested this novel recursive combinational method with median method replacement to estimate the standard deviation of normal to normal (SDNN) beat intervals of congestive heart failure (CHF) and normal sinus rhythm subjects. This work discusses the improvement in pre-processing over single use of impulse rejection filter and removal of abnormal beats for heart rate variability for the estimation of SDNN and Poncar plot descriptors (SD1, SD2, and SD1/SD2) in detail. We have found the 22 ms value of SDNN and 36 ms value of SD2 descriptor of Poincar plot as clinical indicators in discriminating the normal cases from CHF cases. The pre-processing is also useful in calculation of Lyapunov exponent which is a nonlinear index as Lyapunov exponents calculated after proposed pre-processing modified in a way that it start following the notion of less complex behaviour of diseased states. PMID:26104469

  3. Using Recursive Regression to Explore Nonlinear Relationships and Interactions: A Tutorial Applied to a Multicultural Education Study

    ERIC Educational Resources Information Center

    Strang, Kenneth David

    2009-01-01

    This paper discusses how a seldom-used statistical procedure, recursive regression (RR), can numerically and graphically illustrate data-driven nonlinear relationships and interaction of variables. This routine falls into the family of exploratory techniques, yet a few interesting features make it a valuable compliment to factor analysis and…

  4. Question Utilization in Solution-Focused Brief Therapy: A Recursive Frame Analysis of Insoo Kim Berg's Solution Talk

    ERIC Educational Resources Information Center

    Cotton, Jeffrey

    2010-01-01

    Recursive frame analysis (RFA) was used to conduct a single case investigation of Insoo Kim Berg's question utilization talk in a solution-focused brief therapy (SFBT) session. Due to the lack of process research that explores how SFBT questions facilitate change, the author investigated how Berg's solution language influenced a client to respond…

  5. An Introduction to Recursive Partitioning: Rationale, Application, and Characteristics of Classification and Regression Trees, Bagging, and Random Forests

    ERIC Educational Resources Information Center

    Strobl, Carolin; Malley, James; Tutz, Gerhard

    2009-01-01

    Recursive partitioning methods have become popular and widely used tools for nonparametric regression and classification in many scientific fields. Especially random forests, which can deal with large numbers of predictor variables even in the presence of complex interactions, have been applied successfully in genetics, clinical medicine, and…

  6. Technical Note: Analytical sensitivity analysis of a two parameter recursive digital baseflow separation filter

    NASA Astrophysics Data System (ADS)

    Eckhardt, K.

    2011-10-01

    A sensitivity analysis for a well established baseflow separation technique, a two parameter recursive digital filter, is presented. The propagation of errors or uncertainties of the two filter parameters into the calculated baseflow index is analytically ascertained. Representative sensitivity indices (defined as the ratio between the relative error of the baseflow index and the relative error of the respective parameter) are derived by application of the resulting equations to a great number of catchments. It is found that in the mean the parameter a, the recession constant, has a stronger influence on the calculated baseflow index than the second filter parameter BFImax. This is favourable in that a can be determined by a recession analysis and therefore should be less uncertain. Whether this finding also applies for a specific catchment can easily be checked by means of the derived equations.

  7. Recursive multiport schemes for implementing quantum algorithms with photonic integrated circuits

    NASA Astrophysics Data System (ADS)

    Tabia, Gelo Noel M.

    2016-01-01

    We present recursive multiport schemes for implementing quantum Fourier transforms and the inversion step in Grover's algorithm on an integrated linear optics device. In particular, each scheme shows how to execute a quantum operation on 2 d modes using a pair of circuits for the same operation on d modes. The circuits operate on path-encoded qudits and realize d -dimensional unitary transformations on these states using linear optical networks with O (d2) optical elements. To evaluate the schemes against realistic errors, we ran simulations of proof-of-principle experiments using a simple fabrication model of silicon-based photonic integrated devices that employ directional couplers and thermo-optic modulators for beam splitters and phase shifters, respectively. We find that high-fidelity performance is achievable with our multiport circuits for 2-qubit and 3-qubit quantum Fourier transforms, and for quantum search on four-item and eight-item databases.

  8. Recursive algorithms for bias and gain nonuniformity correction in infrared videos.

    PubMed

    Pipa, Daniel R; da Silva, Eduardo A B; Pagliari, Carla L; Diniz, Paulo S R

    2012-12-01

    Infrared focal-plane array (IRFPA) detectors suffer from fixed-pattern noise (FPN) that degrades image quality, which is also known as spatial nonuniformity. FPN is still a serious problem, despite recent advances in IRFPA technology. This paper proposes new scene-based correction algorithms for continuous compensation of bias and gain nonuniformity in FPA sensors. The proposed schemes use recursive least-square and affine projection techniques that jointly compensate for both the bias and gain of each image pixel, presenting rapid convergence and robustness to noise. The synthetic and real IRFPA videos experimentally show that the proposed solutions are competitive with the state-of-the-art in FPN reduction, by presenting recovered images with higher fidelity. PMID:22997263

  9. A Recursive Multiscale Correlation-Averaging Algorithm for an Automated Distributed Road Condition Monitoring System

    SciTech Connect

    Ndoye, Mandoye; Barker, Alan M; Krogmeier, James; Bullock, Darcy

    2011-01-01

    A signal processing approach is proposed to jointly filter and fuse spatially indexed measurements captured from many vehicles. It is assumed that these measurements are influenced by both sensor noise and measurement indexing uncertainties. Measurements from low-cost vehicle-mounted sensors (e.g., accelerometers and Global Positioning System (GPS) receivers) are properly combined to produce higher quality road roughness data for cost-effective road surface condition monitoring. The proposed algorithms are recursively implemented and thus require only moderate computational power and memory space. These algorithms are important for future road management systems, which will use on-road vehicles as a distributed network of sensing probes gathering spatially indexed measurements for condition monitoring, in addition to other applications, such as environmental sensing and/or traffic monitoring. Our method and the related signal processing algorithms have been successfully tested using field data.

  10. Management of Large-Scale Wireless Sensor Networks Utilizing Multi-Parent Recursive Area Hierarchies

    SciTech Connect

    Cree, Johnathan V.; Delgado-Frias, Jose

    2013-04-19

    Autonomously configuring and self-healing a largescale wireless sensor network requires a light-weight maintenance protocol that is scalable. Further, in a battery powered wireless sensor network duty-cycling a node’s radio can reduce the power consumption of a device and extend the lifetime of a network. With duty-cycled nodes the power consumption of a node’s radio depends on the amount of communication is must perform and by reducing the communication the power consumption can also be reduced. Multi-parent hierarchies can be used to reduce the communication cost when constructing a recursive area clustering hierarchy when compared to singleparent solutions that utilize inefficient communication methods such as flooding and information propagation via single-hop broadcasts. The multi-parent hierarchies remain scalable and provides a level of redundancy for the hierarchy.

  11. Human motion planning based on recursive dynamics and optimal control techniques.

    PubMed

    Lo, Janzen; Huang, Gang; Metaxas, Dimitris

    2002-11-01

    This paper presents an efficient optimal control and recursive dynamics-based computer animation system for simulating and controlling the motion of articulated figures. A quasi-Newton nonlinear programming technique (super-linear convergence) is implemented to solve minimum torque-based human motion-planning problems. The explicit analytical gradients needed in the dynamics are derived using a matrix exponential formulation and Lie algebra. Cubic spline functions are used to make the search space for an optimal solution finite. Based on our formulations, our method is well conditioned and robust, in addition to being computationally efficient. To better illustrate the efficiency of our method, we present results of natural looking and physically correct human motions for a variety of human motion tasks involving open and closed loop kinematic chains. PMID:12583402

  12. Spatial join optimization among WFSs based on recursive partitioning and filtering rate estimation

    NASA Astrophysics Data System (ADS)

    Lan, Guiwen; Wu, Congcong; Shi, Guangyi; Chen, Qi; Yang, Zhao

    2015-12-01

    Spatial join among Web Feature Services (WFS) is time-consuming for most of non-candidate spatial objects may be encoded by GML and transferred to client side. In this paper, an optimization strategy is proposed to enhance performance of these joins by filtering non-candidate spatial objects as many as possible. By recursive partitioning, the data skew of sub-areas is facilitated to reduce data transmission using spatial semi-join. Moreover filtering rate is used to determine whether a spatial semi-join for a sub-area is profitable and choose a suitable execution plan for it. The experimental results show that the proposed strategy is feasible under most circumstances.

  13. Performance and Accuracy of Recursive Subspace Bisection for Hybrid DFT Calculations

    NASA Astrophysics Data System (ADS)

    Dawson, William; Gygi, Francois

    2015-03-01

    The high cost of computing the Hartree-Fock exchange has resulted in limited use of Hybrid Functionals in DFT calculations. Approximations based on transformation to localized orbitals provide one way to reduce this cost. One such method is the recursive subspace bisection method (RSB). Such localization methods involve truncation of localized orbitals, which introduces an additional approximation. We take advantage of our ability to systematically reduce the error in RSB calculations through a single parameter to study this approximation. We present the errors in ground state energy, forces, and relative energy differences between configurations for a variety of systems, including tungsten oxide, a silicon-water interface, and liquid water including the calculation of empty states. Supported by DOE/BES Grant DE-SC0008938.

  14. Towards Interactive Construction of Topical Hierarchy: A Recursive Tensor Decomposition Approach

    PubMed Central

    Wang, Chi; Liu, Xueqing; Song, Yanglei; Han, Jiawei

    2015-01-01

    Automatic construction of user-desired topical hierarchies over large volumes of text data is a highly desirable but challenging task. This study proposes to give users freedom to construct topical hierarchies via interactive operations such as expanding a branch and merging several branches. Existing hierarchical topic modeling techniques are inadequate for this purpose because (1) they cannot consistently preserve the topics when the hierarchy structure is modified; and (2) the slow inference prevents swift response to user requests. In this study, we propose a novel method, called STROD, that allows efficient and consistent modification of topic hierarchies, based on a recursive generative model and a scalable tensor decomposition inference algorithm with theoretical performance guarantee. Empirical evaluation shows that STROD reduces the runtime of construction by several orders of magnitude, while generating consistent and quality hierarchies. PMID:26705505

  15. Recursive method to obtain the parametric representation of a generic Feynman diagram

    SciTech Connect

    Gonzalez, Ivan; Schmidt, Ivan

    2005-11-15

    A recursive algebraic method which allows one to obtain the Feynman or Schwinger parametric representation of a generic L-loops and (E+1) external lines diagram, in a scalar {phi}{sup 3}+{phi}{sup 4} theory, is presented. The representation is obtained starting from an initial parameters matrix, which relates the scalar products between internal and external momenta, and which appears directly when this parametrization is applied to the momentum space representation of the graph. The final product is an algebraic formula that shows explicitly the external momenta dependence and also an algorithm that can be easily programmed, either in a computer programming language (C/C++, Fortran, ...) or in a symbolic calculation package (Maple, Mathematica, ...)

  16. On the Shock-Response-Spectrum Recursive Algorithm of Kelly and Richman

    NASA Technical Reports Server (NTRS)

    Martin, Justin N.; Sinclair, Andrew J.; Foster, Winfred A.

    2010-01-01

    The monograph Principles and Techniques of Shock Data Analysis written by Kelly and Richman in 1969 has become a seminal reference on the shock response spectrum (SRS) [1]. Because of its clear physical descriptions and mathematical presentation of the SRS, it has been cited in multiple handbooks on the subject [2, 3] and research articles [4 10]. Because of continued interest, two additional versions of the monograph have been published: a second edition by Scavuzzo and Pusey in 1996 [11] and a reprint of the original edition in 2008 [12]. The main purpose of this note is to correct several typographical errors in the manuscript's presentation of a recursive algorithm for SRS calculations. These errors are consistent across all three editions of the monograph. The secondary purpose of this note is to present a Matlab implementation of the corrected algorithm.

  17. Cloud Computing Application for Hotspot Clustering Using Recursive Density Based Clustering (RDBC)

    NASA Astrophysics Data System (ADS)

    Santoso, Aries; Khiyarin Nisa, Karlina

    2016-01-01

    Indonesia has vast areas of tropical forest, but are often burned which causes extensive damage to property and human life. Monitoring hotspots can be one of the forest fire management. Each hotspot is recorded in dataset so that it can be processed and analyzed. This research aims to build a cloud computing application which visualizes hotspots clustering. This application uses the R programming language with Shiny web framework and implements Recursive Density Based Clustering (RDBC) algorithm. Clustering is done on hotspot dataset of the Kalimantan Island and South Sumatra Province to find the spread pattern of hotspots. The clustering results are evaluated using the Silhouette's Coefficient (SC) which yield best value 0.3220798 for Kalimantan dataset. Clustering pattern are displayed in the form of web pages so that it can be widely accessed and become the reference for fire occurrence prediction.

  18. Parallel FEM algorithms based on recursive spatial decomposition. II - Automatic analysis via hierarchical substructuring

    NASA Astrophysics Data System (ADS)

    Saxena, M.; Perucchio, R.

    1993-04-01

    In the companion paper, we discussed the applicability of a recursive spatial decomposition (RSD) based automatic meshing procedure for concurrent implementation. This paper describes an automatic substructuring scheme based on RSD of the domain, that can be closely integrated with the RSD-based automatic meshing procedure. The hierarchical data structures used to represent RSD-based automatically derived meshes also provide a one-to-one mapping between spatially decomposed subdomains and analytical substructures. Such a hierarchical organization of the substructures and the inherent parallelism of RSD are exploited to design the substructuring scheme suitable for parallel processing. A systematic simulation of parallel computational environments is used to evaluate the performance of the substructuring scheme for coarse-grain architectures. Speed-ups from vectorization alone are also reported.

  19. Adaptive speed control of hydrogenerators by recursive least squares identification algorithm

    SciTech Connect

    Bonnett, S.C.; Wozniak, L.

    1995-03-01

    An inherent challenge to hydrogenerator speed regulation is the nonlinear and time varying nature of the process. When a fixed gain controller is used to regulate such a process, globally stable system response can be achieved only at the expense of degraded off design-point performance. This paper considers the application of an adaptive control methodology that allows the controller gains to automatically adjust to changing process variables and thereby provide uniform closed loop response over a wide range of operating conditions. Preliminary findings derived from a calculated plant model, and a standard recursive identification technique, indicate that the adaptive system provides desirable dynamic response despite changes in system operating dynamics, maintaining stable operation in situations where constant gain schedules fail.

  20. Co-clustering with Recursive Elimination for Verb Synonym Extraction from Large Text Corpus

    NASA Astrophysics Data System (ADS)

    Takeuchi, Koichi; Takahashi, Hideyuki

    The extraction of verb synonyms is a key technology to build a verb dictionary as a language resource. This paper presents a co-clustering-based verb synonym extraction approach that increases the number of extracted meanings of polysemous verbs from a large text corpus. For verb synonym extraction with a clustering approach dealing with polysemous verbs can be one problem issue because each polysemous verb should be categorized into different clusters depending on each meaning; thus there is a high possibility of failing to extract some of the meanings of polysemous verbs. Our proposed approach can extract the different meanings of polysemous verbs by recursively eliminating the extracted clusters from the initial data set. The experimental results of verb synonym extraction show that the proposed approach increases the correct verb clusters by about 50% with a 0.9% increase in precision and a 1.5% increase in recall over the previous approach.

  1. Two dimensional recursive digital filters for near real time image processing

    NASA Technical Reports Server (NTRS)

    Olson, D.; Sherrod, E.

    1980-01-01

    A program was designed toward the demonstration of the feasibility of using two dimensional recursive digital filters for subjective image processing applications that require rapid turn around. The concept of the use of a dedicated minicomputer for the processor for this application was demonstrated. The minicomputer used was the HP1000 series E with a RTE 2 disc operating system and 32K words of memory. A Grinnel 256 x 512 x 8 bit display system was used to display the images. Sample images were provided by NASA Goddard on a 800 BPI, 9 track tape. Four 512 x 512 images representing 4 spectral regions of the same scene were provided. These images were filtered with enhancement filters developed during this effort.

  2. Human motion planning based on recursive dynamics and optimal control techniques

    NASA Technical Reports Server (NTRS)

    Lo, Janzen; Huang, Gang; Metaxas, Dimitris

    2002-01-01

    This paper presents an efficient optimal control and recursive dynamics-based computer animation system for simulating and controlling the motion of articulated figures. A quasi-Newton nonlinear programming technique (super-linear convergence) is implemented to solve minimum torque-based human motion-planning problems. The explicit analytical gradients needed in the dynamics are derived using a matrix exponential formulation and Lie algebra. Cubic spline functions are used to make the search space for an optimal solution finite. Based on our formulations, our method is well conditioned and robust, in addition to being computationally efficient. To better illustrate the efficiency of our method, we present results of natural looking and physically correct human motions for a variety of human motion tasks involving open and closed loop kinematic chains.

  3. Recursive anisotropic 2-D Gaussian filtering based on a triple-axis decomposition.

    PubMed

    Lam, Stanley Yiu Man; Shi, Bertram E

    2007-07-01

    We describe a recursive algorithm for anisotropic 2-D Gaussian filtering, based on separating the filter into the cascade of three, rather two, 1-D filters. The filters operate along axes obtained by integer horizontal and/or vertical pixel shifts. This eliminates interpolation, which removes spatial inhomogeneity in the filter, and produces more elliptically shaped kernels. It also results in a more regular filter structure, which facilitates implementation in DSP chips. Finally, it improves matching between filters with the same eccentricity and width, but different orientations. Our analysis and experiments indicate that the computational complexity is similar to an algorithm that operates along two axes (<11 ms for a 512 x 512 image using a 3.2-GHz Pentium 4 PC). On the other hand, given a limited set of basis filter axes, there is an orientation dependent lower bound on the achievable aspect ratios. PMID:17605390

  4. System Simulation by Recursive Feedback: Coupling a Set of Stand-Alone Subsystem Simulations

    NASA Technical Reports Server (NTRS)

    Nixon, D. D.

    2001-01-01

    Conventional construction of digital dynamic system simulations often involves collecting differential equations that model each subsystem, arran g them to a standard form, and obtaining their numerical gin solution as a single coupled, total-system simultaneous set. Simulation by numerical coupling of independent stand-alone subsimulations is a fundamentally different approach that is attractive because, among other things, the architecture naturally facilitates high fidelity, broad scope, and discipline independence. Recursive feedback is defined and discussed as a candidate approach to multidiscipline dynamic system simulation by numerical coupling of self-contained, single-discipline subsystem simulations. A satellite motion example containing three subsystems (orbit dynamics, attitude dynamics, and aerodynamics) has been defined and constructed using this approach. Conventional solution methods are used in the subsystem simulations. Distributed and centralized implementations of coupling have been considered. Numerical results are evaluated by direct comparison with a standard total-system, simultaneous-solution approach.

  5. Social cognition in ADHD: irony understanding and recursive theory of mind.

    PubMed

    Caillies, Stéphanie; Bertot, Vincine; Motte, Jacques; Raynaud, Christine; Abely, Michel

    2014-11-01

    The main goal of the present study was to characterise the social cognition abilities of French children with ADHD, in terms of their understanding of people's recursive mental states and their irony comprehension. We hypothesised that these children have difficulty understanding second-order false beliefs and ironic remarks, owing to the executive dysfunction that is characteristic of ADHD. We therefore conducted an experiment in which children with ADHD and typically developing matched controls performed second-order false-belief and executive function tasks. They then listened to ironic stories and answered questions about the ironic comments and about the speakers' beliefs and attitudes. The groups differed significantly on second-order theory of mind, irony comprehension and executive functions, confirming that children with ADHD have impaired social cognition. PMID:25155741

  6. A Comparison of Evolutionary Algorithms for Tracking Time-Varying Recursive Systems

    NASA Astrophysics Data System (ADS)

    White, Michael S.; Flockton, Stuart J.

    2003-12-01

    A comparison is made of the behaviour of some evolutionary algorithms in time-varying adaptive recursive filter systems. Simulations show that an algorithm including random immigrants outperforms a more conventional algorithm using the breeder genetic algorithm as the mutation operator when the time variation is discontinuous, but neither algorithm performs well when the time variation is rapid but smooth. To meet this deficit, a new hybrid algorithm which uses a hill climber as an additional genetic operator, applied for several steps at each generation, is introduced. A comparison is made of the effect of applying the hill climbing operator a few times to all members of the population or a larger number of times solely to the best individual; it is found that applying to the whole population yields the better results, substantially improved compared with those obtained using earlier methods.

  7. Fault detection in an air-handling unit using residual and recursive parameter identification methods

    SciTech Connect

    Lee, W.Y.; Park, C.; Kelly, G.E.

    1996-11-01

    A scheme for detecting faults in an air-handling unit using residual and parameter identification methods is presented. Faults can be detected by comparing the normal or expected operating condition data with the abnormal, measured data using residuals. Faults can also be detected by examining unmeasurable parameter changes in a model of a controlled system using a system parameter identification technique. In this study, autoregressive moving average with exogenous input (ARMAX) and autoregressive with exogenous input (ARX) models with both single-input/single-output (SISO) and multi-input/single-output (MISO) structures are examined. Model parameters are determined using the Kalman filter recursive identification method. This approach is tested using experimental data from a laboratory`s variable-air-volume (VAV) air-handling unit operated with and without faults.

  8. RecRWR: a recursive random walk method for improved identification of diseases.

    PubMed

    Arrais, Joel Perdiz; Oliveira, José Luís

    2015-01-01

    High-throughput methods such as next-generation sequencing or DNA microarrays lack precision, as they return hundreds of genes for a single disease profile. Several computational methods applied to physical interaction of protein networks have been successfully used in identification of the best disease candidates for each expression profile. An open problem for these methods is the ability to combine and take advantage of the wealth of biomedical data publicly available. We propose an enhanced method to improve selection of the best disease targets for a multilayer biomedical network that integrates PPI data annotated with stable knowledge from OMIM diseases and GO biological processes. We present a comprehensive validation that demonstrates the advantage of the proposed approach, Recursive Random Walk with Restarts (RecRWR). The obtained results outline the superiority of the proposed approach, RecRWR, in identifying disease candidates, especially with high levels of biological noise and benefiting from all data available. PMID:25874227

  9. RecRWR: A Recursive Random Walk Method for Improved Identification of Diseases

    PubMed Central

    Perdiz Arrais, Joel; Oliveira, José Luís

    2015-01-01

    High-throughput methods such as next-generation sequencing or DNA microarrays lack precision, as they return hundreds of genes for a single disease profile. Several computational methods applied to physical interaction of protein networks have been successfully used in identification of the best disease candidates for each expression profile. An open problem for these methods is the ability to combine and take advantage of the wealth of biomedical data publicly available. We propose an enhanced method to improve selection of the best disease targets for a multilayer biomedical network that integrates PPI data annotated with stable knowledge from OMIM diseases and GO biological processes. We present a comprehensive validation that demonstrates the advantage of the proposed approach, Recursive Random Walk with Restarts (RecRWR). The obtained results outline the superiority of the proposed approach, RecRWR, in identifying disease candidates, especially with high levels of biological noise and benefiting from all data available. PMID:25874227

  10. Parallel Implementation of the Recursive Approximation of an Unsupervised Hierarchical Segmentation Algorithm. Chapter 5

    NASA Technical Reports Server (NTRS)

    Tilton, James C.; Plaza, Antonio J. (Editor); Chang, Chein-I. (Editor)

    2008-01-01

    The hierarchical image segmentation algorithm (referred to as HSEG) is a hybrid of hierarchical step-wise optimization (HSWO) and constrained spectral clustering that produces a hierarchical set of image segmentations. HSWO is an iterative approach to region grooving segmentation in which the optimal image segmentation is found at N(sub R) regions, given a segmentation at N(sub R+1) regions. HSEG's addition of constrained spectral clustering makes it a computationally intensive algorithm, for all but, the smallest of images. To counteract this, a computationally efficient recursive approximation of HSEG (called RHSEG) has been devised. Further improvements in processing speed are obtained through a parallel implementation of RHSEG. This chapter describes this parallel implementation and demonstrates its computational efficiency on a Landsat Thematic Mapper test scene.

  11. Use of recursively generated intermediates in state selective multireference coupled-cluster method: A numerical example

    SciTech Connect

    Ghose, K.B.; Adamowicz, L.

    1995-12-01

    The present work represents the first attempt to utilize the idea of recursively generated intermediates (RGI) in the framework of the state-selective multi-reference coupled-cluster method truncated at triple excitations [SS CCSD(T)]. The expressions for stepwise generation of intermediates are so structured that the spin and point symmetry simplifications can be easily applied during computation. Suitable modifications in SS CCSD(T) equations are introduced to allow for optional quasilinearization of nonlinear terms in difficult convergence situations. The computational code is, as expected, much faster than the SS CCSD(T) code without RGI adaptation. This has been numerically demonstrated by potential energy surface (PES) calculation of the HF molecule using a double zeta basis. {copyright} {ital 1995} {ital American} {ital Institute} {ital of} {ital Physics}.

  12. nD statistical shape model building via recursive boundary subdivision

    NASA Astrophysics Data System (ADS)

    Rueda, Sylvia; Udupa, Jayaram K.

    2009-02-01

    Landmark based statistical object modeling techniques, such as Active Shape Modeling, have proven useful in medical image analysis. Identification of the same homologous set of points in a training set of object shapes is the most crucial step in ASM, which has encountered challenges, the most crucial among these being (C1) defining and characterizing landmarks; (C2) ensuring homology; (C3) generalizing to n > 2 dimensions; (C4) achieving practical computations. In this paper, we propose a novel global-to-local strategy that attempts to address C3 and C4 directly and works in Rn. The 3D version of it attempts to address C1 and C2 indirectly by starting from three initial corresponding points determined in all training shapes via a method α, and subsequently by subdividing the shapes into connected boundary segments by a plane determined by these points. A shape analysis method β is applied on each segment to determine a landmark on the segment. This point introduces more triplets of points, the planes defined by which are used to further subdivide the boundary segments. This recursive boundary subdivision (RBS) process continues simultaneously on all training shapes, maintaining synchrony of the level of recursion, and thereby keeping correspondence among generated points automatically by the correspondence of the homologous shape segments in all training shapes. The process terminates when no subdividing planes are left to be considered that indicate (as per method β) that a point can continue to be selected on the associated segment. Several examples of α and β are provided as well as some preliminary results on 3D shapes.

  13. Poster — Thur Eve — 55: An automated XML technique for isocentre verification on the Varian TrueBeam

    SciTech Connect

    Asiev, Krum; Mullins, Joel; DeBlois, François; Liang, Liheng; Syme, Alasdair

    2014-08-15

    Isocentre verification tests, such as the Winston-Lutz (WL) test, have gained popularity in the recent years as techniques such as stereotactic radiosurgery/radiotherapy (SRS/SRT) treatments are more commonly performed on radiotherapy linacs. These highly conformal treatments require frequent monitoring of the geometrical accuracy of the isocentre to ensure proper radiation delivery. At our clinic, the WL test is performed by acquiring with the EPID a collection of 8 images of a WL phantom fixed on the couch for various couch/gantry angles. This set of images is later analyzed to determine the isocentre size. The current work addresses the acquisition process. A manual WL test acquisition performed by and experienced physicist takes in average 25 minutes and is prone to user manipulation errors. We have automated this acquisition on a Varian TrueBeam STx linac (Varian, Palo Alto, USA). The Varian developer mode allows the execution of custom-made XML script files to control all aspects of the linac operation. We have created an XML-WL script that cycles through each couch/gantry combinations taking an EPID image at each position. This automated acquisition is done in less than 4 minutes. The reproducibility of the method was verified by repeating the execution of the XML file 5 times. The analysis of the images showed variation of the isocenter size less than 0.1 mm along the X, Y and Z axes and compares favorably to a manual acquisition for which we typically observe variations up to 0.5 mm.

  14. Recursive mass matrix factorization and inversion: An operator approach to open- and closed-chain multibody dynamics

    NASA Technical Reports Server (NTRS)

    Rodriguez, G.; Kreutz, K.

    1988-01-01

    This report advances a linear operator approach for analyzing the dynamics of systems of joint-connected rigid bodies.It is established that the mass matrix M for such a system can be factored as M=(I+H phi L)D(I+H phi L) sup T. This yields an immediate inversion M sup -1=(I-H psi L) sup T D sup -1 (I-H psi L), where H and phi are given by known link geometric parameters, and L, psi and D are obtained recursively by a spatial discrete-step Kalman filter and by the corresponding Riccati equation associated with this filter. The factors (I+H phi L) and (I-H psi L) are lower triangular matrices which are inverses of each other, and D is a diagonal matrix. This factorization and inversion of the mass matrix leads to recursive algortihms for forward dynamics based on spatially recursive filtering and smoothing. The primary motivation for advancing the operator approach is to provide a better means to formulate, analyze and understand spatial recursions in multibody dynamics. This is achieved because the linear operator notation allows manipulation of the equations of motion using a very high-level analytical framework (a spatial operator algebra) that is easy to understand and use. Detailed lower-level recursive algorithms can readily be obtained for inspection from the expressions involving spatial operators. The report consists of two main sections. In Part 1, the problem of serial chain manipulators is analyzed and solved. Extensions to a closed-chain system formed by multiple manipulators moving a common task object are contained in Part 2. To retain ease of exposition in the report, only these two types of multibody systems are considered. However, the same methods can be easily applied to arbitrary multibody systems formed by a collection of joint-connected regid bodies.

  15. DbMap: improving database interoperability issues in medical software using a simple, Java-Xml based solution.

    PubMed Central

    Karadimas, H.; Hemery, F.; Roland, P.; Lepage, E.

    2000-01-01

    In medical software development, the use of databases plays a central role. However, most of the databases have heterogeneous encoding and data models. To deal with these variations in the application code directly is error-prone and reduces the potential reuse of the produced software. Several approaches to overcome these limitations have been proposed in the medical database literature, which will be presented. We present a simple solution, based on a Java library, and a central Metadata description file in XML. This development approach presents several benefits in software design and development cycles, the main one being the simplicity in maintenance. PMID:11079915

  16. The Open Microscopy Environment (OME) Data Model and XML file: open tools for informatics and quantitative analysis in biological imaging

    PubMed Central

    Goldberg, Ilya G; Allan, Chris; Burel, Jean-Marie; Creager, Doug; Falconi, Andrea; Hochheiser, Harry; Johnston, Josiah; Mellen, Jeff; Sorger, Peter K; Swedlow, Jason R

    2005-01-01

    The Open Microscopy Environment (OME) defines a data model and a software implementation to serve as an informatics framework for imaging in biological microscopy experiments, including representation of acquisition parameters, annotations and image analysis results. OME is designed to support high-content cell-based screening as well as traditional image analysis applications. The OME Data Model, expressed in Extensible Markup Language (XML) and realized in a traditional database, is both extensible and self-describing, allowing it to meet emerging imaging and analysis needs. PMID:15892875

  17. Using the eXtensible Markup Language (XML) in a regional electronic patient record for patients with malignant diseases.

    PubMed

    Wolff, A C; Mludek, V; van der Haak, M; Bork, W; Bülzebruck, H; Drings, P; Schmücker, P; Wannenmacher, M; Haux, R

    2001-01-01

    Communication between different institutions which are responsible for the treatment of the same patient is of outstanding significance, especially in the field of tumor diseases. Regional electronic patient records could support the co-operation of different institutions by providing ac-cess to all necessary information whether it belongs to the own institution or to a partner. The Department of Medical Informatics, University of Heidelberg is performing a project in co-operation with the Thoraxclinic-Heidelberg and the Department of Clinical Radiology, University of Heidelberg with the goal: to define an architectural concept for interlinking the electronic patient record of the two clinical institutions to build a common virtual electronic patient record and carry out an exemplary implementation, to examine composition, structure and content of medical documents for tumor patients with the aim of defining an XML-based markup language allowing summarizing overviews and suitable granularities, and to integrate clinical practice guidelines and other external knowledge with the electronic patient record using XML-technologies to support the physician in the daily decision process. This paper will show, how a regional electronic patient record could be built on an architectural level and describe elementary steps towards a on content-oriented structuring of medical records. PMID:11604828

  18. An enhanced security solution for electronic medical records based on AES hybrid technique with SOAP/XML and SHA-1.

    PubMed

    Kiah, M L Mat; Nabi, Mohamed S; Zaidan, B B; Zaidan, A A

    2013-10-01

    This study aims to provide security solutions for implementing electronic medical records (EMRs). E-Health organizations could utilize the proposed method and implement recommended solutions in medical/health systems. Majority of the required security features of EMRs were noted. The methods used were tested against each of these security features. In implementing the system, the combination that satisfied all of the security features of EMRs was selected. Secure implementation and management of EMRs facilitate the safeguarding of the confidentiality, integrity, and availability of e-health organization systems. Health practitioners, patients, and visitors can use the information system facilities safely and with confidence anytime and anywhere. After critically reviewing security and data transmission methods, a new hybrid method was proposed to be implemented on EMR systems. This method will enhance the robustness, security, and integration of EMR systems. The hybrid of simple object access protocol/extensible markup language (XML) with advanced encryption standard and secure hash algorithm version 1 has achieved the security requirements of an EMR system with the capability of integrating with other systems through the design of XML messages. PMID:24037086

  19. [Emergency medicine: update 2008].

    PubMed

    Carron, Pierre-Nicolas; Rutschmann, Olivier

    2009-01-01

    The emergency medicine appears more and more as a transversal discipline, leaning on specific competences regularly updated with evidence-based medicine concepts. This selection of recent articles presents an update on frequent conditions, including the place of neuroimaging for patients with seizures or minor head injuries, the management of acute cocaine intoxications, the diagnosis of aortic dissections, or the management of cardiopulmonary arrest. The primary care physician will find elements of diagnostic or therapeutic strategies. This selection reflects the dynamism of emergency medicine. PMID:19216323

  20. Automatic representation and reconstruction of DBM from LiDAR data using Recursive Minimum Bounding Rectangle

    NASA Astrophysics Data System (ADS)

    Kwak, Eunju; Habib, Ayman

    2014-07-01

    Three-dimensional building models are important for various applications, such as disaster management and urban planning. The development of laser scanning sensor technologies has resulted in many different approaches for efficient building model generation using LiDAR data. Despite this effort, generation of these models lacks economical and reliable techniques that fully exploit the advantage of LiDAR data. Therefore, this research aims to develop a framework for fully-automated building model generation by integrating data-driven and model-driven methods using LiDAR datasets. The building model generation starts by employing LiDAR data for building detection and approximate boundary determination. The generated building boundaries are then integrated into a model-based processing strategy because LiDAR derived planes show irregular boundaries due to the nature of LiDAR point acquisition. The focus of the research is generating models for the buildings with right-angled-corners, which can be described with a collection of rectangles under the assumption that the majority of the buildings in urban areas belong to this category. Therefore, by applying the Minimum Bounding Rectangle (MBR) algorithm recursively, the LiDAR boundaries are decomposed into sets of rectangles for further processing. At the same time, the quality of the MBRs is examined to verify that the buildings, from which the boundaries are generated, are buildings with right-angled-corners. The parameters that define the model primitives are adjusted through a model-based boundary fitting procedure using LiDAR boundaries. The level of details in the final Digital Building Model is based on the number of recursions during the MBR processing, which in turn are determined by the LiDAR point density. The model-based boundary fitting improves the quality of the generated boundaries and as seen in experimental results, the quality depends on the average LiDAR point spacing. This research thus develops an approach which not only automates the building model generation, but also achieves the best accuracy of the model while utilizing only LiDAR data.