Science.gov

Sample records for updating recursive xml

  1. Soft sensor modelling by time difference, recursive partial least squares and adaptive model updating

    NASA Astrophysics Data System (ADS)

    Fu, Y.; Yang, W.; Xu, O.; Zhou, L.; Wang, J.

    2017-04-01

    To investigate time-variant and nonlinear characteristics in industrial processes, a soft sensor modelling method based on time difference, moving-window recursive partial least square (PLS) and adaptive model updating is proposed. In this method, time difference values of input and output variables are used as training samples to construct the model, which can reduce the effects of the nonlinear characteristic on modelling accuracy and retain the advantages of recursive PLS algorithm. To solve the high updating frequency of the model, a confidence value is introduced, which can be updated adaptively according to the results of the model performance assessment. Once the confidence value is updated, the model can be updated. The proposed method has been used to predict the 4-carboxy-benz-aldehyde (CBA) content in the purified terephthalic acid (PTA) oxidation reaction process. The results show that the proposed soft sensor modelling method can reduce computation effectively, improve prediction accuracy by making use of process information and reflect the process characteristics accurately.

  2. XML in Libraries.

    ERIC Educational Resources Information Center

    Tennant, Roy, Ed.

    This book presents examples of how libraries are using XML (eXtensible Markup Language) to solve problems, expand services, and improve systems. Part I contains papers on using XML in library catalog records: "Updating MARC Records with XMLMARC" (Kevin S. Clarke, Stanford University) and "Searching and Retrieving XML Records via the…

  3. SU-E-T-327: The Update of a XML Composing Tool for TrueBeam Developer Mode

    SciTech Connect

    Yan, Y; Mao, W; Jiang, S

    2014-06-01

    Purpose: To introduce a major upgrade of a novel XML beam composing tool to scientists and engineers who strive to translate certain capabilities of TrueBeam Developer Mode to future clinical benefits of radiation therapy. Methods: TrueBeam Developer Mode provides the users with a test bed for unconventional plans utilizing certain unique features not accessible at the clinical mode. To access the full set of capabilities, a XML beam definition file accommodating all parameters including kV/MV imaging triggers in the plan can be locally loaded at this mode, however it is difficult and laborious to compose one in a text editor. In this study, a stand-along interactive XML beam composing application, TrueBeam TeachMod, was developed on Windows platforms to assist users in making their unique plans in a WYSWYG manner. A conventional plan can be imported in a DICOM RT object as the start of the beam editing process in which trajectories of all axes of a TrueBeam machine can be modified to the intended values at any control point. TeachMod also includes libraries of predefined imaging and treatment procedures to further expedite the process. Results: The TeachMod application is a major of the TeachMod module within DICOManTX. It fully supports TrueBeam 2.0. Trajectories of all axes including all MLC leaves can be graphically rendered and edited as needed. The time for XML beam composing has been reduced to a negligible amount regardless the complexity of the plan. A good understanding of XML language and TrueBeam schema is not required though preferred. Conclusion: Creating XML beams manually in a text editor will be a lengthy error-prone process for sophisticated plans. A XML beam composing tool is highly desirable for R and D activities. It will bridge the gap between scopes of TrueBeam capabilities and their clinical application potentials.

  4. XML under the Hood.

    ERIC Educational Resources Information Center

    Scharf, David

    2002-01-01

    Discusses XML (extensible markup language), particularly as it relates to libraries. Topics include organizing information; cataloging; metadata; similarities to HTML; organizations dealing with XML; making XML useful; a history of XML; the semantic Web; related technologies; XML at the Library of Congress; and its role in improving the…

  5. Recursive SAR imaging

    NASA Astrophysics Data System (ADS)

    Moses, Randolph L.; Ash, Joshua N.

    2008-04-01

    We investigate a recursive procedure for synthetic aperture imaging. We consider a concept in which a SAR system persistently interrogates a scene, for example as it flies along or around that scene. In traditional SAR imaging, the radar measurements are processed in blocks, by partitioning the data into a set of non-overlapping or overlapping azimuth angles, then processing each block. We consider a recursive update approach, in which the SAR image is continually updated, as a linear combination of a small number of previous images and a term containing the current radar measurement. We investigate the crossrange sidelobes realized by such an imaging approach. We show that a first-order autoregression of the image gives crossrange sidelobes similar to a rectangular azimuth window, while a third-order autoregression gives sidelobes comparable to those obtained from widely-used windows in block-processing image formation. The computational and memory requirements of the recursive imaging approach are modest - on the order of M • N2 where M is the recursion order (typically <= 3) and N2 is the image size. We compare images obtained from the recursive and block processing techniques, both for a synthetic scene and for X-band SAR measurements from the Gotcha data set.

  6. Recursion Mathematics.

    ERIC Educational Resources Information Center

    Olson, Alton T.

    1989-01-01

    Discusses the use of the recursive method to permutations of n objects and a problem making c cents in change using pennies and nickels when order is important. Presents a LOGO program for the examples. (YP)

  7. ScotlandsPlaces XML: Bespoke XML or XML Mapping?

    ERIC Educational Resources Information Center

    Beamer, Ashley; Gillick, Mark

    2010-01-01

    Purpose: The purpose of this paper is to investigate web services (in the form of parameterised URLs), specifically in the context of the ScotlandsPlaces project. This involves cross-domain querying, data retrieval and display via the development of a bespoke XML standard rather than existing XML formats and mapping between them.…

  8. On recursion

    PubMed Central

    Watumull, Jeffrey; Hauser, Marc D.; Roberts, Ian G.; Hornstein, Norbert

    2014-01-01

    It is a truism that conceptual understanding of a hypothesis is required for its empirical investigation. However, the concept of recursion as articulated in the context of linguistic analysis has been perennially confused. Nowhere has this been more evident than in attempts to critique and extend Hauseretal's. (2002) articulation. These authors put forward the hypothesis that what is uniquely human and unique to the faculty of language—the faculty of language in the narrow sense (FLN)—is a recursive system that generates and maps syntactic objects to conceptual-intentional and sensory-motor systems. This thesis was based on the standard mathematical definition of recursion as understood by Gödel and Turing, and yet has commonly been interpreted in other ways, most notably and incorrectly as a thesis about the capacity for syntactic embedding. As we explain, the recursiveness of a function is defined independent of such output, whether infinite or finite, embedded or unembedded—existent or non-existent. And to the extent that embedding is a sufficient, though not necessary, diagnostic of recursion, it has not been established that the apparent restriction on embedding in some languages is of any theoretical import. Misunderstanding of these facts has generated research that is often irrelevant to the FLN thesis as well as to other theories of language competence that focus on its generative power of expression. This essay is an attempt to bring conceptual clarity to such discussions as well as to future empirical investigations by explaining three criterial properties of recursion: computability (i.e., rules in intension rather than lists in extension); definition by induction (i.e., rules strongly generative of structure); and mathematical induction (i.e., rules for the principled—and potentially unbounded—expansion of strongly generated structure). By these necessary and sufficient criteria, the grammars of all natural languages are recursive. PMID

  9. XML Style Guide

    DTIC Science & Technology

    2015-07-01

    Acronyms ASCII American Standard Code for Information Interchange DAU data acquisition unit DDML data display markup language IHAL...Transfer Standard URI uniform resource identifier W3C World Wide Web Consortium XML extensible markup language XSD XML schema definition XML Style...Style Guide, RCC 125-15, July 2015 1 Introduction The next generation of telemetry systems will rely heavily on extensible markup language (XML

  10. XML-BASED REPRESENTATION

    SciTech Connect

    R. KELSEY

    2001-02-01

    For focused applications with limited user and use application communities, XML can be the right choice for representation. It is easy to use, maintain, and extend and enjoys wide support in commercial and research sectors. When the knowledge and information to be represented is object-based and use of that knowledge and information is a high priority, then XML-based representation should be considered. This paper discusses some of the issues involved in using XML-based representation and presents an example application that successfully uses an XML-based representation.

  11. XML: A Publisher's Perspective.

    ERIC Educational Resources Information Center

    Andrews, Timothy M.

    1999-01-01

    Explains eXtensible Markup Language (XML) and describes how Dow Jones Interactive is using it to improve the news-gathering and dissemination process through intranets and the World Wide Web. Discusses benefits of using XML, the relationship to HyperText Markup Language (HTML), lack of available software tools and industry support, and future…

  12. Adaptable Iterative and Recursive Kalman Filter Schemes

    NASA Technical Reports Server (NTRS)

    Zanetti, Renato

    2014-01-01

    Nonlinear filters are often very computationally expensive and usually not suitable for real-time applications. Real-time navigation algorithms are typically based on linear estimators, such as the extended Kalman filter (EKF) and, to a much lesser extent, the unscented Kalman filter. The Iterated Kalman filter (IKF) and the Recursive Update Filter (RUF) are two algorithms that reduce the consequences of the linearization assumption of the EKF by performing N updates for each new measurement, where N is the number of recursions, a tuning parameter. This paper introduces an adaptable RUF algorithm to calculate N on the go, a similar technique can be used for the IKF as well.

  13. XML: An Introduction.

    ERIC Educational Resources Information Center

    Lewis, John D.

    1998-01-01

    Describes XML (extensible markup language), a new language classification submitted to the World Wide Web Consortium that is defined in terms of both SGML (Standard Generalized Markup Language) and HTML (Hypertext Markup Language), specifically designed for the Internet. Limitations of PDF (Portable Document Format) files for electronic journals…

  14. Recursion, Language, and Starlings

    ERIC Educational Resources Information Center

    Corballis, Michael C.

    2007-01-01

    It has been claimed that recursion is one of the properties that distinguishes human language from any other form of animal communication. Contrary to this claim, a recent study purports to demonstrate center-embedded recursion in starlings. I show that the performance of the birds in this study can be explained by a counting strategy, without any…

  15. How Does XML Help Libraries?

    ERIC Educational Resources Information Center

    Banerjee, Kyle

    2002-01-01

    Discusses XML, how it has transformed the way information is managed and delivered, and its impact on libraries. Topics include how XML differs from other markup languages; the document object model (DOM); style sheets; practical applications for archival materials, interlibrary loans, digital collections, and MARC data; and future possibilities.…

  16. "The Wonder Years" of XML.

    ERIC Educational Resources Information Center

    Gazan, Rich

    2000-01-01

    Surveys the current state of Extensible Markup Language (XML), a metalanguage for creating structured documents that describe their own content, and its implications for information professionals. Predicts that XML will become the common language underlying Web, word processing, and database formats. Also discusses Extensible Stylesheet Language…

  17. XML Schema Languages: Beyond DTD.

    ERIC Educational Resources Information Center

    Ioannides, Demetrios

    2000-01-01

    Discussion of XML (extensible markup language) and the traditional DTD (document type definition) format focuses on efforts of the World Wide Web Consortium's XML schema working group to develop a schema language to replace DTD that will be capable of defining the set of constraints of any possible data resource. (Contains 14 references.) (LRW)

  18. Recursive Deadbeat Controller Design

    NASA Technical Reports Server (NTRS)

    Juang, Jer-Nan; Phan, Minh Q.

    1997-01-01

    This paper presents a recursive algorithm for a deadbeat predictive controller design. The method combines together the concepts of system identification and deadbeat controller designs. It starts with the multi-step output prediction equation and derives the control force in terms of past input and output time histories. The formulation thus derived satisfies simultaneously system identification and deadbeat controller design requirements. As soon as the coefficient matrices are identified satisfying the output prediction equation, no further work is required to compute the deadbeat control gain matrices. The method can be implemented recursively just as any typical recursive system identification techniques.

  19. Baysian recursive image estimation.

    NASA Technical Reports Server (NTRS)

    Nahi, N. E.; Assefi, T.

    1972-01-01

    Discussion of a statistical procedure for treatment of noise-affected images to recover unaffected images by recursive processing with noise background elimination. The feasibility of the application of a recursive linear Kalman filtering technique to image processing is demonstrated. The procedure is applicable to images which are characterized statistically by mean and correlation functions. A time invariant dynamic model is proposed to provide stationary statistics for the scanner output.

  20. Distinctive signatures of recursion

    PubMed Central

    Martins, Maurício Dias

    2012-01-01

    Although recursion has been hypothesized to be a necessary capacity for the evolution of language, the multiplicity of definitions being used has undermined the broader interpretation of empirical results. I propose that only a definition focused on representational abilities allows the prediction of specific behavioural traits that enable us to distinguish recursion from non-recursive iteration and from hierarchical embedding: only subjects able to represent recursion, i.e. to represent different hierarchical dependencies (related by parenthood) with the same set of rules, are able to generalize and produce new levels of embedding beyond those specified a priori (in the algorithm or in the input). The ability to use such representations may be advantageous in several domains: action sequencing, problem-solving, spatial navigation, social navigation and for the emergence of conventionalized communication systems. The ability to represent contiguous hierarchical levels with the same rules may lead subjects to expect unknown levels and constituents to behave similarly, and this prior knowledge may bias learning positively. Finally, a new paradigm to test for recursion is presented. Preliminary results suggest that the ability to represent recursion in the spatial domain recruits both visual and verbal resources. Implications regarding language evolution are discussed. PMID:22688640

  1. Cytometry metadata in XML

    NASA Astrophysics Data System (ADS)

    Leif, Robert C.; Leif, Stephanie H.

    2016-04-01

    Introduction: The International Society for Advancement of Cytometry (ISAC) has created a standard for the Minimum Information about a Flow Cytometry Experiment (MIFlowCyt 1.0). CytometryML will serve as a common metadata standard for flow and image cytometry (digital microscopy). Methods: The MIFlowCyt data-types were created, as is the rest of CytometryML, in the XML Schema Definition Language (XSD1.1). The datatypes are primarily based on the Flow Cytometry and the Digital Imaging and Communication (DICOM) standards. A small section of the code was formatted with standard HTML formatting elements (p, h1, h2, etc.). Results:1) The part of MIFlowCyt that describes the Experimental Overview including the specimen and substantial parts of several other major elements has been implemented as CytometryML XML schemas (www.cytometryml.org). 2) The feasibility of using MIFlowCyt to provide the combination of an overview, table of contents, and/or an index of a scientific paper or a report has been demonstrated. Previously, a sample electronic publication, EPUB, was created that could contain both MIFlowCyt metadata as well as the binary data. Conclusions: The use of CytometryML technology together with XHTML5 and CSS permits the metadata to be directly formatted and together with the binary data to be stored in an EPUB container. This will facilitate: formatting, data- mining, presentation, data verification, and inclusion in structured research, clinical, and regulatory documents, as well as demonstrate a publication's adherence to the MIFlowCyt standard, promote interoperability and should also result in the textual and numeric data being published using web technology without any change in composition.

  2. XML Files: MedlinePlus

    MedlinePlus

    ... Topic Group XML (11 K) ( DTD , 3 K) Definitions of Health Terms These files contain English definitions of health terms. The files contain health terms definitions NIH sources for the definitions These files are ...

  3. Language and Recursion

    NASA Astrophysics Data System (ADS)

    Lowenthal, Francis

    2010-11-01

    This paper examines whether the recursive structure imbedded in some exercises used in the Non Verbal Communication Device (NVCD) approach is actually the factor that enables this approach to favor language acquisition and reacquisition in the case of children with cerebral lesions. For that a definition of the principle of recursion as it is used by logicians is presented. The two opposing approaches to the problem of language development are explained. For many authors such as Chomsky [1] the faculty of language is innate. This is known as the Standard Theory; the other researchers in this field, e.g. Bates and Elman [2], claim that language is entirely constructed by the young child: they thus speak of Language Acquisition. It is also shown that in both cases, a version of the principle of recursion is relevant for human language. The NVCD approach is defined and the results obtained in the domain of language while using this approach are presented: young subjects using this approach acquire a richer language structure or re-acquire such a structure in the case of cerebral lesions. Finally it is shown that exercises used in this framework imply the manipulation of recursive structures leading to regular grammars. It is thus hypothesized that language development could be favored using recursive structures with the young child. It could also be the case that the NVCD like exercises used with children lead to the elaboration of a regular language, as defined by Chomsky [3], which could be sufficient for language development but would not require full recursion. This double claim could reconcile Chomsky's approach with psychological observations made by adherents of the Language Acquisition approach, if it is confirmed by researches combining the use of NVCDs, psychometric methods and the use of Neural Networks. This paper thus suggests that a research group oriented towards this problematic should be organized.

  4. Recursive heuristic classification

    NASA Technical Reports Server (NTRS)

    Wilkins, David C.

    1994-01-01

    The author will describe a new problem-solving approach called recursive heuristic classification, whereby a subproblem of heuristic classification is itself formulated and solved by heuristic classification. This allows the construction of more knowledge-intensive classification programs in a way that yields a clean organization. Further, standard knowledge acquisition and learning techniques for heuristic classification can be used to create, refine, and maintain the knowledge base associated with the recursively called classification expert system. The method of recursive heuristic classification was used in the Minerva blackboard shell for heuristic classification. Minerva recursively calls itself every problem-solving cycle to solve the important blackboard scheduler task, which involves assigning a desirability rating to alternative problem-solving actions. Knowing these ratings is critical to the use of an expert system as a component of a critiquing or apprenticeship tutoring system. One innovation of this research is a method called dynamic heuristic classification, which allows selection among dynamically generated classification categories instead of requiring them to be prenumerated.

  5. Recursion in Aphasia

    ERIC Educational Resources Information Center

    Banreti, Zoltan

    2010-01-01

    This study investigates how aphasic impairment impinges on syntactic and/or semantic recursivity of human language. A series of tests has been conducted with the participation of five Hungarian speaking aphasic subjects and 10 control subjects. Photographs representing simple situations were presented to subjects and questions were asked about…

  6. ADASS Web Database XML Project

    NASA Astrophysics Data System (ADS)

    Barg, M. I.; Stobie, E. B.; Ferro, A. J.; O'Neil, E. J.

    In the spring of 2000, at the request of the ADASS Program Organizing Committee (POC), we began organizing information from previous ADASS conferences in an effort to create a centralized database. The beginnings of this database originated from data (invited speakers, participants, papers, etc.) extracted from HyperText Markup Language (HTML) documents from past ADASS host sites. Unfortunately, not all HTML documents are well formed and parsing them proved to be an iterative process. It was evident at the beginning that if these Web documents were organized in a standardized way, such as XML (Extensible Markup Language), the processing of this information across the Web could be automated, more efficient, and less error prone. This paper will briefly review the many programming tools available for processing XML, including Java, Perl and Python, and will explore the mapping of relational data from our MySQL database to XML.

  7. XWeB: The XML Warehouse Benchmark

    NASA Astrophysics Data System (ADS)

    Mahboubi, Hadj; Darmont, Jérôme

    With the emergence of XML as a standard for representing business data, new decision support applications are being developed. These XML data warehouses aim at supporting On-Line Analytical Processing (OLAP) operations that manipulate irregular XML data. To ensure feasibility of these new tools, important performance issues must be addressed. Performance is customarily assessed with the help of benchmarks. However, decision support benchmarks do not currently support XML features. In this paper, we introduce the XML Warehouse Benchmark (XWeB), which aims at filling this gap. XWeB derives from the relational decision support benchmark TPC-H. It is mainly composed of a test data warehouse that is based on a unified reference model for XML warehouses and that features XML-specific structures, and its associate XQuery decision support workload. XWeB's usage is illustrated by experiments on several XML database management systems.

  8. Setting the Standard: XML on Campus.

    ERIC Educational Resources Information Center

    Rawlins, Mike

    2001-01-01

    Explains what XML (Extensible Markup Language) is; where to find it in a few years (everywhere from Web pages, to database management systems, to common campus applications); issues that will make XML somewhat of an experimental strategy in the near term; and the importance of decision-makers being abreast of XML trends in standards, tools…

  9. Recursive Rational Choice.

    DTIC Science & Technology

    1981-11-01

    402. Putnam , Hilary , [1973], "Recursive Functions and Hierarchies", American Mathematical Monthly, Vol.80, pp.6 8-8 6 . Rice, H.G., [1954...point the reader is also referred to Putnam [19731. The following are useful facts (cf. Kleene [1950]) that we will make reference to subsequently. The...Hierarchy (Cf. Putnam [19731, pp.77-80 and Hermes [1965] pp.192-202). The classification is made in terms of the structure of the definitions that

  10. Recursive scaled DCT

    NASA Astrophysics Data System (ADS)

    Hou, Hsieh-Sheng

    1991-12-01

    Among the various image data compression methods, the discrete cosine transform (DCT) has become the most popular in performing gray-scale image compression and decomposition. However, the computational burden in performing a DCT is heavy. For example, in a regular DCT, at least 11 multiplications are required for processing an 8 X 1 image block. The idea of the scaled-DCT is that more than half the multiplications in a regular DCT are unnecessary, because they can be formulated as scaling factors of the DCT coefficients, and these coefficients may be scaled back in the quantization process. A fast recursive algorithm for computing the scaled-DCT is presented in this paper. The formulations are derived based on practical considerations of applying the scaled-DCT algorithm to image data compression and decompression. These include the considerations of flexibility of processing different sizes of DCT blocks and the actual savings of the required number of arithmetic operations. Due to the recursive nature of this algorithm, a higher-order scaled-DCT can be obtained from two lower-order scaled DCTs. Thus, a scaled-DCT VLSI chip designed according to this algorithm may process different sizes of DCT under software control. To illustrate the unique properties of this recursive scaled-DCT algorithm, the one-dimensional formulations are presented with several examples exhibited in signal flow-graph forms.

  11. gfortran2xml V0.6

    SciTech Connect

    Rasmussen, Craig E.; Rickett, Christopher D.

    2005-06-28

    the software tool, gfortran2xml, creates an XML representation of a parse tree created by the GNU Fortran compiler. The XML output file describes Fortran interfaces (procedure names and associated parameters and return values). gfortran2xml is useful for projects that perform static analysis to modify existing Fortran source files, tools can automatically insert performance monitoring calls to aid in improving the speed of parallel applications. gfortran2xml is used in projects that generate code for Fortran and C/C++ language interoperability.

  12. Recursive implementations of temporal filters for image motion computation.

    PubMed

    Clifford, C W; Langley, K

    2000-05-01

    Efficient algorithms for image motion computation are important for computer vision applications and the modelling of biological vision systems. Intensity-based image motion computation proceeds in two stages: the convolution of linear spatiotemporal filter kernels with the image sequence, followed by the non-linear combination of the filter outputs. If the spatiotemporal extent of the filter kernels is large, then the convolution stage can be very intensive computationally. One effective means of reducing the storage required and computation involved in implementing the temporal convolutions is the introduction of recursive filtering. Non-recursive methods require the number of frames of the image sequence stored at any given time to be equal to the temporal extent of the slowest temporal filter. In contrast, recursive methods encode recent stimulus history implicitly in the values of a small number of variables updated through a series of feedback equations. Recursive filtering reduces the number of values stored in memory during convolution and the number of mathematical operations involved in computing the filters' outputs. This paper extends previous recursive implementations of gradient- and correlation-based motion analysis algorithms [Fleet DJ, Langley K (1995) IEEE PAMI 17: 61-67; Clifford CWG, Ibbotson MR, Langley K (1997) Vis Neurosci 14: 741-749], describing a recursive implementation of causal band-pass temporal filters suitable for use in energy- and phase-based algorithms for image motion computation. It is shown that the filters' temporal frequency tuning curves fit psychophysical estimates of the temporal properties of human visual filters.

  13. Using recursion to compute the inverse of the genomic relationship matrix.

    PubMed

    Misztal, I; Legarra, A; Aguilar, I

    2014-01-01

    Computing the inverse of the genomic relationship matrix using recursion was investigated. A traditional algorithm to invert the numerator relationship matrix is based on the observation that the conditional expectation for an additive effect of 1 animal given the effects of all other animals depends on the effects of its sire and dam only, each with a coefficient of 0.5. With genomic relationships, such an expectation depends on all other genotyped animals, and the coefficients do not have any set value. For each animal, the coefficients plus the conditional variance can be called a genomic recursion. If such recursions are known, the mixed model equations can be solved without explicitly creating the inverse of the genomic relationship matrix. Several algorithms were developed to create genomic recursions. In an algorithm with sequential updates, genomic recursions are created animal by animal. That algorithm can also be used to update a known inverse of a genomic relationship matrix for additional genotypes. In an algorithm with forward updates, a newly computed recursion is immediately applied to update recursions for remaining animals. The computing costs for both algorithms depend on the sparsity pattern of the genomic recursions, but are lower or equal than for regular inversion. An algorithm for proven and young animals assumes that the genomic recursions for young animals contain coefficients only for proven animals. Such an algorithm generates exact genomic EBV in genomic BLUP and is an approximation in single-step genomic BLUP. That algorithm has a cubic cost for the number of proven animals and a linear cost for the number of young animals. The genomic recursions can provide new insight into genomic evaluation and possibly reduce costs of genetic predictions with extremely large numbers of genotypes.

  14. Recursive principal components analysis.

    PubMed

    Voegtlin, Thomas

    2005-10-01

    A recurrent linear network can be trained with Oja's constrained Hebbian learning rule. As a result, the network learns to represent the temporal context associated to its input sequence. The operation performed by the network is a generalization of Principal Components Analysis (PCA) to time-series, called Recursive PCA. The representations learned by the network are adapted to the temporal statistics of the input. Moreover, sequences stored in the network may be retrieved explicitly, in the reverse order of presentation, thus providing a straight-forward neural implementation of a logical stack.

  15. Recursive Objects--An Object Oriented Presentation of Recursion

    ERIC Educational Resources Information Center

    Sher, David B.

    2004-01-01

    Generally, when recursion is introduced to students the concept is illustrated with a toy (Towers of Hanoi) and some abstract mathematical functions (factorial, power, Fibonacci). These illustrate recursion in the same sense that counting to 10 can be used to illustrate a for loop. These are all good illustrations, but do not represent serious…

  16. A recursive technique for adaptive vector quantization

    NASA Technical Reports Server (NTRS)

    Lindsay, Robert A.

    1989-01-01

    Vector Quantization (VQ) is fast becoming an accepted, if not preferred method for image compression. The VQ performs well when compressing all types of imagery including Video, Electro-Optical (EO), Infrared (IR), Synthetic Aperture Radar (SAR), Multi-Spectral (MS), and digital map data. The only requirement is to change the codebook to switch the compressor from one image sensor to another. There are several approaches for designing codebooks for a vector quantizer. Adaptive Vector Quantization is a procedure that simultaneously designs codebooks as the data is being encoded or quantized. This is done by computing the centroid as a recursive moving average where the centroids move after every vector is encoded. When computing the centroid of a fixed set of vectors the resultant centroid is identical to the previous centroid calculation. This method of centroid calculation can be easily combined with VQ encoding techniques. The defined quantizer changes after every encoded vector by recursively updating the centroid of minimum distance which is the selected by the encoder. Since the quantizer is changing definition or states after every encoded vector, the decoder must now receive updates to the codebook. This is done as side information by multiplexing bits into the compressed source data.

  17. Tomcat, Oracle & XML Web Archive

    SciTech Connect

    Cothren, D. C.

    2008-01-01

    The TOX (Tomcat Oracle & XML) web archive is a foundation for development of HTTP-based applications using Tomcat (or some other servlet container) and an Oracle RDBMS. Use of TOX requires coding primarily in PL/SQL, JavaScript, and XSLT, but also in HTML, CSS and potentially Java. Coded in Java and PL/SQL itself, TOX provides the foundation for more complex applications to be built.

  18. Recursive Feature Extraction in Graphs

    SciTech Connect

    2014-08-14

    ReFeX extracts recursive topological features from graph data. The input is a graph as a csv file and the output is a csv file containing feature values for each node in the graph. The features are based on topological counts in the neighborhoods of each nodes, as well as recursive summaries of neighbors' features.

  19. Speed up of XML parsers with PHP language implementation

    NASA Astrophysics Data System (ADS)

    Georgiev, Bozhidar; Georgieva, Adriana

    2012-11-01

    In this paper, authors introduce PHP5's XML implementation and show how to read, parse, and write a short and uncomplicated XML file using Simple XML in a PHP environment. The possibilities for mutual work of PHP5 language and XML standard are described. The details of parsing process with Simple XML are also cleared. A practical project PHP-XML-MySQL presents the advantages of XML implementation in PHP modules. This approach allows comparatively simple search of XML hierarchical data by means of PHP software tools. The proposed project includes database, which can be extended with new data and new XML parsing functions.

  20. XML Translator for Interface Descriptions

    NASA Technical Reports Server (NTRS)

    Boroson, Elizabeth R.

    2009-01-01

    A computer program defines an XML schema for specifying the interface to a generic FPGA from the perspective of software that will interact with the device. This XML interface description is then translated into header files for C, Verilog, and VHDL. User interface definition input is checked via both the provided XML schema and the translator module to ensure consistency and accuracy. Currently, programming used on both sides of an interface is inconsistent. This makes it hard to find and fix errors. By using a common schema, both sides are forced to use the same structure by using the same framework and toolset. This makes for easy identification of problems, which leads to the ability to formulate a solution. The toolset contains constants that allow a programmer to use each register, and to access each field in the register. Once programming is complete, the translator is run as part of the make process, which ensures that whenever an interface is changed, all of the code that uses the header files describing it is recompiled.

  1. XML technology planning database : lessons learned

    NASA Technical Reports Server (NTRS)

    Some, Raphael R.; Neff, Jon M.

    2005-01-01

    A hierarchical Extensible Markup Language(XML) database called XCALIBR (XML Analysis LIBRary) has been developed by Millennium Program to assist in technology investment (ROI) analysis and technology Language Capability the New return on portfolio optimization. The database contains mission requirements and technology capabilities, which are related by use of an XML dictionary. The XML dictionary codifies a standardized taxonomy for space missions, systems, subsystems and technologies. In addition to being used for ROI analysis, the database is being examined for use in project planning, tracking and documentation. During the past year, the database has moved from development into alpha testing. This paper describes the lessons learned during construction and testing of the prototype database and the motivation for moving from an XML taxonomy to a standard XML-based ontology.

  2. XML syntax for clinical laboratory procedure manuals.

    PubMed

    Saadawi, Gilan; Harrison, James H

    2003-01-01

    We have developed a document type description (DTD) in Extensable Markup Language (XML) for clinical laboratory procedures. Our XML syntax can adequately structure a variety of procedure types across different laboratories and is compatible with current procedure standards. The combination of this format with an XML content management system and appropriate style sheets will allow efficient procedure maintenance, distributed access, customized display and effective searching across a large body of test information.

  3. Recursivity in Lingua Cosmica

    NASA Astrophysics Data System (ADS)

    Ollongren, Alexander

    2011-02-01

    In a sequence of papers on the topic of message construction for interstellar communication by means of a cosmic language, the present author has discussed various significant requirements such a lingua should satisfy. The author's Lingua Cosmica is a (meta) system for annotating contents of possibly large-scale messages for ETI. LINCOS, based on formal constructive logic, was primarily designed for dealing with logic contents of messages but is also applicable for denoting structural properties of more general abstractions embedded in such messages. The present paper explains ways and means for achieving this for a special case: recursive entities. As usual two stages are involved: first the domain of discourse is enriched with suitable representations of the entities concerned, after which properties over them can be dealt with within the system itself. As a representative example the case of Russian dolls (Matrjoshka's) is discussed in some detail and relations with linguistic structures in natural languages are briefly exploited.

  4. Recursive adaptive frame integration limited

    NASA Astrophysics Data System (ADS)

    Rafailov, Michael K.

    2006-05-01

    Recursive Frame Integration Limited was proposed as a way to improve frame integration performance and mitigate issues related to high data rate needed for conventional frame integration. The technique applies two thresholds - one tuned for optimum probability of detection, the other to manage required false alarm rate - and allows a non-linear integration process that, along with Signal-to-Noise Ratio (SNR) gain, provides system designers more capability where cost, weight, or power considerations limit system data rate, processing, or memory capability. However, Recursive Frame Integration Limited may have performance issues when single frame SNR is really low. Recursive Adaptive Frame Integration Limited is proposed as a means to improve limited integration performance with really low single frame SNR. It combines the benefits of nonlinear recursive limited frame integration and adaptive thresholds with a kind of conventional frame integration.

  5. Compressing Aviation Data in XML Format

    NASA Technical Reports Server (NTRS)

    Patel, Hemil; Lau, Derek; Kulkarni, Deepak

    2003-01-01

    Design, operations and maintenance activities in aviation involve analysis of variety of aviation data. This data is typically in disparate formats making it difficult to use with different software packages. Use of a self-describing and extensible standard called XML provides a solution to this interoperability problem. XML provides a standardized language for describing the contents of an information stream, performing the same kind of definitional role for Web content as a database schema performs for relational databases. XML data can be easily customized for display using Extensible Style Sheets (XSL). While self-describing nature of XML makes it easy to reuse, it also increases the size of data significantly. Therefore, transfemng a dataset in XML form can decrease throughput and increase data transfer time significantly. It also increases storage requirements significantly. A natural solution to the problem is to compress the data using suitable algorithm and transfer it in the compressed form. We found that XML-specific compressors such as Xmill and XMLPPM generally outperform traditional compressors. However, optimal use of Xmill requires of discovery of optimal options to use while running Xmill. This, in turn, depends on the nature of data used. Manual disc0ver.y of optimal setting can require an engineer to experiment for weeks. We have devised an XML compression advisory tool that can analyze sample data files and recommend what compression tool would work the best for this data and what are the optimal settings to be used with a XML compression tool.

  6. A Novel Navigation Paradigm for XML Repositories.

    ERIC Educational Resources Information Center

    Azagury, Alain; Factor, Michael E.; Maarek, Yoelle S.; Mandler, Benny

    2002-01-01

    Discusses data exchange over the Internet and describes the architecture and implementation of an XML document repository that promotes a navigation paradigm for XML documents based on content and context. Topics include information retrieval and semistructured documents; and file systems as information storage infrastructure, particularly XMLFS.…

  7. XML Content Finally Arrives on the Web!

    ERIC Educational Resources Information Center

    Funke, Susan

    1998-01-01

    Explains extensible markup language (XML) and how it differs from hypertext markup language (HTML) and standard generalized markup language (SGML). Highlights include features of XML, including better formatting of documents, better searching capabilities, multiple uses for hyperlinking, and an increase in Web applications; Web browsers; and what…

  8. Hopf algebras and topological recursion

    NASA Astrophysics Data System (ADS)

    Esteves, João N.

    2015-11-01

    We consider a model for topological recursion based on the Hopf algebra of planar binary trees defined by Loday and Ronco (1998 Adv. Math. 139 293-309 We show that extending this Hopf algebra by identifying pairs of nearest neighbor leaves, and thus producing graphs with loops, we obtain the full recursion formula discovered by Eynard and Orantin (2007 Commun. Number Theory Phys. 1 347-452).

  9. Direct Waveform Inversion: a New Recursive Scheme

    NASA Astrophysics Data System (ADS)

    Zheng, Y.

    2015-12-01

    The goal of the full-waveform inversion (FWI) is to find an Earth's model such that the synthetic waveforms computed using the model fit the observed ones. In practice, such a model is found in the context of the perturbation approach in an iterative fashion. Specifically, to find such a model, one starts from an initial global velocity model and perform model updating iteratively based on the Frechet derivative or single scattering by adjoint methods to minimize some cost function. However, this process often leads to local minima for the nonlinear cost function in the optimization and slow or no convergence when the starting model is far from the true model. To solve for the initial-model dependence and the convergence issue, we show a new direct waveform inversion (DWI) idea to directly invert the waveform data recursively by explicitly enforcing the causality principle. The DWI offers the advantage of assuming no global initial model and no iteration is needed for the model updating. Starting from the source-receiver region, the DWI builds the model outward recursively by fitting the earliest part of the reflection waveforms and the DWI process is always convergent. The DWI combines seismic imaging and velocity model building into one single process and this is in contrast to many industrial applications where seismic imaging/migration and velocity modeling building are done alternatively. The DWI idea is applicable to one-, two-, and three-dimensional spaces. We show numerical examples to support our idea using full waveform data including both free-surface and inter-bed multiples. Using reflection seismic data, we show that the DWI can invert for both velocity and density, separately.

  10. XML, TEI, and Digital Libraries in the Humanities.

    ERIC Educational Resources Information Center

    Nellhaus, Tobin

    2001-01-01

    Describes the history and major features of XML and TEI, discusses their potential utility for the creation of digital libraries, and focuses on XML's application in the humanities, particularly theater and drama studies. Highlights include HTML and hyperlinks; the impact of XML on text encoding and document access; and XML and academic…

  11. An Introduction to the Extensible Markup Language (XML).

    ERIC Educational Resources Information Center

    Bryan, Martin

    1998-01-01

    Describes Extensible Markup Language (XML), a subset of the Standard Generalized Markup Language (SGML) that is designed to make it easy to interchange structured documents over the Internet. Topics include Document Type Definition (DTD), components of XML, the use of XML, text and non-text elements, and uses for XML-coded files. (LRW)

  12. XML Format for SESAME and LEOS

    SciTech Connect

    Durrenberger, J K; Neely, J R; Sterne, P A

    2009-04-29

    The objective of this document is to describe the XML format used by LLNL and LANL to represent the equation-of-state and related material information in the LEOS and SESAME data libraries. The primary purpose of this document is to describe a specific XML format for representing EOS data that is tailored to the nature of the underlying data and is amenable to conversion to both legacy SESAME and LEOS binary formats. The secondary purpose is to describe an XML format that lends itself to a 'natural' representation in a binary file format of the SESAME, pdb or hdf5 form so that this format and related tools can be used for the rapid and efficient development and implementation of prototype data structures. This document describes the XML format only. A working knowledge of LEOS and SESAME formats is assumed.

  13. XML Schema Guide for Primary CDR Submissions

    EPA Pesticide Factsheets

    This document presents the extensible markup language (XML) schema guide for the Office of Pollution Prevention and Toxics’ (OPPT) e-CDRweb tool. E-CDRweb is the electronic, web-based tool provided by Environmental Protection Agency (EPA) for the submission of Chemical Data Reporting (CDR) information. This document provides the user with tips and guidance on correctly using the version 1.7 XML schema. Please note that the order of the elements must match the schema.

  14. XML Flight/Ground Data Dictionary Management

    NASA Technical Reports Server (NTRS)

    Wright, Jesse; Wiklow, Colette

    2007-01-01

    A computer program generates Extensible Markup Language (XML) files that effect coupling between the command- and telemetry-handling software running aboard a spacecraft and the corresponding software running in ground support systems. The XML files are produced by use of information from the flight software and from flight-system engineering. The XML files are converted to legacy ground-system data formats for command and telemetry, transformed into Web-based and printed documentation, and used in developing new ground-system data-handling software. Previously, the information about telemetry and command was scattered in various paper documents that were not synchronized. The process of searching and reading the documents was time-consuming and introduced errors. In contrast, the XML files contain all of the information in one place. XML structures can evolve in such a manner as to enable the addition, to the XML files, of the metadata necessary to track the changes and the associated documentation. The use of this software has reduced the extent of manual operations in developing a ground data system, thereby saving considerable time and removing errors that previously arose in the translation and transcription of software information from the flight to the ground system.

  15. How Learning Logic Programming Affects Recursion Comprehension

    ERIC Educational Resources Information Center

    Haberman, Bruria

    2004-01-01

    Recursion is a central concept in computer science, yet it is difficult for beginners to comprehend. Israeli high-school students learn recursion in the framework of a special modular program in computer science (Gal-Ezer & Harel, 1999). Some of them are introduced to the concept of recursion in two different paradigms: the procedural…

  16. Core Recursive Hierarchical Image Segmentation

    NASA Technical Reports Server (NTRS)

    Tilton, James

    2011-01-01

    The Recursive Hierarchical Image Segmentation (RHSEG) software has been repackaged to provide a version of the RHSEG software that is not subject to patent restrictions and that can be released to the general public through NASA GSFC's Open Source release process. Like the Core HSEG Software Package, this Core RHSEG Software Package also includes a visualization program called HSEGViewer along with a utility program HSEGReader. It also includes an additional utility program called HSEGExtract. The unique feature of the Core RHSEG package is that it is a repackaging of the RHSEG technology designed to specifically avoid the inclusion of the certain software technology. Unlike the Core HSEG package, it includes the recursive portions of the technology, but does not include processing window artifact elimination technology.

  17. XML-based DICOM data format.

    PubMed

    Yu, Cong; Yao, Zhihong

    2010-04-01

    To enhance the readability, improve the structure, and facilitate the sharing of digital imaging and communications in medicine (DICOM) files, this research proposed one kind of XML-based DICOM data format. Because XML Schema offers great flexibility for expressing constraints on the content model of elements, we used it to describe the new format, thus making it consistent with the one originally defined by DICOM. Meanwhile, such schemas can be used in the creation and validation of the XML-encoded DICOM files, acting as a standard for data transmission and sharing on the Web. Upon defining the new data format, we started with representing a single data element and further improved the whole data structure with the method of modularization. In contrast to the original format, the new one possesses better structure without loss of related information. In addition, we demonstrated the application of XSLT and XQuery. All of the advantages mentioned above resulted from this new data format.

  18. XML, Ontologies, and Their Clinical Applications.

    PubMed

    Yu, Chunjiang; Shen, Bairong

    2016-01-01

    The development of information technology has resulted in its penetration into every area of clinical research. Various clinical systems have been developed, which produce increasing volumes of clinical data. However, saving, exchanging, querying, and exploiting these data are challenging issues. The development of Extensible Markup Language (XML) has allowed the generation of flexible information formats to facilitate the electronic sharing of structured data via networks, and it has been used widely for clinical data processing. In particular, XML is very useful in the fields of data standardization, data exchange, and data integration. Moreover, ontologies have been attracting increased attention in various clinical fields in recent years. An ontology is the basic level of a knowledge representation scheme, and various ontology repositories have been developed, such as Gene Ontology and BioPortal. The creation of these standardized repositories greatly facilitates clinical research in related fields. In this chapter, we discuss the basic concepts of XML and ontologies, as well as their clinical applications.

  19. XML Based Scientific Data Management Facility

    NASA Technical Reports Server (NTRS)

    Mehrotra, P.; Zubair, M.; Bushnell, Dennis M. (Technical Monitor)

    2002-01-01

    The World Wide Web consortium has developed an Extensible Markup Language (XML) to support the building of better information management infrastructures. The scientific computing community realizing the benefits of XML has designed markup languages for scientific data. In this paper, we propose a XML based scientific data management ,facility, XDMF. The project is motivated by the fact that even though a lot of scientific data is being generated, it is not being shared because of lack of standards and infrastructure support for discovering and transforming the data. The proposed data management facility can be used to discover the scientific data itself, the transformation functions, and also for applying the required transformations. We have built a prototype system of the proposed data management facility that can work on different platforms. We have implemented the system using Java, and Apache XSLT engine Xalan. To support remote data and transformation functions, we had to extend the XSLT specification and the Xalan package.

  20. XML Schema Guide for Secondary CDR Submissions

    EPA Pesticide Factsheets

    This document presents the extensible markup language (XML) schema guide for the Office of Pollution Prevention and Toxics’ (OPPT) e-CDRweb tool. E-CDRweb is the electronic, web-based tool provided by Environmental Protection Agency (EPA) for the submission of Chemical Data Reporting (CDR) information. This document provides the user with tips and guidance on correctly using the version 1.1 XML schema for the Joint Submission Form. Please note that the order of the elements must match the schema.

  1. A Survey in Indexing and Searching XML Documents.

    ERIC Educational Resources Information Center

    Luk, Robert W. P.; Leong, H. V.; Dillon, Tharam S.; Chan, Alvin T. S.; Croft, W. Bruce; Allan, James

    2002-01-01

    Discussion of XML focuses on indexing techniques for XML documents, grouping them into flat-file, semistructured, and structured indexing paradigms. Highlights include searching techniques, including full text search and multistage search; search result presentations; database and information retrieval system integration; XML query languages; and…

  2. XML and E-Journals: The State of Play.

    ERIC Educational Resources Information Center

    Wusteman, Judith

    2003-01-01

    Discusses the introduction of the use of XML (Extensible Markup Language) in publishing electronic journals. Topics include standards, including DTDs (Document Type Definition), or document type definitions; aggregator requirements; SGML (Standard Generalized Markup Language); benefits of XML for e-journals; XML metadata; the possibility of…

  3. XTCE. XML Telemetry and Command Exchange Tutorial

    NASA Technical Reports Server (NTRS)

    Rice, Kevin; Kizzort, Brad; Simon, Jerry

    2010-01-01

    An XML Telemetry Command Exchange (XTCE) tutoral oriented towards packets or minor frames is shown. The contents include: 1) The Basics; 2) Describing Telemetry; 3) Describing the Telemetry Format; 4) Commanding; 5) Forgotten Elements; 6) Implementing XTCE; and 7) GovSat.

  4. XML-based resources for simulation

    SciTech Connect

    Kelsey, R. L.; Riese, J. M.; Young, G. A.

    2004-01-01

    As simulations and the machines they run on become larger and more complex the inputs and outputs become more unwieldy. Increased complexity makes the setup of simulation problems difficult. It also contributes to the burden of handling and analyzing large amounts of output results. Another problem is that among a class of simulation codes (such as those for physical system simulation) there is often no single standard format or resource for input data. To run the same problem on different simulations requires a different setup for each simulation code. The extensible Markup Language (XML) is used to represent a general set of data resources including physical system problems, materials, and test results. These resources provide a 'plug and play' approach to simulation setup. For example, a particular material for a physical system can be selected from a material database. The XML-based representation of the selected material is then converted to the native format of the simulation being run and plugged into the simulation input file. In this manner a user can quickly and more easily put together a simulation setup. In the case of output data, an XML approach to regression testing includes tests and test results with XML-based representations. This facilitates the ability to query for specific tests and make comparisons between results. Also, output results can easily be converted to other formats for publishing online or on paper.

  5. Implementing XML Schema Naming and Design Rules

    SciTech Connect

    Lubell, Joshua; Kulvatunyou, Boonserm; Morris, Katherine; Harvey, Betty

    2006-08-01

    We are building a methodology and tool kit for encoding XML schema Naming and Design Rules (NDRs) in a computer-interpretable fashion, enabling automated rule enforcement and improving schema quality. Through our experience implementing rules from various NDR specifications, we discuss some issues and offer practical guidance to organizations grappling with NDR development.

  6. Shuttle-Data-Tape XML Translator

    NASA Technical Reports Server (NTRS)

    Barry, Matthew R.; Osborne, Richard N.

    2005-01-01

    JSDTImport is a computer program for translating native Shuttle Data Tape (SDT) files from American Standard Code for Information Interchange (ASCII) format into databases in other formats. JSDTImport solves the problem of organizing the SDT content, affording flexibility to enable users to choose how to store the information in a database to better support client and server applications. JSDTImport can be dynamically configured by use of a simple Extensible Markup Language (XML) file. JSDTImport uses this XML file to define how each record and field will be parsed, its layout and definition, and how the resulting database will be structured. JSDTImport also includes a client application programming interface (API) layer that provides abstraction for the data-querying process. The API enables a user to specify the search criteria to apply in gathering all the data relevant to a query. The API can be used to organize the SDT content and translate into a native XML database. The XML format is structured into efficient sections, enabling excellent query performance by use of the XPath query language. Optionally, the content can be translated into a Structured Query Language (SQL) database for fast, reliable SQL queries on standard database server computers.

  7. Flight Dynamic Model Exchange using XML

    NASA Technical Reports Server (NTRS)

    Jackson, E. Bruce; Hildreth, Bruce L.

    2002-01-01

    The AIAA Modeling and Simulation Technical Committee has worked for several years to develop a standard by which the information needed to develop physics-based models of aircraft can be specified. The purpose of this standard is to provide a well-defined set of information, definitions, data tables and axis systems so that cooperating organizations can transfer a model from one simulation facility to another with maximum efficiency. This paper proposes using an application of the eXtensible Markup Language (XML) to implement the AIAA simulation standard. The motivation and justification for using a standard such as XML is discussed. Necessary data elements to be supported are outlined. An example of an aerodynamic model as an XML file is given. This example includes definition of independent and dependent variables for function tables, definition of key variables used to define the model, and axis systems used. The final steps necessary for implementation of the standard are presented. Software to take an XML-defined model and import/export it to/from a given simulation facility is discussed, but not demonstrated. That would be the next step in final implementation of standards for physics-based aircraft dynamic models.

  8. Converting from XML to HDF-EOS

    NASA Technical Reports Server (NTRS)

    Ullman, Richard; Bane, Bob; Yang, Jingli

    2008-01-01

    A computer program recreates an HDF-EOS file from an Extensible Markup Language (XML) representation of the contents of that file. This program is one of two programs written to enable testing of the schemas described in the immediately preceding article to determine whether the schemas capture all details of HDF-EOS files.

  9. Recursions for statistical multiple alignment

    PubMed Central

    Hein, Jotun; Jensen, Jens Ledet; Pedersen, Christian N. S.

    2003-01-01

    Algorithms are presented that allow the calculation of the probability of a set of sequences related by a binary tree that have evolved according to the Thorne–Kishino–Felsenstein model for a fixed set of parameters. The algorithms are based on a Markov chain generating sequences and their alignment at nodes in a tree. Depending on whether the complete realization of this Markov chain is decomposed into the first transition and the rest of the realization or the last transition and the first part of the realization, two kinds of recursions are obtained that are computationally similar but probabilistically different. The running time of the algorithms is \\documentclass[12pt]{minimal} \\usepackage{amsmath} \\usepackage{wasysym} \\usepackage{amsfonts} \\usepackage{amssymb} \\usepackage{amsbsy} \\usepackage{mathrsfs} \\setlength{\\oddsidemargin}{-69pt} \\begin{document} \\begin{equation*}O({\\Pi}_{i}^{d}=1~L_{i})\\end{equation*}\\end{document}, where Li is the length of the ith observed sequences and d is the number of sequences. An alternative recursion is also formulated that uses only a Markov chain involving the inner nodes of a tree. PMID:14657378

  10. Using Spreadsheets to Help Students Think Recursively

    ERIC Educational Resources Information Center

    Webber, Robert P.

    2012-01-01

    Spreadsheets lend themselves naturally to recursive computations, since a formula can be defined as a function of one of more preceding cells. A hypothesized closed form for the "n"th term of a recursive sequence can be tested easily by using a spreadsheet to compute a large number of the terms. Similarly, a conjecture about the limit of a series…

  11. The Recursive Paradigm: Suppose We Already Knew.

    ERIC Educational Resources Information Center

    Maurer, Stephen B.

    1995-01-01

    Explains the recursive model in discrete mathematics through five examples and problems. Discusses the relationship between the recursive model, mathematical induction, and inductive reasoning and the relevance of these concepts in the school curriculum. Provides ideas for approaching this material with students. (Author/DDD)

  12. Conjugate gradient algorithms using multiple recursions

    SciTech Connect

    Barth, T.; Manteuffel, T.

    1996-12-31

    Much is already known about when a conjugate gradient method can be implemented with short recursions for the direction vectors. The work done in 1984 by Faber and Manteuffel gave necessary and sufficient conditions on the iteration matrix A, in order for a conjugate gradient method to be implemented with a single recursion of a certain form. However, this form does not take into account all possible recursions. This became evident when Jagels and Reichel used an algorithm of Gragg for unitary matrices to demonstrate that the class of matrices for which a practical conjugate gradient algorithm exists can be extended to include unitary and shifted unitary matrices. The implementation uses short double recursions for the direction vectors. This motivates the study of multiple recursion algorithms.

  13. An exponentiation method for XML element retrieval.

    PubMed

    Wichaiwong, Tanakorn

    2014-01-01

    XML document is now widely used for modelling and storing structured documents. The structure is very rich and carries important information about contents and their relationships, for example, e-Commerce. XML data-centric collections require query terms allowing users to specify constraints on the document structure; mapping structure queries and assigning the weight are significant for the set of possibly relevant documents with respect to structural conditions. In this paper, we present an extension to the MEXIR search system that supports the combination of structural and content queries in the form of content-and-structure queries, which we call the Exponentiation function. It has been shown the structural information improve the effectiveness of the search system up to 52.60% over the baseline BM25 at MAP.

  14. Internet-based data interchange with XML

    NASA Astrophysics Data System (ADS)

    Fuerst, Karl; Schmidt, Thomas

    2000-12-01

    In this paper, a complete concept for Internet Electronic Data Interchange (EDI) - a well-known buzzword in the area of logistics and supply chain management to enable the automation of the interactions between companies and their partners - using XML (eXtensible Markup Language) will be proposed. This approach is based on Internet and XML, because the implementation of traditional EDI (e.g. EDIFACT, ANSI X.12) is mostly too costly for small and medium sized enterprises, which want to integrate their suppliers and customers in a supply chain. The paper will also present the results of the implementation of a prototype for such a system, which has been developed for an industrial partner to improve the current situation of parts delivery. The main functions of this system are an early warning system to detect problems during the parts delivery process as early as possible, and a transport following system to pursue the transportation.

  15. XML Based Scientific Data Management Facility

    NASA Technical Reports Server (NTRS)

    Mehrotra, Piyush; Zubair, M.; Ziebartt, John (Technical Monitor)

    2001-01-01

    The World Wide Web consortium has developed an Extensible Markup Language (XML) to support the building of better information management infrastructures. The scientific computing community realizing the benefits of HTML has designed markup languages for scientific data. In this paper, we propose a XML based scientific data management facility, XDMF. The project is motivated by the fact that even though a lot of scientific data is being generated, it is not being shared because of lack of standards and infrastructure support for discovering and transforming the data. The proposed data management facility can be used to discover the scientific data itself, the transformation functions, and also for applying the required transformations. We have built a prototype system of the proposed data management facility that can work on different platforms. We have implemented the system using Java, and Apache XSLT engine Xalan. To support remote data and transformation functions, we had to extend the XSLT specification and the Xalan package.

  16. A Simple XML Producer-Consumer Protocol

    NASA Technical Reports Server (NTRS)

    Smith, Warren; Gunter, Dan; Quesnel, Darcy

    2000-01-01

    This document describes a simple XML-based protocol that can be used for producers of events to communicate with consumers of events. The protocol described here is not meant to be the most efficient protocol, the most logical protocol, or the best protocol in any way. This protocol was defined quickly and it's intent is to give us a reasonable protocol that we can implement relatively easily and then use to gain experience in distributed event services. This experience will help us evaluate proposals for event representations, XML-based encoding of information, and communication protocols. The next section of this document describes how we represent events in this protocol and then defines the two events that we choose to use for our initial experiments. These definitions are made by example so that they are informal and easy to understand. The following section then proceeds to define the producer-consumer protocol we have agreed upon for our initial experiments.

  17. XML Based Markup Languages for Specific Domains

    NASA Astrophysics Data System (ADS)

    Varde, Aparna; Rundensteiner, Elke; Fahrenholz, Sally

    A challenging area in web based support systems is the study of human activities in connection with the web, especially with reference to certain domains. This includes capturing human reasoning in information retrieval, facilitating the exchange of domain-specific knowledge through a common platform and developing tools for the analysis of data on the web from a domain expert's angle. Among the techniques and standards related to such work, we have XML, the eXtensible Markup Language. This serves as a medium of communication for storing and publishing textual, numeric and other forms of data seamlessly. XML tag sets are such that they preserve semantics and simplify the understanding of stored information by users. Often domain-specific markup languages are designed using XML, with a user-centric perspective. Standardization bodies and research communities may extend these to include additional semantics of areas within and related to the domain. This chapter outlines the issues to be considered in developing domain-specific markup languages: the motivation for development, the semantic considerations, the syntactic constraints and other relevant aspects, especially taking into account human factors. Illustrating examples are provided from domains such as Medicine, Finance and Materials Science. Particular emphasis in these examples is on the Materials Markup Language MatML and the semantics of one of its areas, namely, the Heat Treating of Materials. The focus of this chapter, however, is not the design of one particular language but rather the generic issues concerning the development of domain-specific markup languages.

  18. Application of recursive approaches to differential orbit correction of near Earth asteroids

    NASA Astrophysics Data System (ADS)

    Dmitriev, Vasily; Lupovka, Valery; Gritsevich, Maria

    2016-10-01

    Comparison of three approaches to the differential orbit correction of celestial bodies was performed: batch least squares fitting, Kalman filter, and recursive least squares filter. The first two techniques are well known and widely used (Montenbruck, O. & Gill, E., 2000). The most attention is paid to the algorithm and details of program realization of recursive least squares filter. The filter's algorithm was derived based on recursive least squares technique that are widely used in data processing applications (Simon, D, 2006). Usage recursive least squares filter, makes possible to process a new set of observational data, without reprocessing data, which has been processed before. Specific feature of such approach is that number of observation in data set may be variable. This feature makes recursive least squares filter more flexible approach compare to batch least squares (process complete set of observations in each iteration) and Kalman filtering (suppose updating state vector on each epoch with measurements).Advantages of proposed approach are demonstrated by processing of real astrometric observations of near Earth asteroids. The case of 2008 TC3 was studied. 2008 TC3 was discovered just before its impact with Earth. There are a many closely spaced observations of 2008 TC3 on the interval between discovering and impact, which creates favorable conditions for usage of recursive approaches. Each of approaches has very similar precision in case of 2008 TC3. At the same time, recursive least squares approaches have much higher performance. Thus, this approach more favorable for orbit fitting of a celestial body, which was detected shortly before the collision or close approach to the Earth.This work was carried out at MIIGAiK and supported by the Russian Science Foundation, Project no. 14-22-00197.References:O. Montenbruck and E. Gill, "Satellite Orbits, Models, Methods and Applications," Springer-Verlag, 2000, pp. 1-369.D. Simon, "Optimal State Estimation

  19. δ-dependency for privacy-preserving XML data publishing.

    PubMed

    Landberg, Anders H; Nguyen, Kinh; Pardede, Eric; Rahayu, J Wenny

    2014-08-01

    An ever increasing amount of medical data such as electronic health records, is being collected, stored, shared and managed in large online health information systems and electronic medical record systems (EMR) (Williams et al., 2001; Virtanen, 2009; Huang and Liou, 2007) [1-3]. From such rich collections, data is often published in the form of census and statistical data sets for the purpose of knowledge sharing and enabling medical research. This brings with it an increasing need for protecting individual people privacy, and it becomes an issue of great importance especially when information about patients is exposed to the public. While the concept of data privacy has been comprehensively studied for relational data, models and algorithms addressing the distinct differences and complex structure of XML data are yet to be explored. Currently, the common compromise method is to convert private XML data into relational data for publication. This ad hoc approach results in significant loss of useful semantic information previously carried in the private XML data. Health data often has very complex structure, which is best expressed in XML. In fact, XML is the standard format for exchanging (e.g. HL7 version 3(1)) and publishing health information. Lack of means to deal directly with data in XML format is inevitably a serious drawback. In this paper we propose a novel privacy protection model for XML, and an algorithm for implementing this model. We provide general rules, both for transforming a private XML schema into a published XML schema, and for mapping private XML data to the new privacy-protected published XML data. In addition, we propose a new privacy property, δ-dependency, which can be applied to both relational and XML data, and that takes into consideration the hierarchical nature of sensitive data (as opposed to "quasi-identifiers"). Lastly, we provide an implementation of our model, algorithm and privacy property, and perform an experimental analysis

  20. Recursive sequences in first-year calculus

    NASA Astrophysics Data System (ADS)

    Krainer, Thomas

    2016-02-01

    This article provides ready-to-use supplementary material on recursive sequences for a second-semester calculus class. It equips first-year calculus students with a basic methodical procedure based on which they can conduct a rigorous convergence or divergence analysis of many simple recursive sequences on their own without the need to invoke inductive arguments as is typically required in calculus textbooks. The sequences that are accessible to this kind of analysis are predominantly (eventually) monotonic, but also certain recursive sequences that alternate around their limit point as they converge can be considered.

  1. Experimental Evaluation of Processing Time for the Synchronization of XML-Based Business Objects

    NASA Astrophysics Data System (ADS)

    Ameling, Michael; Wolf, Bernhard; Springer, Thomas; Schill, Alexander

    Business objects (BOs) are data containers for complex data structures used in business applications such as Supply Chain Management and Customer Relationship Management. Due to the replication of application logic, multiple copies of BOs are created which have to be synchronized and updated. This is a complex and time consuming task because BOs rigorously vary in their structure according to the distribution, number and size of elements. Since BOs are internally represented as XML documents, the parsing of XML is one major cost factor which has to be considered for minimizing the processing time during synchronization. The prediction of the parsing time for BOs is an significant property for the selection of an efficient synchronization mechanism. In this paper, we present a method to evaluate the influence of the structure of BOs on their parsing time. The results of our experimental evaluation incorporating four different XML parsers examine the dependencies between the distribution of elements and the parsing time. Finally, a general cost model will be validated and simplified according to the results of the experimental setup.

  2. Software Development Of XML Parser Based On Algebraic Tools

    NASA Astrophysics Data System (ADS)

    Georgiev, Bozhidar; Georgieva, Adriana

    2011-12-01

    In this paper, is presented one software development and implementation of an algebraic method for XML data processing, which accelerates XML parsing process. Therefore, the proposed in this article nontraditional approach for fast XML navigation with algebraic tools contributes to advanced efforts in the making of an easier user-friendly API for XML transformations. Here the proposed software for XML documents processing (parser) is easy to use and can manage files with strictly defined data structure. The purpose of the presented algorithm is to offer a new approach for search and restructuring hierarchical XML data. This approach permits fast XML documents processing, using algebraic model developed in details in previous works of the same authors. So proposed parsing mechanism is easy accessible to the web consumer who is able to control XML file processing, to search different elements (tags) in it, to delete and to add a new XML content as well. The presented various tests show higher rapidity and low consumption of resources in comparison with some existing commercial parsers.

  3. Method for implementation of recursive hierarchical segmentation on parallel computers

    NASA Technical Reports Server (NTRS)

    Tilton, James C. (Inventor)

    2005-01-01

    A method, computer readable storage, and apparatus for implementing a recursive hierarchical segmentation algorithm on a parallel computing platform. The method includes setting a bottom level of recursion that defines where a recursive division of an image into sections stops dividing, and setting an intermediate level of recursion where the recursive division changes from a parallel implementation into a serial implementation. The segmentation algorithm is implemented according to the set levels. The method can also include setting a convergence check level of recursion with which the first level of recursion communicates with when performing a convergence check.

  4. Is recursion language-specific? Evidence of recursive mechanisms in the structure of intentional action.

    PubMed

    Vicari, Giuseppe; Adenzato, Mauro

    2014-05-01

    In their 2002 seminal paper Hauser, Chomsky and Fitch hypothesize that recursion is the only human-specific and language-specific mechanism of the faculty of language. While debate focused primarily on the meaning of recursion in the hypothesis and on the human-specific and syntax-specific character of recursion, the present work focuses on the claim that recursion is language-specific. We argue that there are recursive structures in the domain of motor intentionality by way of extending John R. Searle's analysis of intentional action. We then discuss evidence from cognitive science and neuroscience supporting the claim that motor-intentional recursion is language-independent and suggest some explanatory hypotheses: (1) linguistic recursion is embodied in sensory-motor processing; (2) linguistic and motor-intentional recursions are distinct and mutually independent mechanisms. Finally, we propose some reflections about the epistemic status of HCF as presenting an empirically falsifiable hypothesis, and on the possibility of testing recursion in different cognitive domains.

  5. Update on the Development and Validation of MERCURY: A Modern, Monte Carlo Particle Transport Code

    SciTech Connect

    Procassini, R J; Taylor, J M; McKinley, M S; Greenman, G M; Cullen, D E; O'Brien, M J; Beck, B R; Hagmann, C A

    2005-06-06

    An update on the development and validation of the MERCURY Monte Carlo particle transport code is presented. MERCURY is a modern, parallel, general-purpose Monte Carlo code being developed at the Lawrence Livermore National Laboratory. During the past year, several major algorithm enhancements have been completed. These include the addition of particle trackers for 3-D combinatorial geometry (CG), 1-D radial meshes, 2-D quadrilateral unstructured meshes, as well as a feature known as templates for defining recursive, repeated structures in CG. New physics capabilities include an elastic-scattering neutron thermalization model, support for continuous energy cross sections and S ({alpha}, {beta}) molecular bound scattering. Each of these new physics features has been validated through code-to-code comparisons with another Monte Carlo transport code. Several important computer science features have been developed, including an extensible input-parameter parser based upon the XML data description language, and a dynamic load-balance methodology for efficient parallel calculations. This paper discusses the recent work in each of these areas, and describes a plan for future extensions that are required to meet the needs of our ever expanding user base.

  6. Recursive Inversion Of Externally Defined Linear Systems

    NASA Technical Reports Server (NTRS)

    Bach, Ralph E., Jr.; Baram, Yoram

    1992-01-01

    Technical memorandum discusses mathematical technique described in "Recursive Inversion by Finite-Impulse-Response Filters" (ARC-12247). Technique is recursive algorithm yielding finite-impulse-response approximation of unknown single-input/single-output, causal, time-invariant, linear, real system, response of which is sequence of impulses. Useful in such diverse applications as medical diagnoses, identification of military targets, geophysical exploration, and nondestructive testing.

  7. A Simple XML Producer-Consumer Protocol

    NASA Technical Reports Server (NTRS)

    Smith, Warren; Gunter, Dan; Quesnel, Darcy; Biegel, Bryan (Technical Monitor)

    2001-01-01

    There are many different projects from government, academia, and industry that provide services for delivering events in distributed environments. The problem with these event services is that they are not general enough to support all uses and they speak different protocols so that they cannot interoperate. We require such interoperability when we, for example, wish to analyze the performance of an application in a distributed environment. Such an analysis might require performance information from the application, computer systems, networks, and scientific instruments. In this work we propose and evaluate a standard XML-based protocol for the transmission of events in distributed systems. One recent trend in government and academic research is the development and deployment of computational grids. Computational grids are large-scale distributed systems that typically consist of high-performance compute, storage, and networking resources. Examples of such computational grids are the DOE Science Grid, the NASA Information Power Grid (IPG), and the NSF Partnerships for Advanced Computing Infrastructure (PACIs). The major effort to deploy these grids is in the area of developing the software services to allow users to execute applications on these large and diverse sets of resources. These services include security, execution of remote applications, managing remote data, access to information about resources and services, and so on. There are several toolkits for providing these services such as Globus, Legion, and Condor. As part of these efforts to develop computational grids, the Global Grid Forum is working to standardize the protocols and APIs used by various grid services. This standardization will allow interoperability between the client and server software of the toolkits that are providing the grid services. The goal of the Performance Working Group of the Grid Forum is to standardize protocols and representations related to the storage and distribution of

  8. Using Heaps in Recursive Hierarchical Segmentation of Data

    NASA Technical Reports Server (NTRS)

    Tilton, James C.

    2006-01-01

    A modification to increase processing speed has been made in the algorithm and implementing software reported in Modified Recursive Hierarchical Segmentation of Data (GSC-14681-1), NASA Tech Briefs, Vol. 30, No. 6 (June 2006), page 51. That software performs recursive hierarchical segmentation of data having spatial characteristics (e.g., spectral-image data). The segmentation process includes an iterative subprocess, in each iteration of which it is necessary to determine a best pair of regions to merge [merges being justified by one or more measure(s) similarity of pixels in the regions]. In the previously reported version of the algorithm and software, the choice of a best pair of regions to merge involved the use of a fully sorted list of regions. That version was computationally inefficient because a fully sorted list is not needed: what is needed is only the identity of the pair of regions characterized by the smallest measure of dissimilarity. The present modification replaces the use of a fully sorted list with the use of data heaps, which are computationally more efficient for performing the required comparisons among dissimilarity measures. The modification includes the incorporation of standard and modified functions for creating and updating data heaps

  9. An XML file format for exchanging singlet lens specifications

    NASA Astrophysics Data System (ADS)

    Gay, Shawn C.; Gangadhara, Sanjay

    2015-10-01

    Zemax has developed an XML schema for the distribution of singlet lens specifications based on the ISO 10110 standard. In OpticStudio 15, this kind of XML data can be exported from the ISO Element Drawing analysis. The data file is then used in a feature that automates exchange of lens data between designer and manufacturer, the Cost Estimator. This Cost Estimator feature submits the XML data to various manufacturers to obtain cost estimates for prototype lens production. The workflow centered on the XML data exchange facilitates rapid cost estimate retrieval and eliminates the need for redundant manual data entry. The XML Schema Definition (XSD) for the XML format can be used with Microsoft developer tools to automatically create .NET classes to serialize and deserialize the singlet lens data to/from XML files. The format provides flexible unit specification for most parameters. Choosing XML as the basis for the file format has provided several benefits, such as the above mentioned automated serialization capabilities in .NET, a human-readable text-based format, and ready support for consumption by web services.

  10. An Expressive and Efficient Language for XML Information Retrieval.

    ERIC Educational Resources Information Center

    Chinenyanga, Taurai Tapiwa; Kushmerick, Nicholas

    2002-01-01

    Discusses XML and information retrieval and describes a query language, ELIXIR (expressive and efficient language for XML information retrieval), with a textual similarity operator that can be used for similarity joins. Explains the algorithm for answering ELIXIR queries to generate intermediate relational data. (Author/LRW)

  11. EquiX-A Search and Query Language for XML.

    ERIC Educational Resources Information Center

    Cohen, Sara; Kanza, Yaron; Kogan, Yakov; Sagiv, Yehoshua; Nutt, Werner; Serebrenik, Alexander

    2002-01-01

    Describes EquiX, a search language for XML that combines querying with searching to query the data and the meta-data content of Web pages. Topics include search engines; a data model for XML documents; search query syntax; search query semantics; an algorithm for evaluating a query on a document; and indexing EquiX queries. (LRW)

  12. Analysis of Binary XML Suitability for NATO Tactical Messaging

    DTIC Science & Technology

    2005-09-01

    algorithm used and an element named <CipherData> which either encapsulates the actual cipher value or a reference on how to obtain it (Leung, 2004). C...power for unattended systems. Finally, due to the security nature required from some sensors, the ability to encrypt and decrypt without any additional...XML Encryption .....................................................................9 C. EVOLUTION OF A BINARY XML ENCODING

  13. Does Being Technical Matter? XML, Single Source, and Technical Communication.

    ERIC Educational Resources Information Center

    Sapienza, Filipp

    2002-01-01

    Describes XML, a recent Web design language that will enable technical communicators to produce documentation that can reuse information and present it across multiple types of media for diverse audiences. Argues that XML requires more interdisciplinary approaches toward the teaching and research of technical communication, particularly with…

  14. Adding XML to the MIS Curriculum: Lessons from the Classroom

    ERIC Educational Resources Information Center

    Wagner, William P.; Pant, Vik; Hilken, Ralph

    2008-01-01

    eXtensible Markup Language (XML) is a new technology that is currently being extolled by many industry experts and software vendors. Potentially it represents a platform independent language for sharing information over networks in a way that is much more seamless than with previous technologies. It is extensible in that XML serves as a "meta"…

  15. Data Manipulation in an XML-Based Digital Image Library

    ERIC Educational Resources Information Center

    Chang, Naicheng

    2005-01-01

    Purpose: To help to clarify the role of XML tools and standards in supporting transition and migration towards a fully XML-based environment for managing access to information. Design/methodology/approach: The Ching Digital Image Library, built on a three-tier architecture, is used as a source of examples to illustrate a number of methods of data…

  16. Adaptive Hypermedia Educational System Based on XML Technologies.

    ERIC Educational Resources Information Center

    Baek, Yeongtae; Wang, Changjong; Lee, Sehoon

    This paper proposes an adaptive hypermedia educational system using XML technologies, such as XML, XSL, XSLT, and XLink. Adaptive systems are capable of altering the presentation of the content of the hypermedia on the basis of a dynamic understanding of the individual user. The user profile can be collected in a user model, while the knowledge…

  17. Recursive camera-motion estimation with the trifocal tensor.

    PubMed

    Yu, Ying Kin; Wong, Kin Hong; Chang, Michael Ming Yuen; Or, Siu Hang

    2006-10-01

    In this paper, an innovative extended Kalman filter (EKF) algorithm for pose tracking using the trifocal tensor is proposed. In the EKF, a constant-velocity motion model is used as the dynamic system, and the trifocal-tensor constraint is incorporated into the measurement model. The proposed method has the advantages of those structure- and-motion-based approaches in that the pose sequence can be computed with no prior information on the scene structure. It also has the strengths of those model-based algorithms in which no updating of the three-dimensional (3-D) structure is necessary in the computation. This results in a stable, accurate, and efficient algorithm. Experimental results show that the proposed approach outperformed other existing EKFs that tackle the same problem. An extension to the pose-tracking algorithm has been made to demonstrate the application of the trifocal constraint to fast recursive 3-D structure recovery.

  18. Recursive Branching Simulated Annealing Algorithm

    NASA Technical Reports Server (NTRS)

    Bolcar, Matthew; Smith, J. Scott; Aronstein, David

    2012-01-01

    This innovation is a variation of a simulated-annealing optimization algorithm that uses a recursive-branching structure to parallelize the search of a parameter space for the globally optimal solution to an objective. The algorithm has been demonstrated to be more effective at searching a parameter space than traditional simulated-annealing methods for a particular problem of interest, and it can readily be applied to a wide variety of optimization problems, including those with a parameter space having both discrete-value parameters (combinatorial) and continuous-variable parameters. It can take the place of a conventional simulated- annealing, Monte-Carlo, or random- walk algorithm. In a conventional simulated-annealing (SA) algorithm, a starting configuration is randomly selected within the parameter space. The algorithm randomly selects another configuration from the parameter space and evaluates the objective function for that configuration. If the objective function value is better than the previous value, the new configuration is adopted as the new point of interest in the parameter space. If the objective function value is worse than the previous value, the new configuration may be adopted, with a probability determined by a temperature parameter, used in analogy to annealing in metals. As the optimization continues, the region of the parameter space from which new configurations can be selected shrinks, and in conjunction with lowering the annealing temperature (and thus lowering the probability for adopting configurations in parameter space with worse objective functions), the algorithm can converge on the globally optimal configuration. The Recursive Branching Simulated Annealing (RBSA) algorithm shares some features with the SA algorithm, notably including the basic principles that a starting configuration is randomly selected from within the parameter space, the algorithm tests other configurations with the goal of finding the globally optimal

  19. A recursive Bayesian approach for fatigue damage prognosis: An experimental validation at the reliability component level

    NASA Astrophysics Data System (ADS)

    Gobbato, Maurizio; Kosmatka, John B.; Conte, Joel P.

    2014-04-01

    Fatigue-induced damage is one of the most uncertain and highly unpredictable failure mechanisms for a large variety of mechanical and structural systems subjected to cyclic and random loads during their service life. A health monitoring system capable of (i) monitoring the critical components of these systems through non-destructive evaluation (NDE) techniques, (ii) assessing their structural integrity, (iii) recursively predicting their remaining fatigue life (RFL), and (iv) providing a cost-efficient reliability-based inspection and maintenance plan (RBIM) is therefore ultimately needed. In contribution to these objectives, the first part of the paper provides an overview and extension of a comprehensive reliability-based fatigue damage prognosis methodology — previously developed by the authors — for recursively predicting and updating the RFL of critical structural components and/or sub-components in aerospace structures. In the second part of the paper, a set of experimental fatigue test data, available in the literature, is used to provide a numerical verification and an experimental validation of the proposed framework at the reliability component level (i.e., single damage mechanism evolving at a single damage location). The results obtained from this study demonstrate (i) the importance and the benefits of a nearly continuous NDE monitoring system, (ii) the efficiency of the recursive Bayesian updating scheme, and (iii) the robustness of the proposed framework in recursively updating and improving the RFL estimations. This study also demonstrates that the proposed methodology can lead to either an extent of the RFL (with a consequent economical gain without compromising the minimum safety requirements) or an increase of safety by detecting a premature fault and therefore avoiding a very costly catastrophic failure.

  20. An XML-hierarchical data structure for ENSDF

    SciTech Connect

    Hurst, Aaron M.

    2016-03-15

    A data structure based on an eXtensible Markup Language (XML) hierarchy according to experimental nuclear structure data in the Evaluated Nuclear Structure Data File (ENSDF) is presented. A Python-coded translator has been developed to interpret the standard one-card records of the ENSDF datasets, together with their associated quantities defined according to field position, and generate corresponding representative XML output. The quantities belonging to this mixed-record format are described in the ENSDF manual. Of the 16 ENSDF records in total, XML output has been successfully generated for 15 records. An XML-translation for the Comment Record is yet to be implemented; this will be considered in a separate phase of the overall translation effort. Continuation records, not yet implemented, will also be treated in a future phase of this work. Several examples are presented in this document to illustrate the XML schema and methods for handling the various ENSDF data types. However, the proposed nomenclature for the XML elements and attributes need not necessarily be considered as a fixed set of constructs. Indeed, better conventions may be suggested and a consensus can be achieved amongst the various groups of people interested in this project. The main purpose here is to present an initial phase of the translation effort to demonstrate the feasibility of interpreting ENSDF datasets and creating a representative XML-structured hierarchy for data storage.

  1. Recursion and feedback in image algebra

    NASA Astrophysics Data System (ADS)

    Ritter, Gerhard X.; Davidson, Jennifer L.

    1991-04-01

    Recursion and feedback are two important processes in image processing. Image algebra, a unified algebraic structure developed for use in image processing and image analysis, provides a common mathematical environment for expressing image processing transforms. It is only recently that image algebra has been extended to include recursive operations [1]. Recently image algebra was shown to incorporate neural nets [2], including a new type of neural net, the morphological neural net [3]. This paper presents the relationship of the recursive image algebra to the field of fractions of the ring of matrices, and gives the two dimensional moving average filter as an example. Also, the popular multilayer perceptron with back propagation and a morphology neural network with learning rule are presented in image algebra notation. These examples show that image algebra can express these important feedback concepts in a succinct way.

  2. Using XML to Separate Content from the Presentation Software in eLearning Applications

    ERIC Educational Resources Information Center

    Merrill, Paul F.

    2005-01-01

    This paper has shown how XML (extensible Markup Language) can be used to mark up content. Since XML documents, with meaningful tags, can be interpreted easily by humans as well as computers, they are ideal for the interchange of information. Because XML tags can be defined by an individual or organization, XML documents have proven useful in a…

  3. Recursive algorithms for vector extrapolation methods

    NASA Technical Reports Server (NTRS)

    Ford, William F.; Sidi, Avram

    1988-01-01

    Three classes of recursion relations are devised for implementing some extrapolation methods for vector sequences. One class of recursion relations can be used to implement methods like the modified minimal polynomial extrapolation and the topological epsilon algorithm; another allows implementation of methods like minimal polynomial and reduced rank extrapolation; while the remaining class can be employed in the implementation of the vector E-algorithm. Operation counts and storage requirements for these methods are also discussed, and some related techniques for special applications are also presented. Included are methods for the rapid evaluations of the vector E-algorithm.

  4. Recursive Implementations of the Consider Filter

    NASA Technical Reports Server (NTRS)

    Zanetti, Renato; DSouza, Chris

    2012-01-01

    One method to account for parameters errors in the Kalman filter is to consider their effect in the so-called Schmidt-Kalman filter. This work addresses issues that arise when implementing a consider Kalman filter as a real-time, recursive algorithm. A favorite implementation of the Kalman filter as an onboard navigation subsystem is the UDU formulation. A new way to implement a UDU consider filter is proposed. The non-optimality of the recursive consider filter is also analyzed, and a modified algorithm is proposed to overcome this limitation.

  5. Recursive integral method for transmission eigenvalues

    NASA Astrophysics Data System (ADS)

    Huang, Ruihao; Struthers, Allan A.; Sun, Jiguang; Zhang, Ruming

    2016-12-01

    Transmission eigenvalue problems arise from inverse scattering theory for inhomogeneous media. These non-selfadjoint problems are numerically challenging because of a complicated spectrum. In this paper, we propose a novel recursive contour integral method for matrix eigenvalue problems from finite element discretizations of transmission eigenvalue problems. The technique tests (using an approximate spectral projection) if a region contains eigenvalues. Regions that contain eigenvalues are subdivided and tested recursively until eigenvalues are isolated with a specified precision. The method is fully parallel and requires no a priori spectral information. Numerical examples show the method is effective and robust.

  6. Parallel scheduling of recursively defined arrays

    NASA Technical Reports Server (NTRS)

    Myers, T. J.; Gokhale, M. B.

    1986-01-01

    A new method of automatic generation of concurrent programs which constructs arrays defined by sets of recursive equations is described. It is assumed that the time of computation of an array element is a linear combination of its indices, and integer programming is used to seek a succession of hyperplanes along which array elements can be computed concurrently. The method can be used to schedule equations involving variable length dependency vectors and mutually recursive arrays. Portions of the work reported here have been implemented in the PS automatic program generation system.

  7. The Block recursion library: accurate calculation of resolvent submatrices using the block recursion method

    NASA Astrophysics Data System (ADS)

    Godin, T. J.; Haydock, Roger

    1991-04-01

    The Block Recursion Library, a collection of FORTRAN subroutines, calculates submatrices of the resolvent of a linear operator. The resolvent, in matrix theory, is a powerful tool for extracting information about solutions of linear systems. The routines use the block recursion method and achieve high accuracy for very large systems of coupled equations. This technique is a generalization of the scalar recursion method, an accurate technique for finding the local density of states. A sample program uses these routines to find the quantum mechanical transmittance of a randomly disordered two-dimensional cluster of atoms.

  8. A Recursive Method for Calculating Certain Partition Functions.

    ERIC Educational Resources Information Center

    Woodrum, Luther; And Others

    1978-01-01

    Describes a simple recursive method for calculating the partition function and average energy of a system consisting of N electrons and L energy levels. Also, presents an efficient APL computer program to utilize the recursion relation. (Author/GA)

  9. Recursive stochastic subspace identification for structural parameter estimation

    NASA Astrophysics Data System (ADS)

    Chang, C. C.; Li, Z.

    2009-03-01

    Identification of structural parameters under ambient condition is an important research topic for structural health monitoring and damage identification. This problem is especially challenging in practice as these structural parameters could vary with time under severe excitation. Among the techniques developed for this problem, the stochastic subspace identification (SSI) is a popular time-domain method. The SSI can perform parametric identification for systems with multiple outputs which cannot be easily done using other time-domain methods. The SSI uses the orthogonal-triangular decomposition (RQ) and the singular value decomposition (SVD) to process measured data, which makes the algorithm efficient and reliable. The SSI however processes data in one batch hence cannot be used in an on-line fashion. In this paper, a recursive SSI method is proposed for on-line tracking of time-varying modal parameters for a structure under ambient excitation. The Givens rotation technique, which can annihilate the designated matrix elements, is used to update the RQ decomposition. Instead of updating the SVD, the projection approximation subspace tracking technique which uses an unconstrained optimization technique to track the signal subspace is employed. The proposed technique is demonstrated on the Phase I ASCE benchmark structure. Results show that the technique can identify and track the time-varying modal properties of the building under ambient condition.

  10. A Survey on Teaching and Learning Recursive Programming

    ERIC Educational Resources Information Center

    Rinderknecht, Christian

    2014-01-01

    We survey the literature about the teaching and learning of recursive programming. After a short history of the advent of recursion in programming languages and its adoption by programmers, we present curricular approaches to recursion, including a review of textbooks and some programming methodology, as well as the functional and imperative…

  11. Scheduling Topics for Improved Student Comprehension of Recursion

    ERIC Educational Resources Information Center

    Zmuda, Michael; Hatch, Melanie

    2007-01-01

    This paper presents the results of an experiment conducted to assess the affects of teaching recursion in two disjoint, non-consecutive units of instruction. One group of students was taught basic and advanced recursion topics in four consecutive class periods, while a second group was taught recursion in two two-period blocks that were separated…

  12. Health Topic XML File Description: MedlinePlus

    MedlinePlus

    ... https://medlineplus.gov/xmldescription.html Health Topic XML File Description: MedlinePlus To use the sharing features on ... Tags Definitions of every possible tag in the file, with examples and their use on MedlinePlus.

  13. XML DTD and Schemas for HDF-EOS

    NASA Technical Reports Server (NTRS)

    Ullman, Richard; Yang, Jingli

    2008-01-01

    An Extensible Markup Language (XML) document type definition (DTD) standard for the structure and contents of HDF-EOS files and their contents, and an equivalent standard in the form of schemas, have been developed.

  14. Adaptable recursive binary entropy coding technique

    NASA Astrophysics Data System (ADS)

    Kiely, Aaron B.; Klimesh, Matthew A.

    2002-07-01

    We present a novel data compression technique, called recursive interleaved entropy coding, that is based on recursive interleaving of variable-to variable length binary source codes. A compression module implementing this technique has the same functionality as arithmetic coding and can be used as the engine in various data compression algorithms. The encoder compresses a bit sequence by recursively encoding groups of bits that have similar estimated statistics, ordering the output in a way that is suited to the decoder. As a result, the decoder has low complexity. The encoding process for our technique is adaptable in that each bit to be encoded has an associated probability-of-zero estimate that may depend on previously encoded bits; this adaptability allows more effective compression. Recursive interleaved entropy coding may have advantages over arithmetic coding, including most notably the admission of a simple and fast decoder. Much variation is possible in the choice of component codes and in the interleaving structure, yielding coder designs of varying complexity and compression efficiency; coder designs that achieve arbitrarily small redundancy can be produced. We discuss coder design and performance estimation methods. We present practical encoding and decoding algorithms, as well as measured performance results.

  15. TORTIS (Toddler's Own Recursive Turtle Interpreter System).

    ERIC Educational Resources Information Center

    Perlman, Radia

    TORTIS (Toddler's Own Recursive Turtle Interpreter System) is a device which can be used to study or nurture the cognitive development of preschool children. The device consists of a "turtle" which the child can control by use of buttons on a control panel. The "turtle" can be made to move in prescribed directions, to take a…

  16. Recursive splicing in long vertebrate genes.

    PubMed

    Sibley, Christopher R; Emmett, Warren; Blazquez, Lorea; Faro, Ana; Haberman, Nejc; Briese, Michael; Trabzuni, Daniah; Ryten, Mina; Weale, Michael E; Hardy, John; Modic, Miha; Curk, Tomaž; Wilson, Stephen W; Plagnol, Vincent; Ule, Jernej

    2015-05-21

    It is generally believed that splicing removes introns as single units from precursor messenger RNA transcripts. However, some long Drosophila melanogaster introns contain a cryptic site, known as a recursive splice site (RS-site), that enables a multi-step process of intron removal termed recursive splicing. The extent to which recursive splicing occurs in other species and its mechanistic basis have not been examined. Here we identify highly conserved RS-sites in genes expressed in the mammalian brain that encode proteins functioning in neuronal development. Moreover, the RS-sites are found in some of the longest introns across vertebrates. We find that vertebrate recursive splicing requires initial definition of an 'RS-exon' that follows the RS-site. The RS-exon is then excluded from the dominant mRNA isoform owing to competition with a reconstituted 5' splice site formed at the RS-site after the first splicing step. Conversely, the RS-exon is included when preceded by cryptic promoters or exons that fail to reconstitute an efficient 5' splice site. Most RS-exons contain a premature stop codon such that their inclusion can decrease mRNA stability. Thus, by establishing a binary splicing switch, RS-sites demarcate different mRNA isoforms emerging from long genes by coupling cryptic elements with inclusion of RS-exons.

  17. FRPA: A Framework for Recursive Parallel Algorithms

    DTIC Science & Technology

    2015-05-01

    Math Ker- nel Library (MKL) [4] matrix multiplication routine on “skinny” matrices. Our double-precision Strassen- Winograd implementation, at just...Optimal Par- allel Recursive Rectangular Matrix Multiplication,” in IEEE International Parallel & Distributed Processing Symposium, 2013. [4] Intel, “ Math

  18. Convergence of a Linear Recursive Sequence

    ERIC Educational Resources Information Center

    Tay, E. G.; Toh, T. L.; Dong, F. M.; Lee, T. Y.

    2004-01-01

    A necessary and sufficient condition is found for a linear recursive sequence to be convergent, no matter what initial values are given. Its limit is also obtained when the sequence is convergent. Methods from various areas of mathematics are used to obtain the results.

  19. Recursive Generation of Space-Times

    NASA Astrophysics Data System (ADS)

    Marks, Dennis

    2015-04-01

    Space-times can be generated recursively from a time-like unit basis vector T and a space-like one S. T is unique up to sign, corresponding to particles and antiparticles. S has the form of qubits. Qubits can make quantum transitions, suggesting spontaneous generation of space-time. Recursive generation leads from 2 dimensions to 4, with grades of the resulting algebra corresponding to space-time, spin-area, momentum-energy, and action. Dimensions can be open (like space-time) or closed. A closed time-like dimension has the symmetry of electromagnetism; 3 closed space-like dimensions have the symmetry of the weak force. The 4 open dimensions and the 4 closed dimensions produce an 8-dimensional space with a symmetry that is the product of the Yang regularization of the Heisenberg-Poincaré group and the GUT regularization of the Standard Model. After 8 dimensions, the pattern of real geometric algebras repeats itself, producing a recursive lattice of spontaneously expanding space-time with the physics of the Standard Model at each point of the lattice, implying conservation laws by Noether's theorem. The laws of nature are not preexistent; rather, they are consequences of the uniformity of space-time. The uniformity of space-time is a consequence of its recursive generation.

  20. Recursive inversion of externally defined linear systems

    NASA Technical Reports Server (NTRS)

    Bach, Ralph E., Jr.; Baram, Yoram

    1988-01-01

    The approximate inversion of an internally unknown linear system, given by its impulse response sequence, by an inverse system having a finite impulse response, is considered. The recursive least squares procedure is shown to have an exact initialization, based on the triangular Toeplitz structure of the matrix involved. The proposed approach also suggests solutions to the problems of system identification and compensation.

  1. A Space Surveillance Ontology: Captured in an XML Schema

    DTIC Science & Technology

    2000-10-01

    The commercial industry is making significant investments in XML, especially in the area of electronic business . The Air Force can capitalize upon...migration path could be to add new data as optional elements until all data consumers can handle this new data. (An alternative approach would be to...use XSL to extract the data of interest and send only that subset to the consumer .) Optional elements also allow the same XML schema to be used to

  2. Recursive least-squares algorithms for fast discrete frequency domain equalization

    NASA Astrophysics Data System (ADS)

    Picchi, G.; Prati, G.

    A simple least-squares initialization algorithm (IA) is defined for use with a self-orthogonalizing equalization algorithm in the discrete frequency domain (DFD). A parallel recursive relation is formulated for updating the Kalman vector in the Kalman/Godard algorithm. The DFD is shown to be a modified LS algorithm, thus permitting an exact solution of the LS problem during the equalizer fill-up stage when the data correlation matrix is singular. The solution to the LS problem provides a basis for initialization of the DFD equalizer coefficients. The results of a simulation of on-line initialization of a DFD equalizer with a recursive initialization algorithm demonstrate a weighting capability that minimizes the effects of mean square errors of poorly estimated small-value taps.

  3. Container-component model and XML in ALMA ACS

    NASA Astrophysics Data System (ADS)

    Sommer, Heiko; Chiozzi, Gianluca; Zagar, Klemen; Voelter, Markus

    2004-09-01

    ALMA software, from high-level data flow applications down to instrument control, is built using the ACS framework. To meet the challenges of developing distributed software in distributed teams, ACS offers a container/component model that integrates the use of XML transfer objects. ACS containers are built on top of CORBA and are available for C++, Java, and Python, so that ALMA software can be written as components in any of these languages. The containers perform technical aspects of the software system, while components can focus on the implementation of functional requirements. Like Web services, components can use XML to exchange structured data by value. For Java components, the container seamlessly integrates the use of XML binding classes, which are Java classes that encapsulate access to XML data through type-safe methods. Binding classes are generated from XML schemas, allowing the Java compiler to enforce compliance of application code with the XML schemas. This presentation will explain the capabilities of the ACS container/component model, and how it relates to other middleware technologies that are popular in industry.

  4. Acoustic multiple scattering using recursive algorithms

    NASA Astrophysics Data System (ADS)

    Amirkulova, Feruza A.; Norris, Andrew N.

    2015-10-01

    Acoustic multiple scattering by a cluster of cylinders in an acoustic medium is considered. A fast recursive technique is described which takes advantage of the multilevel Block Toeplitz structure of the linear system. A parallelization technique is described that enables efficient application of the proposed recursive algorithm for solving multilevel Block Toeplitz systems on high performance computer clusters. Numerical comparisons of CPU time and total elapsed time taken to solve the linear system using the direct LAPACK and TOEPLITZ libraries on Intel FORTRAN, show the advantage of the TOEPLITZ solver. Computations are optimized by multi-threading which displays improved efficiency of the TOEPLITZ solver with the increase of the number of scatterers and frequency.

  5. Geomagnetic modeling by optimal recursive filtering

    NASA Technical Reports Server (NTRS)

    Gibbs, B. P.; Estes, R. H.

    1981-01-01

    The results of a preliminary study to determine the feasibility of using Kalman filter techniques for geomagnetic field modeling are given. Specifically, five separate field models were computed using observatory annual means, satellite, survey and airborne data for the years 1950 to 1976. Each of the individual field models used approximately five years of data. These five models were combined using a recursive information filter (a Kalman filter written in terms of information matrices rather than covariance matrices.) The resulting estimate of the geomagnetic field and its secular variation was propogated four years past the data to the time of the MAGSAT data. The accuracy with which this field model matched the MAGSAT data was evaluated by comparisons with predictions from other pre-MAGSAT field models. The field estimate obtained by recursive estimation was found to be superior to all other models.

  6. The recursion relation in Lagrangian perturbation theory

    SciTech Connect

    Rampf, Cornelius

    2012-12-01

    We derive a recursion relation in the framework of Lagrangian perturbation theory, appropriate for studying the inhomogeneities of the large scale structure of the universe. We use the fact that the perturbative expansion of the matter density contrast is in one-to-one correspondence with standard perturbation theory (SPT) at any order. This correspondence has been recently shown to be valid up to fourth order for a non-relativistic, irrotational and dust-like component. Assuming it to be valid at arbitrary (higher) order, we express the Lagrangian displacement field in terms of the perturbative kernels of SPT, which are itself given by their own and well-known recursion relation. We argue that the Lagrangian solution always contains more non-linear information in comparison with the SPT solution, (mainly) if the non-perturbative density contrast is restored after the displacement field is obtained.

  7. Recursive prescription for logarithmic jet rate coefficients

    NASA Astrophysics Data System (ADS)

    Gerwick, Erik

    2013-11-01

    We derive a recursion relation for the analytic leading logarithmic coefficients of a final state gluon cascade. We demonstrate the potential of our method by analytically computing the rate coefficients for the emission of up to 80 gluons in both the exclusive-kt (Durham) and generalized inclusive-kt class of jet algorithms. There is a particularly simple form for the ratios of resolved coefficients. We suggest potential applications for our method including the efficient generation of shower histories.

  8. Realization Of Algebraic Processor For XML Documents Processing

    SciTech Connect

    Georgiev, Bozhidar; Georgieva, Adriana

    2010-10-25

    In this paper, are presented some possibilities concerning the implementation of an algebraic method for XML hierarchical data processing which makes faster the XML search mechanism. Here is offered a different point of view for creation of advanced algebraic processor (with all necessary software tools and programming modules respectively). Therefore, this nontraditional approach for fast XML navigation with the presented algebraic processor may help to build an easier user-friendly interface provided XML transformations, which can avoid the difficulties in the complicated language constructions of XSL, XSLT and XPath. This approach allows comparatively simple search of XML hierarchical data by means of the following types of functions: specification functions and so named build-in functions. The choice of programming language Java may appear strange at first, but it isn't when you consider that the applications can run on different kinds of computers. The specific search mechanism based on the linear algebra theory is faster in comparison with MSXML parsers (on the basis of the developed examples with about 30%). Actually, there exists the possibility for creating new software tools based on the linear algebra theory, which cover the whole navigation and search techniques characterizing XSLT/XPath. The proposed method is able to replace more complicated operations in other SOA components.

  9. The SGML Standardization Framework and the Introduction of XML

    PubMed Central

    Grütter, Rolf

    2000-01-01

    Extensible Markup Language (XML) is on its way to becoming a global standard for the representation, exchange, and presentation of information on the World Wide Web (WWW). More than that, XML is creating a standardization framework, in terms of an open network of meta-standards and mediators that allows for the definition of further conventions and agreements in specific business domains. Such an approach is particularly needed in the healthcare domain; XML promises to especially suit the particularities of patient records and their lifelong storage, retrieval, and exchange. At a time when change rather than steadiness is becoming the faithful feature of our society, standardization frameworks which support a diversified growth of specifications that are appropriate to the actual needs of the users are becoming more and more important; and efforts should be made to encourage this new attempt at standardization to grow in a fruitful direction. Thus, the introduction of XML reflects a standardization process which is neither exclusively based on an acknowledged standardization authority, nor a pure market standard. Instead, a consortium of companies, academic institutions, and public bodies has agreed on a common recommendation based on an existing standardization framework. The consortium's process of agreeing to a standardization framework will doubtlessly be successful in the case of XML, and it is suggested that it should be considered as a generic model for standardization processes in the future. PMID:11720931

  10. Development Life Cycle and Tools for XML Content Models

    SciTech Connect

    Kulvatunyou, Boonserm; Morris, Katherine; Buhwan, Jeong; Goyal, Puja

    2004-11-01

    Many integration projects today rely on shared semantic models based on standards represented using Extensible Mark up Language (XML) technologies. Shared semantic models typically evolve and require maintenance. In addition, to promote interoperability and reduce integration costs, the shared semantics should be reused as much as possible. Semantic components must be consistent and valid in terms of agreed upon standards and guidelines. In this paper, we describe an activity model for creation, use, and maintenance of a shared semantic model that is coherent and supports efficient enterprise integration. We then use this activity model to frame our research and the development of tools to support those activities. We provide overviews of these tools primarily in the context of the W3C XML Schema. At the present, we focus our work on the W3C XML Schema as the representation of choice, due to its extensive adoption by industry.

  11. Semantic reasoning with XML-based biomedical information models.

    PubMed

    O'Connor, Martin J; Das, Amar

    2010-01-01

    The Extensible Markup Language (XML) is increasingly being used for biomedical data exchange. The parallel growth in the use of ontologies in biomedicine presents opportunities for combining the two technologies to leverage the semantic reasoning services provided by ontology-based tools. There are currently no standardized approaches for taking XML-encoded biomedical information models and representing and reasoning with them using ontologies. To address this shortcoming, we have developed a workflow and a suite of tools for transforming XML-based information models into domain ontologies encoded using OWL. In this study, we applied semantics reasoning methods to these ontologies to automatically generate domain-level inferences. We successfully used these methods to develop semantic reasoning methods for information models in the HIV and radiological image domains.

  12. KAT: A Flexible XML-based Knowledge Authoring Environment

    PubMed Central

    Hulse, Nathan C.; Rocha, Roberto A.; Del Fiol, Guilherme; Bradshaw, Richard L.; Hanna, Timothy P.; Roemer, Lorrie K.

    2005-01-01

    As part of an enterprise effort to develop new clinical information systems at Intermountain Health Care, the authors have built a knowledge authoring tool that facilitates the development and refinement of medical knowledge content. At present, users of the application can compose order sets and an assortment of other structured clinical knowledge documents based on XML schemas. The flexible nature of the application allows the immediate authoring of new types of documents once an appropriate XML schema and accompanying Web form have been developed and stored in a shared repository. The need for a knowledge acquisition tool stems largely from the desire for medical practitioners to be able to write their own content for use within clinical applications. We hypothesize that medical knowledge content for clinical use can be successfully created and maintained through XML-based document frameworks containing structured and coded knowledge. PMID:15802477

  13. Utilizing the Structure and Content Information for XML Document Clustering

    NASA Astrophysics Data System (ADS)

    Tran, Tien; Kutty, Sangeetha; Nayak, Richi

    This paper reports on the experiments and results of a clustering approach used in the INEX 2008 document mining challenge. The clustering approach utilizes both the structure and content information of the Wikipedia XML document collection. A latent semantic kernel (LSK) is used to measure the semantic similarity between XML documents based on their content features. The construction of a latent semantic kernel involves the computing of singular vector decomposition (SVD). On a large feature space matrix, the computation of SVD is very expensive in terms of time and memory requirements. Thus in this clustering approach, the dimension of the document space of a term-document matrix is reduced before performing SVD. The document space reduction is based on the common structural information of the Wikipedia XML document collection. The proposed clustering approach has shown to be effective on the Wikipedia collection in the INEX 2008 document mining challenge.

  14. The Harer-Zagier recursion for an irregular spectral curve

    NASA Astrophysics Data System (ADS)

    Chekhov, Leonid O.

    2016-12-01

    We derive the Do and Norbury recursion formula for the one-loop mean of an irregular spectral curve from a variant of replica method by Brezín and Hikami. We express this recursion in special times in which all terms W1(g) of the genus expansion of the one-loop mean are polynomials. We generalize this recursion to the case of generalized Laguerre ensembles corresponding to unicellular bicolored maps with different weights of vertices of different colors.

  15. A standard XML Schema for computerised anaesthetic records.

    PubMed

    Gardner, M; Peachey, T

    2002-12-01

    We believe that one of the most influential developments for the practice of anaesthesia in this decade will be the introduction of a national (or possibly international) standard XML Schema for computerised anaesthetic records, and that such development should be actively promoted by appropriate professional groups. We discuss the implications of such a schema, and make suggestions regarding its requirements. We also report on one approach to the development of an XML Schema for anaesthetic records (provisionally named SnowSchema in honour of Dr John Snow), and compare the current version with the suggested requirements.

  16. XML in an Adaptive Framework for Instrument Control

    NASA Technical Reports Server (NTRS)

    Ames, Troy J.

    2004-01-01

    NASA Goddard Space Flight Center is developing an extensible framework for instrument command and control, known as Instrument Remote Control (IRC), that combines the platform independent processing capabilities of Java with the power of the Extensible Markup Language (XML). A key aspect of the architecture is software that is driven by an instrument description, written using the Instrument Markup Language (IML). IML is an XML dialect used to describe interfaces to control and monitor the instrument, command sets and command formats, data streams, communication mechanisms, and data processing algorithms.

  17. Applying Analogical Reasoning Techniques for Teaching XML Document Querying Skills in Database Classes

    ERIC Educational Resources Information Center

    Mitri, Michel

    2012-01-01

    XML has become the most ubiquitous format for exchange of data between applications running on the Internet. Most Web Services provide their information to clients in the form of XML. The ability to process complex XML documents in order to extract relevant information is becoming as important a skill for IS students to master as querying…

  18. XML: A Language To Manage the World Wide Web. ERIC Digest.

    ERIC Educational Resources Information Center

    Davis-Tanous, Jennifer R.

    This digest provides an overview of XML (Extensible Markup Language), a markup language used to construct World Wide Web pages. Topics addressed include: (1) definition of a markup language, including comparison of XML with SGML (Standard Generalized Markup Language) and HTML (HyperText Markup Language); (2) how XML works, including sample tags,…

  19. Non-recursive sequential input deconvolution

    NASA Astrophysics Data System (ADS)

    Bernal, Dionisio

    2017-01-01

    A scheme for sequential deconvolution of inputs from measured outputs is presented. The key feature in the formulation is elimination of the initial state from the input-output relations by projecting the output in the left null space of the observability block. Removal of the initial state allows the sequential format of the deconvolution, essential for computational reasons, to be implemented non-recursively, assuring unconditional stability. Identifiability is realized when the input-output arrangement does not have transmission zeros, and observability and controllability are shown immaterial. Comparison of results from the scheme with those from Dynamic Programming highlights the benefits of eliminating the initial state.

  20. Recursion equations in gauge field theories

    NASA Astrophysics Data System (ADS)

    Migdal, A. A.

    An approximate recursion equation is formulated, describing the scale transformation of the effective action of a gauge field. In two-dimensional space-time the equation becomes exact. In four-dimensional theories it reproduces asymptotic freedom to an accuracy of 30% in the coefficients of the β-function. In the strong-coupling region the β-function remains negative and this results in an asymptotic prison in the infrared region. Possible generalizations and applications to the quark-gluon gauge theory are discussed.

  1. The Lehmer Matrix and Its Recursive Analogue

    DTIC Science & Technology

    2010-01-01

    for failing to comply with a collection of information if it does not display a currently valid OMB control number . 1. REPORT DATE 2010 2. REPORT...TYPE 3. DATES COVERED 00-00-2010 to 00-00-2010 4. TITLE AND SUBTITLE The Lehmer matrix and its recursive analogue 5a. CONTRACT NUMBER 5b...GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER 6. AUTHOR(S) 5d. PROJECT NUMBER 5e. TASK NUMBER 5f. WORK UNIT NUMBER 7. PERFORMING ORGANIZATION NAME(S) AND

  2. Modified Recursive Hierarchical Segmentation of Data

    NASA Technical Reports Server (NTRS)

    Tilton, James C.

    2006-01-01

    An algorithm and a computer program that implements the algorithm that performs recursive hierarchical segmentation (RHSEG) of data have been developed. While the current implementation is for two-dimensional data having spatial characteristics (e.g., image, spectral, or spectral-image data), the generalized algorithm also applies to three-dimensional or higher dimensional data and also to data with no spatial characteristics. The algorithm and software are modified versions of a prior RHSEG algorithm and software, the outputs of which often contain processing-window artifacts including, for example, spurious segmentation-image regions along the boundaries of processing-window edges.

  3. Personalization of XML Content Browsing Based on User Preferences

    ERIC Educational Resources Information Center

    Encelle, Benoit; Baptiste-Jessel, Nadine; Sedes, Florence

    2009-01-01

    Personalization of user interfaces for browsing content is a key concept to ensure content accessibility. In this direction, we introduce concepts that result in the generation of personalized multimodal user interfaces for browsing XML content. User requirements concerning the browsing of a specific content type can be specified by means of…

  4. An XML-based communication protocol for accelerator distributed controls

    NASA Astrophysics Data System (ADS)

    Catani, L.

    2008-03-01

    This paper presents the development of XMLvRPC, an RPC-like communication protocol based, for this particular application, on the TCP/IP and XML (eXtensible Markup Language) tools built-in in LabVIEW. XML is used to format commands and data passed between client and server while socket interface for communication uses either TCP or UDP transmission protocols. This implementation extends the features of these general purpose libraries and incorporates solutions that might provide, with limited modifications, full compatibility with well established and more general communication protocol, i.e. XML-RPC, while preserving portability to different platforms supported by LabVIEW. The XMLvRPC suite of software has been equipped with specific tools for its deployment in distributed control systems as, for instance, a quasi-automatic configuration and registration of the distributed components and a simple plug-and-play approach to the installation of new services. Key feature is the management of large binary arrays that allow coding of large binary data set, e.g. raw images, more efficiently with respect to the standard XML coding.

  5. A Conversion Tool for Mathematical Expressions in Web XML Files.

    ERIC Educational Resources Information Center

    Ohtake, Nobuyuki; Kanahori, Toshihiro

    2003-01-01

    This article discusses the conversion of mathematical equations into Extensible Markup Language (XML) on the World Wide Web for individuals with visual impairments. A program is described that converts the presentation markup style to the content markup style in MathML to allow browsers to render mathematical expressions without other programs.…

  6. Castles Made of Sand: Building Sustainable Digitized Collections Using XML.

    ERIC Educational Resources Information Center

    Ragon, Bart

    2003-01-01

    Describes work at the University of Virginia library to digitize special collections. Discusses the use of XML (Extensible Markup Language); providing access to original source materials; DTD (Document Type Definition); TEI (Text Encoding Initiative); metadata; XSL (Extensible Style Language); and future possibilities. (LRW)

  7. Recursive support vector machines for dimensionality reduction.

    PubMed

    Tao, Qing; Chu, Dejun; Wang, Jue

    2008-01-01

    The usual dimensionality reduction technique in supervised learning is mainly based on linear discriminant analysis (LDA), but it suffers from singularity or undersampled problems. On the other hand, a regular support vector machine (SVM) separates the data only in terms of one single direction of maximum margin, and the classification accuracy may be not good enough. In this letter, a recursive SVM (RSVM) is presented, in which several orthogonal directions that best separate the data with the maximum margin are obtained. Theoretical analysis shows that a completely orthogonal basis can be derived in feature subspace spanned by the training samples and the margin is decreasing along the recursive components in linearly separable cases. As a result, a new dimensionality reduction technique based on multilevel maximum margin components and then a classifier with high accuracy are achieved. Experiments in synthetic and several real data sets show that RSVM using multilevel maximum margin features can do efficient dimensionality reduction and outperform regular SVM in binary classification problems.

  8. A recursive algorithm for Zernike polynomials

    NASA Technical Reports Server (NTRS)

    Davenport, J. W.

    1982-01-01

    The analysis of a function defined on a rotationally symmetric system, with either a circular or annular pupil is discussed. In order to numerically analyze such systems it is typical to expand the given function in terms of a class of orthogonal polynomials. Because of their particular properties, the Zernike polynomials are especially suited for numerical calculations. Developed is a recursive algorithm that can be used to generate the Zernike polynomials up to a given order. The algorithm is recursively defined over J where R(J,N) is the Zernike polynomial of degree N obtained by orthogonalizing the sequence R(J), R(J+2), ..., R(J+2N) over (epsilon, 1). The terms in the preceding row - the (J-1) row - up to the N+1 term is needed for generating the (J,N)th term. Thus, the algorith generates an upper left-triangular table. This algorithm was placed in the computer with the necessary support program also included.

  9. Tetramethyleneethane Equivalents: Recursive Reagents for Serialized Cycloadditions

    PubMed Central

    2015-01-01

    New reactions and reagents that allow for multiple bond-forming events per synthetic operation are required to achieve structural complexity and thus value with step-, time-, cost-, and waste-economy. Here we report a new class of reagents that function like tetramethyleneethane (TME), allowing for back-to-back [4 + 2] cycloadditions, thereby amplifying the complexity-increasing benefits of Diels–Alder and metal-catalyzed cycloadditions. The parent recursive reagent, 2,3-dimethylene-4-trimethylsilylbutan-1-ol (DMTB), is readily available from the metathesis of ethylene and THP-protected 4-trimethylsilylbutyn-1-ol. DMTB and related reagents engage diverse dienophiles in an initial Diels–Alder or metal-catalyzed [4 + 2] cycloaddition, triggering a subsequent vinylogous Peterson elimination that recursively generates a new diene for a second cycloaddition. Overall, this multicomponent catalytic cascade produces in one operation carbo- and heterobicyclic building blocks for the synthesis of a variety of natural products, therapeutic leads, imaging agents, and materials. Its application to the three step synthesis of a new solvatochromic fluorophore, N-ethyl(6-N,N-dimethylaminoanthracene-2,3-dicarboximide) (6-DMA), and the photophysical characterization of this fluorophore are described. PMID:25961416

  10. Tetramethyleneethane Equivalents: Recursive Reagents for Serialized Cycloadditions.

    PubMed

    Wender, Paul A; Jeffreys, Matthew S; Raub, Andrew G

    2015-07-22

    New reactions and reagents that allow for multiple bond-forming events per synthetic operation are required to achieve structural complexity and thus value with step-, time-, cost-, and waste-economy. Here we report a new class of reagents that function like tetramethyleneethane (TME), allowing for back-to-back [4 + 2] cycloadditions, thereby amplifying the complexity-increasing benefits of Diels-Alder and metal-catalyzed cycloadditions. The parent recursive reagent, 2,3-dimethylene-4-trimethylsilylbutan-1-ol (DMTB), is readily available from the metathesis of ethylene and THP-protected 4-trimethylsilylbutyn-1-ol. DMTB and related reagents engage diverse dienophiles in an initial Diels-Alder or metal-catalyzed [4 + 2] cycloaddition, triggering a subsequent vinylogous Peterson elimination that recursively generates a new diene for a second cycloaddition. Overall, this multicomponent catalytic cascade produces in one operation carbo- and heterobicyclic building blocks for the synthesis of a variety of natural products, therapeutic leads, imaging agents, and materials. Its application to the three step synthesis of a new solvatochromic fluorophore, N-ethyl(6-N,N-dimethylaminoanthracene-2,3-dicarboximide) (6-DMA), and the photophysical characterization of this fluorophore are described.

  11. Recursion Relations for Double Ramification Hierarchies

    NASA Astrophysics Data System (ADS)

    Buryak, Alexandr; Rossi, Paolo

    2016-03-01

    In this paper we study various properties of the double ramification hierarchy, an integrable hierarchy of hamiltonian PDEs introduced in Buryak (CommunMath Phys 336(3):1085-1107, 2015) using intersection theory of the double ramification cycle in the moduli space of stable curves. In particular, we prove a recursion formula that recovers the full hierarchy starting from just one of the Hamiltonians, the one associated to the first descendant of the unit of a cohomological field theory. Moreover, we introduce analogues of the topological recursion relations and the divisor equation both for the Hamiltonian densities and for the string solution of the double ramification hierarchy. This machinery is very efficient and we apply it to various computations for the trivial and Hodge cohomological field theories, and for the r -spin Witten's classes. Moreover, we prove the Miura equivalence between the double ramification hierarchy and the Dubrovin-Zhang hierarchy for the Gromov-Witten theory of the complex projective line (extended Toda hierarchy).

  12. Improvement in Recursive Hierarchical Segmentation of Data

    NASA Technical Reports Server (NTRS)

    Tilton, James C.

    2006-01-01

    A further modification has been made in the algorithm and implementing software reported in Modified Recursive Hierarchical Segmentation of Data (GSC- 14681-1), NASA Tech Briefs, Vol. 30, No. 6 (June 2006), page 51. That software performs recursive hierarchical segmentation of data having spatial characteristics (e.g., spectral-image data). The output of a prior version of the software contained artifacts, including spurious segmentation-image regions bounded by processing-window edges. The modification for suppressing the artifacts, mentioned in the cited article, was addition of a subroutine that analyzes data in the vicinities of seams to find pairs of regions that tend to lie adjacent to each other on opposite sides of the seams. Within each such pair, pixels in one region that are more similar to pixels in the other region are reassigned to the other region. The present modification provides for a parameter ranging from 0 to 1 for controlling the relative priority of merges between spatially adjacent and spatially non-adjacent regions. At 1, spatially-adjacent-/spatially- non-adjacent-region merges have equal priority. At 0, only spatially-adjacent-region merges (no spectral clustering) are allowed. Between 0 and 1, spatially-adjacent- region merges have priority over spatially- non-adjacent ones.

  13. XML does Real Programmers a Service

    SciTech Connect

    Gorton, Ian

    2008-09-01

    As the sun slowly sets of this first decade of the new millenium, it seems appropriate to update the sojourn of the real programmers as they adapt to their ever changing technical and business environment. Real Programmers were perfectly characterized and differentiated from their quiche-eating, Pascal programming brethren in Ed Post’s seminal “Real Programmers Don’t Use Pascal” (Datamation, 1983). My follow-up ("Real programmers do use Delphi," Software, IEEE , vol.12, no.6, pp.8, 10, 12-, Nov 1995) charted their evolution from FORTRAN-only programmers to embracing a wider range of mainstream languages and tools that still afforded ample opportunity for creativity, game-playing, irregular work hours, and importantly, long-term job security.

  14. Progress on an implementation of MIFlowCyt in XML

    NASA Astrophysics Data System (ADS)

    Leif, Robert C.; Leif, Stephanie H.

    2015-03-01

    Introduction: The International Society for Advancement of Cytometry (ISAC) Data Standards Task Force (DSTF) has created a standard for the Minimum Information about a Flow Cytometry Experiment (MIFlowCyt 1.0). The CytometryML schemas, are based in part upon the Flow Cytometry Standard and Digital Imaging and Communication (DICOM) standards. CytometryML has and will be extended and adapted to include MIFlowCyt, as well as to serve as a common standard for flow and image cytometry (digital microscopy). Methods: The MIFlowCyt data-types were created, as is the rest of CytometryML, in the XML Schema Definition Language (XSD1.1). Individual major elements of the MIFlowCyt schema were translated into XML and filled with reasonable data. A small section of the code was formatted with HTML formatting elements. Results: The differences in the amount of detail to be recorded for 1) users of standard techniques including data analysts and 2) others, such as method and device creators, laboratory and other managers, engineers, and regulatory specialists required that separate data-types be created to describe the instrument configuration and components. A very substantial part of the MIFlowCyt element that describes the Experimental Overview part of the MIFlowCyt and substantial parts of several other major elements have been developed. Conclusions: The future use of structured XML tags and web technology should facilitate searching of experimental information, its presentation, and inclusion in structured research, clinical, and regulatory documents, as well as demonstrate in publications adherence to the MIFlowCyt standard. The use of CytometryML together with XML technology should also result in the textual and numeric data being published using web technology without any change in composition. Preliminary testing indicates that CytometryML XML pages can be directly formatted with the combination of HTML

  15. Recursive representation of Wronskians in confluent supersymmetric quantum mechanics

    NASA Astrophysics Data System (ADS)

    Contreras-Astorga, Alonso; Schulze-Halberg, Axel

    2017-03-01

    A recursive form of arbitrary-order Wronskian associated with transformation functions in the confluent algorithm of supersymmetric quantum mechanics (SUSY) is constructed. With this recursive form regularity conditions for the generated potentials can be analyzed. Moreover, as byproducts we obtain new representations of solutions to Schrödinger equations that underwent a confluent SUSY-transformation.

  16. Language, Mind, Practice: Families of Recursive Thinking in Human Reasoning

    ERIC Educational Resources Information Center

    Josephson, Marika

    2011-01-01

    In 2002, Chomsky, Hauser, and Fitch asserted that recursion may be the one aspect of the human language faculty that makes human language unique in the narrow sense--unique to language and unique to human beings. They also argue somewhat more quietly (as do Pinker and Jackendoff 2005) that recursion may be possible outside of language: navigation,…

  17. XML: James Webb Space Telescope Database Issues, Lessons, and Status

    NASA Technical Reports Server (NTRS)

    Detter, Ryan; Mooney, Michael; Fatig, Curtis

    2003-01-01

    This paper will present the current concept using extensible Markup Language (XML) as the underlying structure for the James Webb Space Telescope (JWST) database. The purpose of using XML is to provide a JWST database, independent of any portion of the ground system, yet still compatible with the various systems using a variety of different structures. The testing of the JWST Flight Software (FSW) started in 2002, yet the launch is scheduled for 2011 with a planned 5-year mission and a 5-year follow on option. The initial database and ground system elements, including the commands, telemetry, and ground system tools will be used for 19 years, plus post mission activities. During the Integration and Test (I&T) phases of the JWST development, 24 distinct laboratories, each geographically dispersed, will have local database tools with an XML database. Each of these laboratories database tools will be used for the exporting and importing of data both locally and to a central database system, inputting data to the database certification process, and providing various reports. A centralized certified database repository will be maintained by the Space Telescope Science Institute (STScI), in Baltimore, Maryland, USA. One of the challenges for the database is to be flexible enough to allow for the upgrade, addition or changing of individual items without effecting the entire ground system. Also, using XML should allow for the altering of the import and export formats needed by the various elements, tracking the verification/validation of each database item, allow many organizations to provide database inputs, and the merging of the many existing database processes into one central database structure throughout the JWST program. Many National Aeronautics and Space Administration (NASA) projects have attempted to take advantage of open source and commercial technology. Often this causes a greater reliance on the use of Commercial-Off-The-Shelf (COTS), which is often limiting

  18. Recursive Partitioning Method on Competing Risk Outcomes

    PubMed Central

    Xu, Wei; Che, Jiahua; Kong, Qin

    2016-01-01

    In some cancer clinical studies, researchers have interests to explore the risk factors associated with competing risk outcomes such as recurrence-free survival. We develop a novel recursive partitioning framework on competing risk data for both prognostic and predictive model constructions. We define specific splitting rules, pruning algorithm, and final tree selection algorithm for the competing risk tree models. This methodology is quite flexible that it can corporate both semiparametric method using Cox proportional hazards model and parametric competing risk model. Both prognostic and predictive tree models are developed to adjust for potential confounding factors. Extensive simulations show that our methods have well-controlled type I error and robust power performance. Finally, we apply both Cox proportional hazards model and flexible parametric model for prognostic tree development on a retrospective clinical study on oropharyngeal cancer patients. PMID:27486300

  19. Syntactic Recursion Facilitates and Working Memory Predicts Recursive Theory of Mind

    PubMed Central

    Arslan, Burcu; Hohenberger, Annette; Verbrugge, Rineke

    2017-01-01

    In this study, we focus on the possible roles of second-order syntactic recursion and working memory in terms of simple and complex span tasks in the development of second-order false belief reasoning. We tested 89 Turkish children in two age groups, one younger (4;6–6;5 years) and one older (6;7–8;10 years). Although second-order syntactic recursion is significantly correlated with the second-order false belief task, results of ordinal logistic regressions revealed that the main predictor of second-order false belief reasoning is complex working memory span. Unlike simple working memory and second-order syntactic recursion tasks, the complex working memory task required processing information serially with additional reasoning demands that require complex working memory strategies. Based on our results, we propose that children’s second-order theory of mind develops when they have efficient reasoning rules to process embedded beliefs serially, thus overcoming a possible serial processing bottleneck. PMID:28072823

  20. T:XML: A Tool Supporting User Interface Model Transformation

    NASA Astrophysics Data System (ADS)

    López-Jaquero, Víctor; Montero, Francisco; González, Pascual

    Model driven development of user interfaces is based on the transformation of an abstract specification into the final user interface the user will interact with. The design of transformation rules to carry out this transformation process is a key issue in any model-driven user interface development approach. In this paper, we introduce T:XML, an integrated development environment for managing, creating and previewing transformation rules. The tool supports the specification of transformation rules by using a graphical notation that works on the basis of the transformation of the input model into a graph-based representation. T:XML allows the design and execution of transformation rules in an integrated development environment. Furthermore, the designer can also preview how the generated user interface looks like after the transformations have been applied. These previewing capabilities can be used to quickly create prototypes to discuss with the users in user-centered design methods.

  1. XML-based product information processing method for product design

    NASA Astrophysics Data System (ADS)

    Zhang, Zhen Yu

    2012-01-01

    Design knowledge of modern mechatronics product is based on information processing as the center of the knowledge-intensive engineering, thus product design innovation is essentially the knowledge and information processing innovation. Analysis of the role of mechatronics product design knowledge and information management features, a unified model of XML-based product information processing method is proposed. Information processing model of product design includes functional knowledge, structural knowledge and their relationships. For the expression of product function element, product structure element, product mapping relationship between function and structure based on the XML model are proposed. The information processing of a parallel friction roller is given as an example, which demonstrates that this method is obviously helpful for knowledge-based design system and product innovation.

  2. XML-based product information processing method for product design

    NASA Astrophysics Data System (ADS)

    Zhang, Zhen Yu

    2011-12-01

    Design knowledge of modern mechatronics product is based on information processing as the center of the knowledge-intensive engineering, thus product design innovation is essentially the knowledge and information processing innovation. Analysis of the role of mechatronics product design knowledge and information management features, a unified model of XML-based product information processing method is proposed. Information processing model of product design includes functional knowledge, structural knowledge and their relationships. For the expression of product function element, product structure element, product mapping relationship between function and structure based on the XML model are proposed. The information processing of a parallel friction roller is given as an example, which demonstrates that this method is obviously helpful for knowledge-based design system and product innovation.

  3. Recursive estimation of prior probabilities using the mixture approach

    NASA Technical Reports Server (NTRS)

    Kazakos, D.

    1974-01-01

    The problem of estimating the prior probabilities q sub k of a mixture of known density functions f sub k(X), based on a sequence of N statistically independent observations is considered. It is shown that for very mild restrictions on f sub k(X), the maximum likelihood estimate of Q is asymptotically efficient. A recursive algorithm for estimating Q is proposed, analyzed, and optimized. For the M = 2 case, it is possible for the recursive algorithm to achieve the same performance with the maximum likelihood one. For M 2, slightly inferior performance is the price for having a recursive algorithm. However, the loss is computable and tolerable.

  4. Enhance Reuse of Standard e-Business XML Schema Documents

    SciTech Connect

    Buhwan, Jeong; Kulvatunyou, Boonserm; Ivezic, Nenad; Jones, Albert

    2005-07-01

    Ideally, e-Business application interfaces would be built from highly reusable specifications of business document standards. Since many of these specifications are poorly understood, users often create new ones or customize existing ones every time a new integration problem arises. Consequently, even though there is a potential for reuse, the lack of a component discovery tool means that the cost of reuse is still prohibitively high. In this paper, we explore the potential of using similarity metrics to discover standard XML Schema documents. Our goal is to enhance reuse of XML Schema document/component standards in new integration contexts through the discovery process. We are motivated by the increasing access to the application interface specifications expressed in the form of XML Schema. These specifications are created to facilitate business documents exchange among software applications. Reuse can reduce both the proliferation of standards and the interoperability costs. To demonstrate these potential benefits, we propose and position our research based on an experimental scenario and a novel evaluation approach to qualify alternative similarity metrics on schema discovery. The edge equality in the evaluation method provides a conservative quality measure. We review a number of fundamental approaches to developing similarity metrics, and we organize these metrics into lexical, structural, and logical categories. For each of the metrics, we discuss its relevance and potential issues in its application to the XML Schema discovery task. We conclude that each of the similarity measures has its own strengths and weaknesses and each is expected to yield different results in different search situations. It is important, in the context of an application of these measures to e-Business standards that a schema discovery engine capable of assigning appropriate weights to different similarity measures be used when the search conditions change. This is a subject of our

  5. A Comparative Analysis of ASCII and XML Logging Systems

    DTIC Science & Technology

    2010-09-01

    4 2.2. EXI vs gzip . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 2.3. EXI vs ASN.1...can now be placed into a database of its own for searching. The meaning / context of a parcel of information is known since XML’s “markup” is self...and technology easier to build and use. XML logging information also becomes extensible, resulting in easier future adaptations. Both of these

  6. XTCE: XML Telemetry and Command Exchange Tutorial, XTCE Version 1

    NASA Technical Reports Server (NTRS)

    Rice, Kevin; Kizzort, Brad

    2008-01-01

    These presentation slides are a tutorial on XML Telemetry and Command Exchange (XTCE). The goal of XTCE is to provide an industry standard mechanism for describing telemetry and command streams (particularly from satellites.) it wiill lower cost and increase validation over traditional formats, and support exchange or native format.XCTE is designed to describe bit streams, that are typical of telemetry and command in the historic space domain.

  7. An XML Driven Graphical User Interface and Application Management Toolkit

    SciTech Connect

    White, Greg R

    2002-01-18

    In the past, the features of a user interface were limited by those available in the existing graphical widgets it used. Now, improvements in processor speed have fostered the emergence of interpreted languages, in which the appropriate method to render a given data object can be loaded at runtime. XML can be used to precisely describe the association of data types with their graphical handling (beans), and Java provides an especially rich environment for programming the graphics. We present a graphical user interface builder based on Java Beans and XML, in which the graphical screens are described textually (in files or a database) in terms of their screen components. Each component may be a simple text read back, or a complex plot. The programming model provides for dynamic data pertaining to a component to be forwarded synchronously or asynchronously, to the appropriate handler, which may be a built-in method, or a complex applet. This work was initially motivated by the need to move the legacy VMS display interface of the SLAC Control Program to another platform while preserving all of its existing functionality. However the model allows us a powerful and generic system for adding new kinds of graphics, such as Matlab, data sources, such as EPICS, middleware, such as AIDA[1], and transport, such as XML and SOAP. The system will also include a management console, which will be able to report on the present usage of the system, for instance who is running it where and connected to which channels.

  8. Rock.XML - Towards a library of rock physics models

    NASA Astrophysics Data System (ADS)

    Jensen, Erling Hugo; Hauge, Ragnar; Ulvmoen, Marit; Johansen, Tor Arne; Drottning, Åsmund

    2016-08-01

    Rock physics modelling provides tools for correlating physical properties of rocks and their constituents to the geophysical observations we measure on a larger scale. Many different theoretical and empirical models exist, to cover the range of different types of rocks. However, upon reviewing these, we see that they are all built around a few main concepts. Based on this observation, we propose a format for digitally storing the specifications for rock physics models which we have named Rock.XML. It does not only contain data about the various constituents, but also the theories and how they are used to combine these building blocks to make a representative model for a particular rock. The format is based on the Extensible Markup Language XML, making it flexible enough to handle complex models as well as scalable towards extending it with new theories and models. This technology has great advantages as far as documenting and exchanging models in an unambiguous way between people and between software. Rock.XML can become a platform for creating a library of rock physics models; making them more accessible to everyone.

  9. On Fusing Recursive Traversals of K-d Trees

    SciTech Connect

    Rajbhandari, Samyam; Kim, Jinsung; Krishnamoorthy, Sriram; Pouchet, Louis-Noel; Rastello, Fabrice; Harrison, Robert J.; Sadayappan, Ponnuswamy

    2016-03-17

    Loop fusion is a key program transformation for data locality optimization that is implemented in production compilers. But optimizing compilers currently cannot exploit fusion opportunities across a set of recursive tree traversal computations with producer-consumer relationships. In this paper, we develop a compile-time approach to dependence characterization and program transformation to enable fusion across recursively specified traversals over k-ary trees. We present the FuseT source-to-source code transformation framework to automatically generate fused composite recursive operators from an input program containing a sequence of primitive recursive operators. We use our framework to implement fused operators for MADNESS, Multiresolution Adaptive Numerical Environment for Scientific Simulation. We show that locality optimization through fusion can offer more than an order of magnitude performance improvement.

  10. A novel extended kernel recursive least squares algorithm.

    PubMed

    Zhu, Pingping; Chen, Badong; Príncipe, José C

    2012-08-01

    In this paper, a novel extended kernel recursive least squares algorithm is proposed combining the kernel recursive least squares algorithm and the Kalman filter or its extensions to estimate or predict signals. Unlike the extended kernel recursive least squares (Ex-KRLS) algorithm proposed by Liu, the state model of our algorithm is still constructed in the original state space and the hidden state is estimated using the Kalman filter. The measurement model used in hidden state estimation is learned by the kernel recursive least squares algorithm (KRLS) in reproducing kernel Hilbert space (RKHS). The novel algorithm has more flexible state and noise models. We apply this algorithm to vehicle tracking and the nonlinear Rayleigh fading channel tracking, and compare the tracking performances with other existing algorithms.

  11. Recursion Operators for CBC system with reductions. Geometric theory

    NASA Astrophysics Data System (ADS)

    Yanovski, A.; Vilasi, G.

    2016-09-01

    We discuss some recent developments of the geometric theory of the Recursion Operators (Generating Operators) for Caudrey-Beals-Coifman systems (CBC systems) on semisimple Lie algebras. As is well known the essence of this interpretation is that the Recursion Operators could be considered as adjoint to Nijenhuis tensors on certain infinite-dimensional manifolds. In particular, we discuss the case when there are Zp reductions of Mikhailov type.

  12. Recursive construction of perfect DNA molecules from imperfect oligonucleotides.

    PubMed

    Linshiz, Gregory; Yehezkel, Tuval Ben; Kaplan, Shai; Gronau, Ilan; Ravid, Sivan; Adar, Rivka; Shapiro, Ehud

    2008-01-01

    Making faultless complex objects from potentially faulty building blocks is a fundamental challenge in computer engineering, nanotechnology and synthetic biology. Here, we show for the first time how recursion can be used to address this challenge and demonstrate a recursive procedure that constructs error-free DNA molecules and their libraries from error-prone oligonucleotides. Divide and Conquer (D&C), the quintessential recursive problem-solving technique, is applied in silico to divide the target DNA sequence into overlapping oligonucleotides short enough to be synthesized directly, albeit with errors; error-prone oligonucleotides are recursively combined in vitro, forming error-prone DNA molecules; error-free fragments of these molecules are then identified, extracted and used as new, typically longer and more accurate, inputs to another iteration of the recursive construction procedure; the entire process repeats until an error-free target molecule is formed. Our recursive construction procedure surpasses existing methods for de novo DNA synthesis in speed, precision, amenability to automation, ease of combining synthetic and natural DNA fragments, and ability to construct designer DNA libraries. It thus provides a novel and robust foundation for the design and construction of synthetic biological molecules and organisms.

  13. Recursive Mahalanobis separability measure for gene subset selection.

    PubMed

    Mao, K Z; Tang, Wenyin

    2011-01-01

    Mahalanobis class separability measure provides an effective evaluation of the discriminative power of a feature subset, and is widely used in feature selection. However, this measure is computationally intensive or even prohibitive when it is applied to gene expression data. In this study, a recursive approach to Mahalanobis measure evaluation is proposed, with the goal of reducing computational overhead. Instead of evaluating Mahalanobis measure directly in high-dimensional space, the recursive approach evaluates the measure through successive evaluations in 2D space. Because of its recursive nature, this approach is extremely efficient when it is combined with a forward search procedure. In addition, it is noted that gene subsets selected by Mahalanobis measure tend to overfit training data and generalize unsatisfactorily on unseen test data, due to small sample size in gene expression problems. To alleviate the overfitting problem, a regularized recursive Mahalanobis measure is proposed in this study, and guidelines on determination of regularization parameters are provided. Experimental studies on five gene expression problems show that the regularized recursive Mahalanobis measure substantially outperforms the nonregularized Mahalanobis measures and the benchmark recursive feature elimination (RFE) algorithm in all five problems.

  14. The XSD-Builder Specification Language—Toward a Semantic View of XML Schema Definition

    NASA Astrophysics Data System (ADS)

    Fong, Joseph; Cheung, San Kuen

    In the present database market, XML database model is a main structure for the forthcoming database system in the Internet environment. As a conceptual schema of XML database, XML Model has its limitation on presenting its data semantics. System analyst has no toolset for modeling and analyzing XML system. We apply XML Tree Model (shown in Figure 2) as a conceptual schema of XML database to model and analyze the structure of an XML database. It is important not only for visualizing, specifying, and documenting structural models, but also for constructing executable systems. The tree model represents inter-relationship among elements inside different logical schema such as XML Schema Definition (XSD), DTD, Schematron, XDR, SOX, and DSD (shown in Figure 1, an explanation of the terms in the figure are shown in Table 1). The XSD-Builder consists of XML Tree Model, source language, translator, and XSD. The source language is called XSD-Source which is mainly for providing an environment with concept of user friendliness while writing an XSD. The source language will consequently be translated by XSD-Translator. Output of XSD-Translator is an XSD which is our target and is called as an object language.

  15. XML — an opportunity for data standards in the geosciences

    NASA Astrophysics Data System (ADS)

    Houlding, Simon W.

    2001-08-01

    Extensible markup language (XML) is a recently introduced meta-language standard on the Web. It provides the rules for development of metadata (markup) standards for information transfer in specific fields. XML allows development of markup languages that describe what information is rather than how it should be presented. This allows computer applications to process the information in intelligent ways. In contrast hypertext markup language (HTML), which fuelled the initial growth of the Web, is a metadata standard concerned exclusively with presentation of information. Besides its potential for revolutionizing Web activities, XML provides an opportunity for development of meaningful data standards in specific application fields. The rapid endorsement of XML by science, industry and e-commerce has already spawned new metadata standards in such fields as mathematics, chemistry, astronomy, multi-media and Web micro-payments. Development of XML-based data standards in the geosciences would significantly reduce the effort currently wasted on manipulating and reformatting data between different computer platforms and applications and would ensure compatibility with the new generation of Web browsers. This paper explores the evolution, benefits and status of XML and related standards in the more general context of Web activities and uses this as a platform for discussion of its potential for development of data standards in the geosciences. Some of the advantages of XML are illustrated by a simple, browser-compatible demonstration of XML functionality applied to a borehole log dataset. The XML dataset and the associated stylesheet and schema declarations are available for FTP download.

  16. Recursive frame integration of limited data: RAFAIL

    NASA Astrophysics Data System (ADS)

    Rafailov, Michael K.; Soli, Robert A.

    2005-08-01

    Real time infrared imaging and tracking usually requires a high probability of target detection along with a low false alarm rate, achievable only with a high "Signal-to-Noise Ratio" (SNR). Frame integration--summing of non-correlated frames--is commonly used to improve the SNR. But conventional frame integration requires significant processing to store full frames and integrate intermediate results, normalize frame data, etc. It may drive acquisition of highly specialized hardware, faster processors, dedicated frame integration circuit cards and extra memory cards. Non-stationary noise, low frequency noise correlation, non-ergodic noise, scene dynamics, or pointing accuracy may also limit performance. Recursive frame integration of limited data--RAFAIL, is proposed as a means to improve frame integration performance and mitigate the issues. The technique applies two thresholds--one tuned for optimum probability of detection, the other to manage required false alarm rate--and allows a non-linear integration process that, along with optimal noise management, provides system designers more capability where cost, weight, or power considerations limit system data rate, processing, or memory capability.

  17. Recursive stochastic effects in valley hybrid inflation

    NASA Astrophysics Data System (ADS)

    Levasseur, Laurence Perreault; Vennin, Vincent; Brandenberger, Robert

    2013-10-01

    Hybrid inflation is a two-field model where inflation ends because of a tachyonic instability, the duration of which is determined by stochastic effects and has important observational implications. Making use of the recursive approach to the stochastic formalism presented in [L. P. Levasseur, preceding article, Phys. Rev. D 88, 083537 (2013)], these effects are consistently computed. Through an analysis of backreaction, this method is shown to converge in the valley but points toward an (expected) instability in the waterfall. It is further shown that the quasistationarity of the auxiliary field distribution breaks down in the case of a short-lived waterfall. We find that the typical dispersion of the waterfall field at the critical point is then diminished, thus increasing the duration of the waterfall phase and jeopardizing the possibility of a short transition. Finally, we find that stochastic effects worsen the blue tilt of the curvature perturbations by an O(1) factor when compared with the usual slow-roll contribution.

  18. Phase Response Design of Recursive All-Pass Digital Filters Using a Modified PSO Algorithm.

    PubMed

    Chang, Wei-Der

    2015-01-01

    This paper develops a new design scheme for the phase response of an all-pass recursive digital filter. A variant of particle swarm optimization (PSO) algorithm will be utilized for solving this kind of filter design problem. It is here called the modified PSO (MPSO) algorithm in which another adjusting factor is more introduced in the velocity updating formula of the algorithm in order to improve the searching ability. In the proposed method, all of the designed filter coefficients are firstly collected to be a parameter vector and this vector is regarded as a particle of the algorithm. The MPSO with a modified velocity formula will force all particles into moving toward the optimal or near optimal solution by minimizing some defined objective function of the optimization problem. To show the effectiveness of the proposed method, two different kinds of linear phase response design examples are illustrated and the general PSO algorithm is compared as well. The obtained results show that the MPSO is superior to the general PSO for the phase response design of digital recursive all-pass filter.

  19. Phase Response Design of Recursive All-Pass Digital Filters Using a Modified PSO Algorithm

    PubMed Central

    Chang, Wei-Der

    2015-01-01

    This paper develops a new design scheme for the phase response of an all-pass recursive digital filter. A variant of particle swarm optimization (PSO) algorithm will be utilized for solving this kind of filter design problem. It is here called the modified PSO (MPSO) algorithm in which another adjusting factor is more introduced in the velocity updating formula of the algorithm in order to improve the searching ability. In the proposed method, all of the designed filter coefficients are firstly collected to be a parameter vector and this vector is regarded as a particle of the algorithm. The MPSO with a modified velocity formula will force all particles into moving toward the optimal or near optimal solution by minimizing some defined objective function of the optimization problem. To show the effectiveness of the proposed method, two different kinds of linear phase response design examples are illustrated and the general PSO algorithm is compared as well. The obtained results show that the MPSO is superior to the general PSO for the phase response design of digital recursive all-pass filter. PMID:26366168

  20. A recursive delayed output-feedback control to stabilize chaotic systems using linear-in-parameter neural networks

    NASA Astrophysics Data System (ADS)

    Yadmellat, Peyman; Nikravesh, S. Kamaleddin Yadavar

    2011-01-01

    In this paper, a recursive delayed output-feedback control strategy is considered for stabilizing unstable periodic orbit of unknown nonlinear chaotic systems. An unknown nonlinearity is directly estimated by a linear-in-parameter neural network which is then used in an observer structure. An on-line modified back propagation algorithm with e-modification is used to update the weights of the network. The globally uniformly ultimately boundedness of overall closed-loop system response is analytically ensured using Razumikhin lemma. To verify the effectiveness of the proposed observer-based controller, a set of simulations is performed on a Rossler system in comparison with several previous methods.

  1. The latest MML (Medical Markup Language) version 2.3--XML-based standard for medical data exchange/storage.

    PubMed

    Guo, Jinqiu; Araki, Kenji; Tanaka, Koji; Sato, Junzo; Suzuki, Muneou; Takada, Akira; Suzuki, Toshiaki; Nakashima, Yusei; Yoshihara, Hiroyuki

    2003-08-01

    As a set of standards, Medical Markup Language (MML) has been developed over the last 8 years to allow the exchange of medical data between different medical information providers MML version 2.21 was characterized by XML as metalanguage and was announced in 1999, at which time full-scale implementation tests were carried out; subsequently, various information and functional inadequacies were discovered in this version. MML was therefore updated to version 2.3 in 2001. At present, MML contains 12 MML modules including the new referral, test result, and report modules. In version 2.3, the group ID element was added; the access right definition and health insurance module were amended.

  2. Recursion to food plants by free-ranging Bornean elephant

    PubMed Central

    Gillespie, Graeme; Goossens, Benoit; Ismail, Sulaiman; Ancrenaz, Marc; Linklater, Wayne

    2015-01-01

    Plant recovery rates after herbivory are thought to be a key factor driving recursion by herbivores to sites and plants to optimise resource-use but have not been investigated as an explanation for recursion in large herbivores. We investigated the relationship between plant recovery and recursion by elephants (Elephas maximus borneensis) in the Lower Kinabatangan Wildlife Sanctuary, Sabah. We identified 182 recently eaten food plants, from 30 species, along 14 × 50 m transects and measured their recovery growth each month over nine months or until they were re-browsed by elephants. The monthly growth in leaf and branch or shoot length for each plant was used to calculate the time required (months) for each species to recover to its pre-eaten length. Elephant returned to all but two transects with 10 eaten plants, a further 26 plants died leaving 146 plants that could be re-eaten. Recursion occurred to 58% of all plants and 12 of the 30 species. Seventy-seven percent of the re-eaten plants were grasses. Recovery times to all plants varied from two to twenty months depending on the species. Recursion to all grasses coincided with plant recovery whereas recursion to most browsed plants occurred four to twelve months before they had recovered to their previous length. The small sample size of many browsed plants that received recursion and uneven plant species distribution across transects limits our ability to generalise for most browsed species but a prominent pattern in plant-scale recursion did emerge. Plant recovery time was a good predictor of time to recursion but varied as a function of growth form (grass, ginger, palm, liana and woody) and differences between sites. Time to plant recursion coincided with plant recovery time for the elephant’s preferred food, grasses, and perhaps also gingers, but not the other browsed species. Elephants are bulk feeders so it is likely that they time their returns to bulk feed on these grass species when quantities have

  3. S-TSP: a novel routing algorithm for In-network processing of recursive computation in wireless sensor networks

    NASA Astrophysics Data System (ADS)

    Tang, Tingfang; Guo, Peng; Liu, Xuefeng

    2016-10-01

    In-network processing is an efficient way to reduce the transmission cost in wireless sensor networks (WSNs). The in-network processing of many domain-specific computation tasks in WSNs usually requires to losslessly distribute the computation of the tasks into the sensor nodes, which is however usually not easy. In this paper we are concerned with such kind of tasks whose computation can only be partitioned into recursive computation mode. To distribute the recursive computations into WSNs, it is required to design an appropriate single in-network processing path, along which the intermediate data is forwarded and updated in the WSNs. We address the recursive computation with constant size of computation result, e.g., distributed least square estimation (D-LSE). Finding the optimal in-network processing path to minimize the total transmission cost in WSNs, is a new problem and seldom studied before. To solve it, we propose a novel routing algorithm called as S-TSP, and compare it with some other greedy algorithms. Extensive simulations are conducted, and the results show the good performance of the proposed S-TSP algorithm.

  4. An XML-based protocol for distributed event services

    SciTech Connect

    Gunter, Dan K.; Smith, Warren; Quesnel, Darcy

    2001-06-25

    A recent trend in distributed computing is the construction of high-performance distributed systems called computational grids. One difficulty we have encountered is that there is no standard format for the representation of performance information and no standard protocol for transmitting this information. This limits the types of performance analysis that can be undertaken in complex distributed systems. To address this problem, we present an XML-based protocol for transmitting performance events in distributed systems and evaluate the performance of this protocol.

  5. An XML-Based Protocol for Distributed Event Services

    NASA Technical Reports Server (NTRS)

    Smith, Warren; Gunter, Dan; Quesnel, Darcy; Biegel, Bryan (Technical Monitor)

    2001-01-01

    A recent trend in distributed computing is the construction of high-performance distributed systems called computational grids. One difficulty we have encountered is that there is no standard format for the representation of performance information and no standard protocol for transmitting this information. This limits the types of performance analysis that can be undertaken in complex distributed systems. To address this problem, we present an XML-based protocol for transmitting performance events in distributed systems and evaluate the performance of this protocol.

  6. Research on Heterogeneous Data Exchange based on XML

    NASA Astrophysics Data System (ADS)

    Li, Huanqin; Liu, Jinfeng

    Integration of multiple data sources is becoming increasingly important for enterprises that cooperate closely with their partners for e-commerce. OLAP enables analysts and decision makers fast access to various materialized views from data warehouses. However, many corporations have internal business applications deployed on different platforms. This paper introduces a model for heterogeneous data exchange based on XML. The system can exchange and share the data among the different sources. The method used to realize the heterogeneous data exchange is given in this paper.

  7. 77 FR 46986 - Revisions to Electric Quarterly Report Filing Process; Availability of Draft XML Schema

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-08-07

    ... Rulemaking. SUMMARY: The Federal Energy Regulatory Commission is making available on its Web site ( http://www.ferc.gov ) a draft of the XML schema that is being developed for use in filing Electric Quarterly... Commission is making available on its Web site ( http://www.ferc.gov ) a draft of the XML schema that...

  8. Efficient XML Interchange (EXI) Compression and Performance Benefits: Development, Implementation and Evaluation

    DTIC Science & Technology

    2010-03-01

    the Noam Chomsky Web site: http://www.chomsky.info/articles/195609--.pdf. Cisco, Inc. (2003). XML processing in the network Cisco position paper to...181 a. Information Grammar Theory ( Chomsky ) ..........................181 b. Events...document. a. Information Grammar Theory ( Chomsky ) Both grammars and events are learned for each XML document by means of a supporting schema or by

  9. Integrated Syntactic/Semantic XML Data Validation with a Reusable Software Component

    ERIC Educational Resources Information Center

    Golikov, Steven

    2013-01-01

    Data integration is a critical component of enterprise system integration, and XML data validation is the foundation for sound data integration of XML-based information systems. Since B2B e-commerce relies on data validation as one of the critical components for enterprise integration, it is imperative for financial industries and e-commerce…

  10. A Practical Introduction to the XML, Extensible Markup Language, by Way of Some Useful Examples

    ERIC Educational Resources Information Center

    Snyder, Robin

    2004-01-01

    XML, Extensible Markup Language, is important as a way to represent and encapsulate the structure of underlying data in a portable way that supports data exchange regardless of the physical storage of the data. This paper (and session) introduces some useful and practical aspects of XML technology for sharing information in a educational setting…

  11. Towards rigorous analysis of the Levitov-Mirlin-Evers recursion

    NASA Astrophysics Data System (ADS)

    Fyodorov, Y. V.; Kupiainen, A.; Webb, C.

    2016-12-01

    This paper aims to develop a rigorous asymptotic analysis of an approximate renormalization group recursion for inverse participation ratios P q of critical powerlaw random band matrices. The recursion goes back to the work by Mirlin and Evers (2000 Phys. Rev. B 62 7920) and earlier works by Levitov (1990 Phys. Rev. Lett. 64 547, 1999 Ann. Phys. 8 697-706) and is aimed to describe the ensuing multifractality of the eigenvectors of such matrices. We point out both similarities and dissimilarities between the LME recursion and those appearing in the theory of multiplicative cascades and branching random walks and show that the methods developed in those fields can be adapted to the present case. In particular the LME recursion is shown to exhibit a phase transition, which we expect is a freezing transition, where the role of temperature is played by the exponent q. However, the LME recursion has features that make its rigorous analysis considerably harder and we point out several open problems for further study.

  12. The redundancy of recursion and infinity for natural language.

    PubMed

    Luuk, Erkki; Luuk, Hendrik

    2011-02-01

    An influential line of thought claims that natural language and arithmetic processing require recursion, a putative hallmark of human cognitive processing (Chomsky in Evolution of human language: biolinguistic perspectives. Cambridge University Press, Cambridge, pp 45-61, 2010; Fitch et al. in Cognition 97(2):179-210, 2005; Hauser et al. in Science 298(5598):1569-1579, 2002). First, we question the need for recursion in human cognitive processing by arguing that a generally simpler and less resource demanding process--iteration--is sufficient to account for human natural language and arithmetic performance. We argue that the only motivation for recursion, the infinity in natural language and arithmetic competence, is equally approachable by iteration and recursion. Second, we submit that the infinity in natural language and arithmetic competence reduces to imagining infinite embedding or concatenation, which is completely independent from the ability to implement infinite processing, and thus, independent from both recursion and iteration. Furthermore, we claim that a property of natural language is physically uncountable finity and not discrete infinity.

  13. Knot Invariants from Topological Recursion on Augmentation Varieties

    NASA Astrophysics Data System (ADS)

    Gu, Jie; Jockers, Hans; Klemm, Albrecht; Soroush, Masoud

    2015-06-01

    Using the duality between Wilson loop expectation values of SU( N) Chern-Simons theory on S 3 and topological open-string amplitudes on the local mirror of the resolved conifold, we study knots on S 3 and their invariants encoded in colored HOMFLY polynomials by means of topological recursion. In the context of the local mirror Calabi-Yau threefold of the resolved conifold, we generalize the topological recursion of the remodelled B-model in order to study branes beyond the class of toric Harvey-Lawson special Lagrangians—as required for analyzing non-trivial knots on S 3. The basic ingredients for the proposed recursion are the spectral curve, given by the augmentation variety of the knot, and the calibrated annulus kernel, encoding the topological annulus amplitudes associated to the knot. We present an explicit construction of the calibrated annulus kernel for torus knots and demonstrate the validity of the topological recursion. We further argue that—if an explicit form of the calibrated annulus kernel is provided for any other knot—the proposed topological recursion should still be applicable. We study the implications of our proposal for knot theory, which exhibit interesting consequences for colored HOMFLY polynomials of mutant knots.

  14. XML schemas and mark-up practices of taxonomic literature

    PubMed Central

    Penev, Lyubomir; Lyal, Christopher HC; Weitzman, Anna; Morse, David R.; King, David; Sautter, Guido; Georgiev, Teodor; Morris, Robert A.; Catapano, Terry; Agosti, Donat

    2011-01-01

    Abstract We review the three most widely used XML schemas used to mark-up taxonomic texts, TaxonX, TaxPub and taXMLit. These are described from the viewpoint of their development history, current status, implementation, and use cases. The concept of “taxon treatment” from the viewpoint of taxonomy mark-up into XML is discussed. TaxonX and taXMLit are primarily designed for legacy literature, the former being more lightweight and with a focus on recovery of taxon treatments, the latter providing a much more detailed set of tags to facilitate data extraction and analysis. TaxPub is an extension of the National Library of Medicine Document Type Definition (NLM DTD) for taxonomy focussed on layout and recovery and, as such, is best suited for mark-up of new publications and their archiving in PubMedCentral. All three schemas have their advantages and shortcomings and can be used for different purposes. PMID:22207808

  15. XML-Based Visual Specification of Multidisciplinary Applications

    NASA Technical Reports Server (NTRS)

    Al-Theneyan, Ahmed; Jakatdar, Amol; Mehrotra, Piyush; Zubair, Mohammad

    2001-01-01

    The advancements in the Internet and Web technologies have fueled a growing interest in developing a web-based distributed computing environment. We have designed and developed Arcade, a web-based environment for designing, executing, monitoring, and controlling distributed heterogeneous applications, which is easy to use and access, portable, and provides support through all phases of the application development and execution. A major focus of the environment is the specification of heterogeneous, multidisciplinary applications. In this paper we focus on the visual and script-based specification interface of Arcade. The web/browser-based visual interface is designed to be intuitive to use and can also be used for visual monitoring during execution. The script specification is based on XML to: (1) make it portable across different frameworks, and (2) make the development of our tools easier by using the existing freely available XML parsers and editors. There is a one-to-one correspondence between the visual and script-based interfaces allowing users to go back and forth between the two. To support this we have developed translators that translate a script-based specification to a visual-based specification, and vice-versa. These translators are integrated with our tools and are transparent to users.

  16. Towards health care process description framework: an XML DTD design.

    PubMed

    Staccini, P; Joubert, M; Quaranta, J F; Aymard, S; Fieschi, D; Fieschi, M

    2001-01-01

    The development of health care and hospital information systems has to meet users needs as well as requirements such as the tracking of all care activities and the support of quality improvement. The use of process-oriented analysis is of-value to provide analysts with: (i) a systematic description of activities; (ii) the elicitation of the useful data to perform and record care tasks; (iii) the selection of relevant decision-making support. But paper-based tools are not a very suitable way to manage and share the documentation produced during this step. The purpose of this work is to propose a method to implement the results of process analysis according to XML techniques (eXtensible Markup Language). It is based on the IDEF0 activity modeling language (Integration DEfinition for Function modeling). A hierarchical description of a process and its components has been defined through a flat XML file with a grammar of proper metadata tags. Perspectives of this method are discussed.

  17. The recursive deterministic perceptron neural network.

    PubMed

    Tajine, Mohamed; Elizondo, David

    1998-12-01

    We introduce a feedforward multilayer neural network which is a generalization of the single layer perceptron topology (SLPT), called recursive deterministic perceptron (RDP). This new model is capable of solving any two-class classification problem, as opposed to the single layer perceptron which can only solve classification problems dealing with linearly separable sets (two subsets X and Y of R(d) are said to be linearly separable if there exists a hyperplane such that the elements of X and Y lie on the two opposite sides of R(d) delimited by this hyperplane). We propose several growing methods for constructing a RDP. These growing methods build a RDP by successively adding intermediate neurons (IN) to the topology (an IN corresponds to a SLPT). Thus, as a result, we obtain a multilayer perceptron topology, which together with the weights, are determined automatically by the constructing algorithms. Each IN augments the affine dimension of the set of input vectors. This augmentation is done by adding the output of each of these INs, as a new component, to every input vector. The construction of a new IN is made by selecting a subset from the set of augmented input vectors which is LS from the rest of this set. This process ends with LS classes in almost n-1 steps where n is the number of input vectors. For this construction, if we assume that the selected LS subsets are of maximum cardinality, the problem is proven to be NP-complete. We also introduce a generalization of the RDP model for classification of m classes (m>2) allowing to always separate m classes. This generalization is based on a new notion of linear separability for m classes, and it follows naturally from the RDP. This new model can be used to compute functions with a finite domain, and thus, to approximate continuous functions. We have also compared - over several classification problems - the percentage of test data correctly classified, or the topology of the 2 and m classes RDPs with that of

  18. Parameter Uncertainty for Aircraft Aerodynamic Modeling using Recursive Least Squares

    NASA Technical Reports Server (NTRS)

    Grauer, Jared A.; Morelli, Eugene A.

    2016-01-01

    A real-time method was demonstrated for determining accurate uncertainty levels of stability and control derivatives estimated using recursive least squares and time-domain data. The method uses a recursive formulation of the residual autocorrelation to account for colored residuals, which are routinely encountered in aircraft parameter estimation and change the predicted uncertainties. Simulation data and flight test data for a subscale jet transport aircraft were used to demonstrate the approach. Results showed that the corrected uncertainties matched the observed scatter in the parameter estimates, and did so more accurately than conventional uncertainty estimates that assume white residuals. Only small differences were observed between batch estimates and recursive estimates at the end of the maneuver. It was also demonstrated that the autocorrelation could be reduced to a small number of lags to minimize computation and memory storage requirements without significantly degrading the accuracy of predicted uncertainty levels.

  19. Haydock's recursive solution of self-adjoint problems. Discrete spectrum

    NASA Astrophysics Data System (ADS)

    Moroz, Alexander

    2014-12-01

    Haydock's recursive solution is shown to underline a number of different concepts such as (i) quasi-exactly solvable models, (ii) exactly solvable models, (iii) three-term recurrence solutions based on Schweber's quantization criterion in Hilbert spaces of entire analytic functions, and (iv) a discrete quantum mechanics of Odake and Sasaki. A recurrent theme of Haydock's recursive solution is that the spectral properties of any self-adjoint problem can be mapped onto a corresponding sequence of polynomials {pn(E) } in energy variable E. The polynomials {pn(E) } are orthonormal with respect to the density of states n0(E) and energy eigenstate | E > is the generating function of {pn(E) } . The generality of Haydock's recursive solution enables one to see the different concepts from a unified perspective and mutually benefiting from each other. Some results obtained within the particular framework of any of (i) to (iv) may have much broader significance.

  20. Elucidating the stop bands of structurally colored systems through recursion

    NASA Astrophysics Data System (ADS)

    Amir, Ariel; Vukusic, Peter

    2013-04-01

    Interference is the source of some of the spectacular colors of animals and plants in nature. In some of these systems, the physical structure consists of an ordered array of layers with alternating high and low refractive indices. This periodicity leads to an optical band structure that is analogous to the electronic band structure encountered in semiconductor physics: specific bands of wavelengths (the stop bands) are perfectly reflected. Here, we present a minimal model for optical band structure in a periodic multilayer structure and solve it using recursion relations. The stop bands emerge in the limit of an infinite number of layers by finding the fixed point of the recursion. We compare to experimental data for various beetles, whose optical structure resembles the proposed model. Thus, using only the phenomenon of interference and the idea of recursion, we are able to elucidate the concept of band structure in the context of the experimentally observed high reflectance and iridescent appearance of structurally colored beetles.

  1. Well-poised generation of Apery-like recursions

    NASA Astrophysics Data System (ADS)

    Zudilin, Wadim

    2005-06-01

    The idea to use classical hypergeometric series and, in particular, well-poised hypergeometric series in diophantine problems of the values of the polylogarithms has led to several novelties in number theory and neighbouring areas of mathematics. Here, we present a systematic approach to derive second-order polynomial recursions for approximations to some values of the Lerch zeta function, depending on the fixed (but not necessarily real) parameter [alpha] satisfying the condition Re([alpha]). Substituting [alpha]=0 into the resulting recurrence equations produces the famous recursions for rational approximations to [zeta](2), [zeta](3) due to Apery, as well as the known recursion for rational approximations to [zeta](4). Multiple integral representations for solutions of the constructed recurrences are also given.

  2. Theory and applications of sampled analog recursive CTD comb filters

    NASA Technical Reports Server (NTRS)

    Tao, T. F.; Holmes, S. V.; Ejaz, A.; Piazza, F.; Saetre, L. T.; Freund, B.

    1976-01-01

    Sampled analog comb filter using a recursive filter implementation was studied. Charge transfer device delay lines were used as the delay elements. The similarities and differences between sampled analog recursive filter and digital recursive filter were pointed out. Both canceller type and integrator type comb filter using the first order or the second order canonical circuits were studied. Experimental results of frequency response are presented to show their difference from that of digital filters. A theoretical analysis was presented based essentially on the digital filter theory but modified to take into account the differences described above. The usefulness of this theory in analyzing the comb filter characteristics was discussed. Several applications of both the canceller type comb filter and the integrator type comb filter were demonstrated.

  3. The limits on combining recursive horn rules with description logics

    SciTech Connect

    Levy, A.Y.; Rousset, M.C.

    1996-12-31

    Horn rule languages have formed the basis for many Artificial Intelligence application languages, but are not expressive enough to model domains with a rich hierarchical structure. Description logics have been designed especially to model rich hierarchies. Several applications would significantly benefit from combining the expressive power of both formalisms. This paper focuses on combining recursive function-free Horn rules with the expressive description logic ALCNR, and shows exactly when a hybrid language with decidable inference can be obtained. First, we show that several of the core constructors of description logics lead by themselves to undecidability of inference when combined with recursive function-free Horn rules. We then show that without these constructors we obtain a maximal subset of ALCNRR that yields a decidable hybrid language. Finally, we describe a restriction on the Horn rules that guarantees decidable inference when combined with all of ALCNR, and covers many of the common usages of recursive rules.

  4. Recursive Heaviside step functions and beginning of the universe

    NASA Astrophysics Data System (ADS)

    Shin, Changsoo; Kim, Seongjai

    2017-04-01

    This article introduces recursive Heaviside step functions, as a potential of the known universe, for the first time in the history of mathematics, science, and engineering. In modern cosmology, various bouncing models have been suggested based on the postulation that the current universe is the result of the collapse of a previous universe. However, all Big Bounce models leave unanswered the question of what powered inflation. Recursive Heaviside step functions are analyzed to represent the warpage of spacetime during the crunch-bounce transition. In particular, the time shift appeared during the transition is modeled in the form of recursive Heaviside step functions and suggested as a possible answer for the immeasurable energy appeared for the Big Bounce.

  5. Detection of small target using recursive higher order statistics

    NASA Astrophysics Data System (ADS)

    Hou, Wang; Sun, Hongyuan; Lei, Zhihui

    2014-02-01

    In this paper, a recursive higher order statistics algorithm is proposed for small target detection in temporal domain. Firstly, the background of image sequence is normalized. Then, the higher order statistics are recursively solved in image sequence to obtain the feature image. Finally, the feature image is segmented with threshold to detect the small target. To validate the algorithm proposed in this paper, five simulated and one semi-simulation image sequences are created. The ROC curves are employed for evaluation of experimental results. Experiment results show that our method is very effective for small target detection.

  6. A decoupled recursive approach for constrained flexible multibody system dynamics

    NASA Technical Reports Server (NTRS)

    Lai, Hao-Jan; Kim, Sung-Soo; Haug, Edward J.; Bae, Dae-Sung

    1989-01-01

    A variational-vector calculus approach is employed to derive a recursive formulation for dynamic analysis of flexible multibody systems. Kinematic relationships for adjacent flexible bodies are derived in a companion paper, using a state vector notation that represents translational and rotational components simultaneously. Cartesian generalized coordinates are assigned for all body and joint reference frames, to explicitly formulate deformation kinematics under small deformation kinematics and an efficient flexible dynamics recursive algorithm is developed. Dynamic analysis of a closed loop robot is performed to illustrate efficiency of the algorithm.

  7. Note on recursion relations for the Q -cut representation

    NASA Astrophysics Data System (ADS)

    Feng, Bo; He, Song; Huang, Rijun; Luo, Ming-xing

    2017-01-01

    In this note, we study the Q -cut representation by combining it with BCFW deformation. As a consequence, the one-loop integrand is expressed in terms of a recursion relation, i.e., n-point one-loop integrand is constructed using tree-level amplitudes and m-point one-loop integrands with m ≤ n - 1. By giving explicit examples, we show that the integrand from the recursion relation is equivalent to that from Feynman diagrams or the original Q -cut construction, up to scale free terms.

  8. Recursive multibody dynamics and discrete-time optimal control

    NASA Technical Reports Server (NTRS)

    Deleuterio, G. M. T.; Damaren, C. J.

    1989-01-01

    A recursive algorithm is developed for the solution of the simulation dynamics problem for a chain of rigid bodies. Arbitrary joint constraints are permitted, that is, joints may allow translational and/or rotational degrees of freedom. The recursive procedure is shown to be identical to that encountered in a discrete-time optimal control problem. For each relevant quantity in the multibody dynamics problem, there exists an analog in the context of optimal control. The performance index that is minimized in the control problem is identified as Gibbs' function for the chain of bodies.

  9. Measurement updating using the U-D factorization. [for Kalman matrix filter error covariance

    NASA Technical Reports Server (NTRS)

    Bierman, G. J.

    1975-01-01

    A new mechanization of the Kalman updating algorithm based on a U-D factorization of the estimate error covariance is introduced. Efficient and stable updating recursions are developed for the unit upper triangular factor U and the diagonal factor D, treating only the parameter estimation problem. Properties of the factorization update performed here include efficient one point at a time processing that requires little more computation than the optimal but numerically unstable conventional Kalman measurement update algorithm, and stability that compares with the square root filter.

  10. Creating preservation metadata from XML-metadata profiles

    NASA Astrophysics Data System (ADS)

    Ulbricht, Damian; Bertelmann, Roland; Gebauer, Petra; Hasler, Tim; Klump, Jens; Kirchner, Ingo; Peters-Kottig, Wolfgang; Mettig, Nora; Rusch, Beate

    2014-05-01

    Metadata Encoding and Transmission Standard (METS). To find datasets in future portals and to make use of this data in own scientific work, proper selection of discovery metadata and application metadata is very important. Some XML-metadata profiles are not suitable for preservation, because version changes are very fast and make it nearly impossible to automate the migration. For other XML-metadata profiles schema definitions are changed after publication of the profile or the schema definitions become inaccessible, which might cause problems during validation of the metadata inside the preservation system [2]. Some metadata profiles are not used widely enough and might not even exist in the future. Eventually, discovery and application metadata have to be embedded into the mdWrap-subtree of the METS-XML. [1] http://www.archivematica.org [2] http://dx.doi.org/10.2218/ijdc.v7i1.215

  11. Teaching object concepts for XML-based representations.

    SciTech Connect

    Kelsey, R. L.

    2002-01-01

    Students learned about object-oriented design concepts and knowledge representation through the use of a set of toy blocks. The blocks represented a limited and focused domain of knowledge and one that was physical and tangible. The blocks helped the students to better visualize, communicate, and understand the domain of knowledge as well as how to perform object decomposition. The blocks were further abstracted to an engineering design kit for water park design. This helped the students to work on techniques for abstraction and conceptualization. It also led the project from tangible exercises into software and programming exercises. Students employed XML to create object-based knowledge representations and Java to use the represented knowledge. The students developed and implemented software allowing a lay user to design and create their own water slide and then to take a simulated ride on their slide.

  12. Managing and Querying Image Annotation and Markup in XML

    PubMed Central

    Wang, Fusheng; Pan, Tony; Sharma, Ashish; Saltz, Joel

    2010-01-01

    Proprietary approaches for representing annotations and image markup are serious barriers for researchers to share image data and knowledge. The Annotation and Image Markup (AIM) project is developing a standard based information model for image annotation and markup in health care and clinical trial environments. The complex hierarchical structures of AIM data model pose new challenges for managing such data in terms of performance and support of complex queries. In this paper, we present our work on managing AIM data through a native XML approach, and supporting complex image and annotation queries through native extension of XQuery language. Through integration with xService, AIM databases can now be conveniently shared through caGrid. PMID:21218167

  13. Managing and Querying Image Annotation and Markup in XML.

    PubMed

    Wang, Fusheng; Pan, Tony; Sharma, Ashish; Saltz, Joel

    2010-01-01

    Proprietary approaches for representing annotations and image markup are serious barriers for researchers to share image data and knowledge. The Annotation and Image Markup (AIM) project is developing a standard based information model for image annotation and markup in health care and clinical trial environments. The complex hierarchical structures of AIM data model pose new challenges for managing such data in terms of performance and support of complex queries. In this paper, we present our work on managing AIM data through a native XML approach, and supporting complex image and annotation queries through native extension of XQuery language. Through integration with xService, AIM databases can now be conveniently shared through caGrid.

  14. Teaching and Learning Recursive Programming: A Review of the Research Literature

    ERIC Educational Resources Information Center

    McCauley, Renée; Grissom, Scott; Fitzgerald, Sue; Murphy, Laurie

    2015-01-01

    Hundreds of articles have been published on the topics of teaching and learning recursion, yet fewer than 50 of them have published research results. This article surveys the computing education research literature and presents findings on challenges students encounter in learning recursion, mental models students develop as they learn recursion,…

  15. Using XML and Java Technologies for Astronomical Instrument Control

    NASA Technical Reports Server (NTRS)

    Ames, Troy; Case, Lynne; Powers, Edward I. (Technical Monitor)

    2001-01-01

    Traditionally, instrument command and control systems have been highly specialized, consisting mostly of custom code that is difficult to develop, maintain, and extend. Such solutions are initially very costly and are inflexible to subsequent engineering change requests, increasing software maintenance costs. Instrument description is too tightly coupled with details of implementation. NASA Goddard Space Flight Center, under the Instrument Remote Control (IRC) project, is developing a general and highly extensible framework that applies to any kind of instrument that can be controlled by a computer. The software architecture combines the platform independent processing capabilities of Java with the power of the Extensible Markup Language (XML), a human readable and machine understandable way to describe structured data. A key aspect of the object-oriented architecture is that the software is driven by an instrument description, written using the Instrument Markup Language (IML), a dialect of XML. IML is used to describe the command sets and command formats of the instrument, communication mechanisms, format of the data coming from the instrument, and characteristics of the graphical user interface to control and monitor the instrument. The IRC framework allows the users to define a data analysis pipeline which converts data coming out of the instrument. The data can be used in visualizations in order for the user to assess the data in real-time, if necessary. The data analysis pipeline algorithms can be supplied by the user in a variety of forms or programming languages. Although the current integration effort is targeted for the High-resolution Airborne Wideband Camera (HAWC) and the Submillimeter and Far Infrared Experiment (SAFIRE), first-light instruments of the Stratospheric Observatory for Infrared Astronomy (SOFIA), the framework is designed to be generic and extensible so that it can be applied to any instrument. Plans are underway to test the framework

  16. Semantics Boosts Syntax in Artificial Grammar Learning Tasks with Recursion

    ERIC Educational Resources Information Center

    Fedor, Anna; Varga, Mate; Szathmary, Eors

    2012-01-01

    Center-embedded recursion (CER) in natural language is exemplified by sentences such as "The malt that the rat ate lay in the house." Parsing center-embedded structures is in the focus of attention because this could be one of the cognitive capacities that make humans distinct from all other animals. The ability to parse CER is usually…

  17. Recursion and the Competence/Performance Distinction in AGL Tasks

    ERIC Educational Resources Information Center

    Lobina, David J.

    2011-01-01

    The term "recursion" is used in at least four distinct theoretical senses within cognitive science. Some of these senses in turn relate to the different levels of analysis described by David Marr some 20 years ago; namely, the underlying competence capacity (the "computational" level), the performance operations used in real-time processing (the…

  18. Recursive inversion of externally defined linear systems by FIR filters

    NASA Technical Reports Server (NTRS)

    Bach, Ralph E., Jr.; Baram, Yoram

    1989-01-01

    The approximate inversion of an internally unknown linear system, given by its impulse response sequence, by an inverse system having a finite impulse response, is considered. The recursive least-squares procedure is shown to have an exact initialization, based on the triangular Toeplitz structure of the matrix involved. The proposed approach also suggests solutions to the problem of system identification and compensation.

  19. Recursivity: A Working Paper on Rhetoric and "Mnesis"

    ERIC Educational Resources Information Center

    Stormer, Nathan

    2013-01-01

    This essay proposes the genealogical study of remembering and forgetting as recursive rhetorical capacities that enable discourse to place itself in an ever-changing present. "Mnesis" is a meta-concept for the arrangements of remembering and forgetting that enable rhetoric to function. Most of the essay defines the materiality of "mnesis", first…

  20. Exploring the Recursive Nature of Food and Family Communication

    ERIC Educational Resources Information Center

    Manning, Linda D.

    2006-01-01

    Family meals act as a barometer to signify the changing nature of family life. The primary objective of this activity is to allow students to experience the many ways in which a recursive relationship exists between the food families eat and the patterns of communication families enact. Through this activity, students experience how food and…

  1. Recursive confidence band construction for an unknown distribution function.

    PubMed

    Kiatsupaibul, Seksan; Hayter, Anthony J

    2015-01-01

    Given a sample X1,...,Xn of independent observations from an unknown continuous distribution function F, the problem of constructing a confidence band for F is considered, which is a fundamental problem in statistical inference. This confidence band provides simultaneous inferences on all quantiles and also on all of the cumulative probabilities of the distribution, and so they are among the most important inference procedures that address the issue of multiplicity. A fully nonparametric approach is taken where no assumptions are made about the distribution function F. Historical approaches to this problem, such as Kolmogorov's famous () procedure, represent some of the earliest inference methodologies that address the issue of multiplicity. This is because a confidence band at a given confidence level 1-α allows inferences on all of the quantiles of the distribution, and also on all of the cumulative probabilities, at that specified confidence level. In this paper it is shown how recursive methodologies can be employed to construct both one-sided and two-sided confidence bands of various types. The first approach operates by putting bounds on the cumulative probabilities at the data points, and a recursive integration approach is described. The second approach operates by providing bounds on certain specified quantiles of the distribution, and its implementation using recursive summations of multinomial probabilities is described. These recursive methodologies are illustrated with examples, and R code is available for their implementation.

  2. Recursive Vocal Pattern Learning and Generalization in Starlings

    ERIC Educational Resources Information Center

    Bloomfield, Tiffany Corinna

    2012-01-01

    Among known communication systems, human language alone exhibits open-ended productivity of meaning. Interest in the psychological mechanisms supporting this ability, and their evolutionary origins, has resurged following the suggestion that the only uniquely human ability underlying language is a mechanism of recursion. This "Unique…

  3. Recursive Inversion By Finite-Impulse-Response Filters

    NASA Technical Reports Server (NTRS)

    Bach, Ralph E., Jr.; Baram, Yoram

    1991-01-01

    Recursive approximation gives least-squares best fit to exact response. Algorithm yields finite-impulse-response approximation of unknown single-input/single-output, causal, time-invariant, linear, real system, response of which is sequence of impulses. Applicable to such system-inversion problems as suppression of echoes and identification of target from its scatter response to incident impulse.

  4. A new constant memory recursion for hidden Markov models.

    PubMed

    Bartolucci, Francesco; Pandolfi, Silvia

    2014-02-01

    We develop the recursion for hidden Markov (HM) models proposed by Bartolucci and Besag (2002), and we show how it may be used to implement an estimation algorithm for these models that requires an amount of memory not depending on the length of the observed series of data. This recursion allows us to obtain the conditional distribution of the latent state at every occasion, given the previous state and the observed data. With respect to the estimation algorithm based on the well-known Baum-Welch recursions, which requires an amount of memory that increases with the sample size, the proposed algorithm also has the advantage of not requiring dummy renormalizations to avoid numerical problems. Moreover, it directly allows us to perform global decoding of the latent sequence of states, without the need of a Viterbi method and with a consistent reduction of the memory requirement with respect to the latter. The proposed approach is compared, in terms of computing time and memory requirement, with the algorithm based on the Baum-Welch recursions and with the so-called linear memory algorithm of Churbanov and Winters-Hilt. The comparison is also based on a series of simulations involving an HM model for continuous time-series data.

  5. Recursive Frame Analysis: A Practitioner's Tool for Mapping Therapeutic Conversation

    ERIC Educational Resources Information Center

    Keeney, Hillary; Keeney, Bradford; Chenail, Ronald J.

    2012-01-01

    Recursive frame analysis (RFA), both a practical therapeutic tool and an advanced qualitative research method that maps the structure of therapeutic conversation, is introduced with a clinical case vignette. We present and illustrate a means of mapping metaphorical themes that contextualize the performance taking place in the room, recursively…

  6. On recursive least-squares filtering algorithms and implementations. Ph.D. Thesis

    NASA Technical Reports Server (NTRS)

    Hsieh, Shih-Fu

    1990-01-01

    In many real-time signal processing applications, fast and numerically stable algorithms for solving least-squares problems are necessary and important. In particular, under non-stationary conditions, these algorithms must be able to adapt themselves to reflect the changes in the system and take appropriate adjustments to achieve optimum performances. Among existing algorithms, the QR-decomposition (QRD)-based recursive least-squares (RLS) methods have been shown to be useful and effective for adaptive signal processing. In order to increase the speed of processing and achieve high throughput rate, many algorithms are being vectorized and/or pipelined to facilitate high degrees of parallelism. A time-recursive formulation of RLS filtering employing block QRD will be considered first. Several methods, including a new non-continuous windowing scheme based on selectively rejecting contaminated data, were investigated for adaptive processing. Based on systolic triarrays, many other forms of systolic arrays are shown to be capable of implementing different algorithms. Various updating and downdating systolic algorithms and architectures for RLS filtering are examined and compared in details, which include Householder reflector, Gram-Schmidt procedure, and Givens rotation. A unified approach encompassing existing square-root-free algorithms is also proposed. For the sinusoidal spectrum estimation problem, a judicious method of separating the noise from the signal is of great interest. Various truncated QR methods are proposed for this purpose and compared to the truncated SVD method. Computer simulations provided for detailed comparisons show the effectiveness of these methods. This thesis deals with fundamental issues of numerical stability, computational efficiency, adaptivity, and VLSI implementation for the RLS filtering problems. In all, various new and modified algorithms and architectures are proposed and analyzed; the significance of any of the new method depends

  7. A Tool Kit for Implementing XML Schema Naming and Design Rules

    SciTech Connect

    Lubell, Joshua; Kulvatunyou, Boonserm

    2005-11-01

    A tool kit being developed at the National Institute of Standards and Technology (NIST)encodes XML schema Naming and Design Rules in a computer-interpretable fashion,enabling automated rule enforcement and improving schema quality.

  8. XML-based approaches for the integration of heterogeneous bio-molecular data

    PubMed Central

    Mesiti, Marco; Jiménez-Ruiz, Ernesto; Sanz, Ismael; Berlanga-Llavori, Rafael; Perlasca, Paolo; Valentini, Giorgio; Manset, David

    2009-01-01

    Background The today's public database infrastructure spans a very large collection of heterogeneous biological data, opening new opportunities for molecular biology, bio-medical and bioinformatics research, but raising also new problems for their integration and computational processing. Results In this paper we survey the most interesting and novel approaches for the representation, integration and management of different kinds of biological data by exploiting XML and the related recommendations and approaches. Moreover, we present new and interesting cutting edge approaches for the appropriate management of heterogeneous biological data represented through XML. Conclusion XML has succeeded in the integration of heterogeneous biomolecular information, and has established itself as the syntactic glue for biological data sources. Nevertheless, a large variety of XML-based data formats have been proposed, thus resulting in a difficult effective integration of bioinformatics data schemes. The adoption of a few semantic-rich standard formats is urgent to achieve a seamless integration of the current biological resources. PMID:19828083

  9. XML based tools for assessing potential impact of advanced technology space validation

    NASA Technical Reports Server (NTRS)

    Some, Raphael R.; Weisbin, Charles

    2004-01-01

    A hierarchical XML database and related analysis tools are being developed by the New Millennium Program to provide guidance on the relative impact, to future NASA missions, of advanced technologies under consideration for developmental funding.

  10. A novel approach to investigate recursion and iteration in visual hierarchical processing.

    PubMed

    Martins, Maurício Dias; Martins, Isabel Pavão; Fitch, W Tecumseh

    2016-12-01

    We describe a new method to explore recursive cognition in the visual domain. We define recursion as the ability to represent multiple hierarchical levels using the same rule, entailing the ability to generate new levels beyond those previously encountered. With this definition recursion can be distinguished from general hierarchical embedding. To investigate this recursion/hierarchy distinction in the visual domain, we developed two novel methods: The Visual Recursion Task (VRT), in which an inferred rule is used to represent new hierarchical levels, and the Embedded Iteration Task (EIT), in which additional elements are added to an existing hierarchical level. We found that adult humans can represent recursion in the visuo-spatial domain, and that this ability is distinct from both general intelligence and the ability to represent iterative processes embedded within hierarchical structures. Compared with embedded iteration, visual recursion correlated positively with other recursive planning tasks (Tower of Hanoi), but not with specific visuo-spatial resources (spatial short-term memory and working memory). We conclude that humans are able to use recursive representations to process complex visuo-spatial hierarchies and that our visual recursion task taps into specific cognitive resources. This method opens exciting opportunities to explore the relationship between visual recursion and language.

  11. NeXML: rich, extensible, and verifiable representation of comparative data and metadata.

    PubMed

    Vos, Rutger A; Balhoff, James P; Caravas, Jason A; Holder, Mark T; Lapp, Hilmar; Maddison, Wayne P; Midford, Peter E; Priyam, Anurag; Sukumaran, Jeet; Xia, Xuhua; Stoltzfus, Arlin

    2012-07-01

    In scientific research, integration and synthesis require a common understanding of where data come from, how much they can be trusted, and what they may be used for. To make such an understanding computer-accessible requires standards for exchanging richly annotated data. The challenges of conveying reusable data are particularly acute in regard to evolutionary comparative analysis, which comprises an ever-expanding list of data types, methods, research aims, and subdisciplines. To facilitate interoperability in evolutionary comparative analysis, we present NeXML, an XML standard (inspired by the current standard, NEXUS) that supports exchange of richly annotated comparative data. NeXML defines syntax for operational taxonomic units, character-state matrices, and phylogenetic trees and networks. Documents can be validated unambiguously. Importantly, any data element can be annotated, to an arbitrary degree of richness, using a system that is both flexible and rigorous. We describe how the use of NeXML by the TreeBASE and Phenoscape projects satisfies user needs that cannot be satisfied with other available file formats. By relying on XML Schema Definition, the design of NeXML facilitates the development and deployment of software for processing, transforming, and querying documents. The adoption of NeXML for practical use is facilitated by the availability of (1) an online manual with code samples and a reference to all defined elements and attributes, (2) programming toolkits in most of the languages used commonly in evolutionary informatics, and (3) input-output support in several widely used software applications. An active, open, community-based development process enables future revision and expansion of NeXML.

  12. A browser-based tool for conversion between Fortran NAMELIST and XML/HTML

    NASA Astrophysics Data System (ADS)

    Naito, O.

    A browser-based tool for conversion between Fortran NAMELIST and XML/HTML is presented. It runs on an HTML5 compliant browser and generates reusable XML files to aid interoperability. It also provides a graphical interface for editing and annotating variables in NAMELIST, hence serves as a primitive code documentation environment. Although the tool is not comprehensive, it could be viewed as a test bed for integrating legacy codes into modern systems.

  13. NeXML: Rich, Extensible, and Verifiable Representation of Comparative Data and Metadata

    PubMed Central

    Vos, Rutger A.; Balhoff, James P.; Caravas, Jason A.; Holder, Mark T.; Lapp, Hilmar; Maddison, Wayne P.; Midford, Peter E.; Priyam, Anurag; Sukumaran, Jeet; Xia, Xuhua; Stoltzfus, Arlin

    2012-01-01

    In scientific research, integration and synthesis require a common understanding of where data come from, how much they can be trusted, and what they may be used for. To make such an understanding computer-accessible requires standards for exchanging richly annotated data. The challenges of conveying reusable data are particularly acute in regard to evolutionary comparative analysis, which comprises an ever-expanding list of data types, methods, research aims, and subdisciplines. To facilitate interoperability in evolutionary comparative analysis, we present NeXML, an XML standard (inspired by the current standard, NEXUS) that supports exchange of richly annotated comparative data. NeXML defines syntax for operational taxonomic units, character-state matrices, and phylogenetic trees and networks. Documents can be validated unambiguously. Importantly, any data element can be annotated, to an arbitrary degree of richness, using a system that is both flexible and rigorous. We describe how the use of NeXML by the TreeBASE and Phenoscape projects satisfies user needs that cannot be satisfied with other available file formats. By relying on XML Schema Definition, the design of NeXML facilitates the development and deployment of software for processing, transforming, and querying documents. The adoption of NeXML for practical use is facilitated by the availability of (1) an online manual with code samples and a reference to all defined elements and attributes, (2) programming toolkits in most of the languages used commonly in evolutionary informatics, and (3) input–output support in several widely used software applications. An active, open, community-based development process enables future revision and expansion of NeXML. PMID:22357728

  14. The Role of Efficient XML Interchange (EXI) in Navy Wide-Area Network (WAN) Optimization

    DTIC Science & Technology

    2015-03-01

    Compression Ratio Info-Leak Made Easy DOD Department of Defense DSCS Defense Satellite Communications System EFX Efficient XML EMCON Emission...Identifiable Information QoS Quality of Service RTT Round Trip Time SaaS Software as a Service SATCOM Satellite Communications SBIR Small Business...Application Protocol WBXML WAP Binary XML WGS Wideband Global SATCOM; Wideband Gapfiller Satellite WML Wireless Markup Language WOC WAN

  15. Integrated Design and Production Reference Integration with ArchGenXML V1.00

    SciTech Connect

    Barter, R H

    2004-07-20

    ArchGenXML is a tool that allows easy creation of Zope products through the use of Archetypes. The Integrated Design and Production Reference (IDPR) should be highly configurable in order to meet the needs of a diverse engineering community. Ease of configuration is key to the success of IDPR. The purpose of this paper is to describe a method of using a UML diagram editor to configure IDPR through ArchGenXML and Archetypes.

  16. An adaptable XML based approach for scientific data management and integration

    NASA Astrophysics Data System (ADS)

    Wang, Fusheng; Thiel, Florian; Furrer, Daniel; Vergara-Niedermayr, Cristobal; Qin, Chen; Hackenberg, Georg; Bourgue, Pierre-Emmanuel; Kaltschmidt, David; Wang, Mo

    2008-03-01

    Increased complexity of scientific research poses new challenges to scientific data management. Meanwhile, scientific collaboration is becoming increasing important, which relies on integrating and sharing data from distributed institutions. We develop SciPort, a Web-based platform on supporting scientific data management and integration based on a central server based distributed architecture, where researchers can easily collect, publish, and share their complex scientific data across multi-institutions. SciPort provides an XML based general approach to model complex scientific data by representing them as XML documents. The documents capture not only hierarchical structured data, but also images and raw data through references. In addition, SciPort provides an XML based hierarchical organization of the overall data space to make it convenient for quick browsing. To provide generalization, schemas and hierarchies are customizable with XML-based definitions, thus it is possible to quickly adapt the system to different applications. While each institution can manage documents on a Local SciPort Server independently, selected documents can be published to a Central Server to form a global view of shared data across all sites. By storing documents in a native XML database, SciPort provides high schema extensibility and supports comprehensive queries through XQuery. By providing a unified and effective means for data modeling, data access and customization with XML, SciPort provides a flexible and powerful platform for sharing scientific data for scientific research communities, and has been successfully used in both biomedical research and clinical trials.

  17. XML-based information system for planetary sciences

    NASA Astrophysics Data System (ADS)

    Carraro, F.; Fonte, S.; Turrini, D.

    2009-04-01

    EuroPlaNet (EPN in the following) has been developed by the planetological community under the "Sixth Framework Programme" (FP6 in the following), the European programme devoted to the improvement of the European research efforts through the creation of an internal market for science and technology. The goal of the EPN programme is the creation of a European network aimed to the diffusion of data produced by space missions dedicated to the study of the Solar System. A special place within the EPN programme is that of I.D.I.S. (Integrated and Distributed Information Service). The main goal of IDIS is to offer to the planetary science community a user-friendly access to the data and information produced by the various types of research activities, i.e. Earth-based observations, space observations, modeling, theory and laboratory experiments. During the FP6 programme IDIS development consisted in the creation of a series of thematic nodes, each of them specialized in a specific scientific domain, and a technical coordination node. The four thematic nodes are the Atmosphere node, the Plasma node, the Interiors & Surfaces node and the Small Bodies & Dust node. The main task of the nodes have been the building up of selected scientific cases related with the scientific domain of each node. The second work done by EPN nodes have been the creation of a catalogue of resources related to their main scientific theme. Both these efforts have been used as the basis for the development of the main IDIS goal, i.e. the integrated distributed service. An XML-based data model have been developed to describe resources using meta-data and to store the meta-data within an XML-based database called eXist. A search engine has been then developed in order to allow users to search resources within the database. Users can select the resource type and can insert one or more values or can choose a value among those present in a list, depending on selected resource. The system searches for all

  18. Update '98.

    ERIC Educational Resources Information Center

    Mock, Karen R.

    1998-01-01

    Updates cases and issues previously discussed in this regular column on human rights in Canada, including racism and anti-Semitism, laws on hate crimes, hate sites on the World Wide Web, the use of the "free speech" defense by hate groups, and legal challenges to antiracist groups by individuals criticized by them. (DSK)

  19. Recursive three-dimensional model reconstruction based on Kalman filtering.

    PubMed

    Yu, Ying Kin; Wong, Kin Hong; Chang, Michael Ming Yuen

    2005-06-01

    A recursive two-step method to recover structure and motion from image sequences based on Kalman filtering is described in this paper. The algorithm consists of two major steps. The first step is an extended Kalman filter (EKF) for the estimation of the object's pose. The second step is a set of EKFs, one for each model point, for the refinement of the positions of the model features in the three-dimensional (3-D) space. These two steps alternate from frame to frame. The initial model converges to the final structure as the image sequence is scanned sequentially. The performance of the algorithm is demonstrated with both synthetic data and real-world objects. Analytical and empirical comparisons are made among our approach, the interleaved bundle adjustment method, and the Kalman filtering-based recursive algorithm by Azarbayejani and Pentland. Our approach outperformed the other two algorithms in terms of computation speed without loss in the quality of model reconstruction.

  20. Development of a recursion RNG-based turbulence model

    NASA Technical Reports Server (NTRS)

    Zhou, YE; Vahala, George; Thangam, S.

    1993-01-01

    Reynolds stress closure models based on the recursion renormalization group theory are developed for the prediction of turbulent separated flows. The proposed model uses a finite wavenumber truncation scheme to account for the spectral distribution of energy. In particular, the model incorporates effects of both local and nonlocal interactions. The nonlocal interactions are shown to yield a contribution identical to that from the epsilon-renormalization group (RNG), while the local interactions introduce higher order dispersive effects. A formal analysis of the model is presented and its ability to accurately predict separated flows is analyzed from a combined theoretical and computational stand point. Turbulent flow past a backward facing step is chosen as a test case and the results obtained based on detailed computations demonstrate that the proposed recursion -RNG model with finite cut-off wavenumber can yield very good predictions for the backstep problem.

  1. EEG and MEG source localization using recursively applied (RAP) MUSIC

    SciTech Connect

    Mosher, J.C.; Leahy, R.M.

    1996-12-31

    The multiple signal characterization (MUSIC) algorithm locates multiple asynchronous dipolar sources from electroencephalography (EEG) and magnetoencephalography (MEG) data. A signal subspace is estimated from the data, then the algorithm scans a single dipole model through a three-dimensional head volume and computes projections onto this subspace. To locate the sources, the user must search the head volume for local peaks in the projection metric. Here we describe a novel extension of this approach which we refer to as RAP (Recursively APplied) MUSIC. This new procedure automatically extracts the locations of the sources through a recursive use of subspace projections, which uses the metric of principal correlations as a multidimensional form of correlation analysis between the model subspace and the data subspace. The dipolar orientations, a form of `diverse polarization,` are easily extracted using the associated principal vectors.

  2. Recursive utility in a Markov environment with stochastic growth

    PubMed Central

    Hansen, Lars Peter; Scheinkman, José A.

    2012-01-01

    Recursive utility models that feature investor concerns about the intertemporal composition of risk are used extensively in applied research in macroeconomics and asset pricing. These models represent preferences as the solution to a nonlinear forward-looking difference equation with a terminal condition. In this paper we study infinite-horizon specifications of this difference equation in the context of a Markov environment. We establish a connection between the solution to this equation and to an arguably simpler Perron–Frobenius eigenvalue equation of the type that occurs in the study of large deviations for Markov processes. By exploiting this connection, we establish existence and uniqueness results. Moreover, we explore a substantive link between large deviation bounds for tail events for stochastic consumption growth and preferences induced by recursive utility. PMID:22778428

  3. Berends-Giele recursion for double-color-ordered amplitudes

    NASA Astrophysics Data System (ADS)

    Mafra, Carlos R.

    2016-07-01

    Tree-level double-color-ordered amplitudes are computed using Berends-Giele recursion relations applied to the bi-adjoint cubic scalar theory. The standard notion of Berends-Giele currents is generalized to double-currents and their recursions are derived from a perturbiner expansion of linearized fields that solve the non-linear field equations. Two applications are given. Firstly, we prove that the entries of the inverse KLT matrix are equal to Berends-Giele double-currents (and are therefore easy to compute). And secondly, a simple formula to generate tree-level BCJ-satisfying numerators for arbitrary multiplicity is proposed by evaluating the field-theory limit of tree-level string amplitudes for various color orderings using double-color-ordered amplitudes.

  4. Fully flexible unit cell simulation with recursive thermostat chains.

    PubMed

    Choi, Kwangseok; Cho, Maenghyo

    2006-11-14

    The recursive thermostat chained fully flexible cell molecular dynamic simulation (NsigmaT ensemble) is performed. The ensemble is based on the metric tensor, whose components are used as extended variables. These variables are combined with Nosé-Poincaré recursive thermostat chains. This extended Hamiltonian approach preserves Hamiltonian in structure, and the partition function satisfies the NsigmaT ensemble state in phase space. In the present study, the generalized leap frog method was employed for time integration. The resulting molecular dynamics simulation was performed for bulk and thin film solid materials in the face-centered-cubic crystal structure. Uniaxial tension test and simple shear test are performed to predict the behaviors of a solid material in the bulk state and nanoscale thin film state. The proposed flexible cell method should serve as a powerful tool for the prediction of mechanical and thermal properties of solid materials including nanoscale behavior.

  5. A Precision Recursive Estimate for Ephemeris Refinement (PREFER)

    NASA Technical Reports Server (NTRS)

    Gibbs, B.

    1980-01-01

    A recursive filter/smoother orbit determination program was developed to refine the ephemerides produced by a batch orbit determination program (e.g., CELEST, GEODYN). The program PREFER can handle a variety of ground and satellite to satellite tracking types as well as satellite altimetry. It was tested on simulated data which contained significant modeling errors and the results clearly demonstrate the superiority of the program compared to batch estimation.

  6. The NITE XML Toolkit: flexible annotation for multimodal language data.

    PubMed

    Carletta, Jean; Evert, Stefan; Heid, Ulrich; Kilgour, Jonathan; Robertson, Judy; Voormann, Holger

    2003-08-01

    Multimodal corpora that show humans interacting via language are now relatively easy to collect. Current tools allow one either to apply sets of time-stamped codes to the data and consider their timing and sequencing or to describe some specific linguistic structure that is present in the data, built over the top of some form of transcription. To further our understanding of human communication, the research community needs code sets with both timings and structure, designed flexibly to address the research questions at hand. The NITE XML Toolkit offers library support that software developers can call upon when writing tools for such code sets and, thus, enables richer analyses than have previously been possible. It includes data handling, a query language containing both structural and temporal constructs, components that can be used to build graphical interfaces, sample programs that demonstrate how to use the libraries, a tool for running queries, and an experimental engine that builds interfaces on the basis of declarative specifications.

  7. Using XML and Java for Astronomical Instrumentation Control

    NASA Technical Reports Server (NTRS)

    Ames, Troy; Koons, Lisa; Sall, Ken; Warsaw, Craig

    2000-01-01

    Traditionally, instrument command and control systems have been highly specialized, consisting mostly of custom code that is difficult to develop, maintain, and extend. Such solutions are initially very costly and are inflexible to subsequent engineering change requests, increasing software maintenance costs. Instrument description is too tightly coupled with details of implementation. NASA Goddard Space Flight Center is developing a general and highly extensible framework that applies to any kind of instrument that can be controlled by a computer. The software architecture combines the platform independent processing capabilities of Java with the power of the Extensible Markup Language (XML), a human readable and machine understandable way to describe structured data. A key aspect of the object-oriented architecture is software that is driven by an instrument description, written using the Instrument Markup Language (IML). ]ML is used to describe graphical user interfaces to control and monitor the instrument, command sets and command formats, data streams, and communication mechanisms. Although the current effort is targeted for the High-resolution Airborne Wideband Camera, a first-light instrument of the Stratospheric Observatory for Infrared Astronomy, the framework is designed to be generic and extensible so that it can be applied to any instrument.

  8. An XML based middleware for ECG format conversion.

    PubMed

    Li, Xuchen; Vojisavljevic, Vuk; Fang, Qiang

    2009-01-01

    With the rapid development of information and communication technologies, various e-health solutions have been proposed. The digitized medical images as well as the mono-dimension medical signals are two major forms of medical information that are stored and manipulated within an electronic medical environment. Though a variety of industrial and international standards such as DICOM and HL7 have been proposed, many proprietary formats are still pervasively used by many Hospital Information System (HIS) and Picture Archiving and Communication System (PACS) vendors. Those proprietary formats are the big hurdle to form a nationwide or even worldwide e-health network. Thus there is an imperative need to solve the medical data integration problem. Moreover, many small clinics, many hospitals in developing countries and some regional hospitals in developed countries, which have limited budget, have been shunned from embracing the latest medical information technologies due to their high costs. In this paper, we propose an XML based middleware which acts as a translation engine to seamlessly integrate clinical ECG data from a variety of proprietary data formats. Furthermore, this ECG translation engine is designed in a way that it can be integrated into an existing PACS to provide a low cost medical information integration and storage solution.

  9. Recursive-operator method in vibration problems for rod systems

    NASA Astrophysics Data System (ADS)

    Rozhkova, E. V.

    2009-12-01

    Using linear differential equations with constant coefficients describing one-dimensional dynamical processes as an example, we show that the solutions of these equations and systems are related to the solution of the corresponding numerical recursion relations and one does not have to compute the roots of the corresponding characteristic equations. The arbitrary functions occurring in the general solution of the homogeneous equations are determined by the initial and boundary conditions or are chosen from various classes of analytic functions. The solutions of the inhomogeneous equations are constructed in the form of integro-differential series acting on the right-hand side of the equation, and the coefficients of the series are determined from the same recursion relations. The convergence of formal solutions as series of a more general recursive-operator construction was proved in [1]. In the special case where the solutions of the equation can be represented in separated variables, the power series can be effectively summed, i.e., expressed in terms of elementary functions, and coincide with the known solutions. In this case, to determine the natural vibration frequencies, one obtains algebraic rather than transcendental equations, which permits exactly determining the imaginary and complex roots of these equations without using the graphic method [2, pp. 448-449]. The correctness of the obtained formulas (differentiation formulas, explicit expressions for the series coefficients, etc.) can be verified directly by appropriate substitutions; therefore, we do not prove them here.

  10. Implicit learning of a recursive rule in an artificial grammar.

    PubMed

    Poletiek, Fenna H

    2002-11-01

    Participants performed an artificial grammar learning task, in which the standard finite state grammar (J. Verb. Learn. Verb. Behavior 6 (1967) 855) was extended with a recursive rule generating self-embedded sequences. We studied the learnability of such a rule in two experiments. The results verify the general hypothesis that recursivity can be learned in an artificial grammar learning task. However this learning seems to be rather based on recognising chunks than on abstract rule induction. First, performance was better for strings with more than one level of self-embedding in the sequence, uncovering more clearly the self-embedding pattern. Second, the infinite repeatability of the recursive rule application was not spontaneously induced from the training, but it was when an additional cue about this possibility was given. Finally, participants were able to verbalise their knowledge of the fragments making up the sequences-especially in the crucial front and back positions-, whereas knowledge of the underlying structure, to the extent it was acquired, was not articulatable. The results are discussed in relation to previous studies on the implicit learnability of complex and abstract rules.

  11. Parallelizable approximate solvers for recursions arising in preconditioning

    SciTech Connect

    Shapira, Y.

    1996-12-31

    For the recursions used in the Modified Incomplete LU (MILU) preconditioner, namely, the incomplete decomposition, forward elimination and back substitution processes, a parallelizable approximate solver is presented. The present analysis shows that the solutions of the recursions depend only weakly on their initial conditions and may be interpreted to indicate that the inexact solution is close, in some sense, to the exact one. The method is based on a domain decomposition approach, suitable for parallel implementations with message passing architectures. It requires a fixed number of communication steps per preconditioned iteration, independently of the number of subdomains or the size of the problem. The overlapping subdomains are either cubes (suitable for mesh-connected arrays of processors) or constructed by the data-flow rule of the recursions (suitable for line-connected arrays with possibly SIMD or vector processors). Numerical examples show that, in both cases, the overhead in the number of iterations required for convergence of the preconditioned iteration is small relatively to the speed-up gained.

  12. Recursive linearization of multibody dynamics equations of motion

    NASA Technical Reports Server (NTRS)

    Lin, Tsung-Chieh; Yae, K. Harold

    1989-01-01

    The equations of motion of a multibody system are nonlinear in nature, and thus pose a difficult problem in linear control design. One approach is to have a first-order approximation through the numerical perturbations at a given configuration, and to design a control law based on the linearized model. Here, a linearized model is generated analytically by following the footsteps of the recursive derivation of the equations of motion. The equations of motion are first written in a Newton-Euler form, which is systematic and easy to construct; then, they are transformed into a relative coordinate representation, which is more efficient in computation. A new computational method for linearization is obtained by applying a series of first-order analytical approximations to the recursive kinematic relationships. The method has proved to be computationally more efficient because of its recursive nature. It has also turned out to be more accurate because of the fact that analytical perturbation circumvents numerical differentiation and other associated numerical operations that may accumulate computational error, thus requiring only analytical operations of matrices and vectors. The power of the proposed linearization algorithm is demonstrated, in comparison to a numerical perturbation method, with a two-link manipulator and a seven degrees of freedom robotic manipulator. Its application to control design is also demonstrated.

  13. An Accelerated Recursive Doubling Algorithm for Block Tridiagonal Systems

    SciTech Connect

    Seal, Sudip K

    2014-01-01

    Block tridiagonal systems of linear equations arise in a wide variety of scientific and engineering applications. Recursive doubling algorithm is a well-known prefix computation-based numerical algorithm that requires O(M^3(N/P + log P)) work to compute the solution of a block tridiagonal system with N block rows and block size M on P processors. In real-world applications, solutions of tridiagonal systems are most often sought with multiple, often hundreds and thousands, of different right hand sides but with the same tridiagonal matrix. Here, we show that a recursive doubling algorithm is sub-optimal when computing solutions of block tridiagonal systems with multiple right hand sides and present a novel algorithm, called the accelerated recursive doubling algorithm, that delivers O(R) improvement when solving block tridiagonal systems with R distinct right hand sides. Since R is typically about 100 1000, this improvement translates to very significant speedups in practice. Detailed complexity analyses of the new algorithm with empirical confirmation of runtime improvements are presented. To the best of our knowledge, this algorithm has not been reported before in the literature.

  14. Recursive linear optical networks for realizing quantum algorithms

    NASA Astrophysics Data System (ADS)

    Tabia, Gelo Noel

    Linear optics has played a leading role in the development of practical quantum technologies. In recent years, advances in integrated quantum photonics have significantly improved the functionality and scalability of linear optical devices. In this talk, I present recursive schemes for implementing quantum Fourier transforms and inversion about the mean in Grover's algorithm with photonic integrated circuits. By recursive, I mean that two copies of a d-dimensional unitary operation is used to build the corresponding unitary operation on 2 d modes. The linear optical networks operate on path-encoded qudits and realize d-dimensional unitary operations using O (d2) elements. To demonstrate that the recursive circuits are viable in practice, I conducted simulations of proof-of-principle experiments using a fabrication model of realistic errors in silicon-based photonic integrated devices. The results indicate high-fidelity performance in the circuits for 2-qubit and 3-qubit quantum Fourier transforms, and for quantum search on 4-item and 8-item databases. This work was funded by institutional research grant IUT2-1 from the Estonian Research Council and by the European Union through the European Regional Development Fund.

  15. Recursive Bayesian electromagnetic refractivity estimation from radar sea clutter

    NASA Astrophysics Data System (ADS)

    Vasudevan, Sathyanarayanan; Anderson, Richard H.; Kraut, Shawn; Gerstoft, Peter; Rogers, L. Ted; Krolik, Jeffrey L.

    2007-04-01

    Estimation of the range- and height-dependent index of refraction over the sea surface facilitates prediction of ducted microwave propagation loss. In this paper, refractivity estimation from radar clutter returns is performed using a Markov state space model for microwave propagation. Specifically, the parabolic approximation for numerical solution of the wave equation is used to formulate the refractivity from clutter (RFC) problem within a nonlinear recursive Bayesian state estimation framework. RFC under this nonlinear state space formulation is more efficient than global fitting of refractivity parameters when the total number of range-varying parameters exceeds the number of basis functions required to represent the height-dependent field at a given range. Moreover, the range-recursive nature of the estimator can be easily adapted to situations where the refractivity modeling changes at discrete ranges, such as at a shoreline. A fast range-recursive solution for obtaining range-varying refractivity is achieved by using sequential importance sampling extensions to state estimation techniques, namely, the forward and Viterbi algorithms. Simulation and real data results from radar clutter collected off Wallops Island, Virginia, are presented which demonstrate the ability of this method to produce propagation loss estimates that compare favorably with ground truth refractivity measurements.

  16. Enhancing Network Communication in NPSNET-V Virtual Environments Using XML-Described Dynamic Behavior (DBP) Protocols

    DTIC Science & Technology

    2001-09-01

    Language (XML) is a meta -language used to define structured-data documents for other languages. XML looks similar to HTML, but is very different and much...and are not simply limited to formatting as in HTML (Harold, 99). XML provides a universal meta -markup language that can be used to describe data...some temporary variables double jox , joy, joz, jow; // fill the blank ADUData packet with the corresponding fields from the

  17. Using a Combination of UML, C2RM, XML, and Metadata Registries to Support Long-Term Development/Engineering

    DTIC Science & Technology

    2003-01-01

    eCommerce XML/EDI Universal Business Language (UBL) Internet & Computing Human Resources (HR-XML) Semantic KEY XML SPECIFICATIONS...f i c a t i o n s Messaging Service Registry Workflow / Process User Interface E- Business Vocabulary M e s s a g e - O r i e n t e d S p e c i f...Relevancy Certainty Admissibility Accuracy Flexibility Portability Capability Supportability Producibility Consumability Resource B

  18. James Webb Space Telescope XML Database: From the Beginning to Today

    NASA Technical Reports Server (NTRS)

    Gal-Edd, Jonathan; Fatig, Curtis C.

    2005-01-01

    The James Webb Space Telescope (JWST) Project has been defining, developing, and exercising the use of a common eXtensible Markup Language (XML) for the command and telemetry (C&T) database structure. JWST is the first large NASA space mission to use XML for databases. The JWST project started developing the concepts for the C&T database in 2002. The database will need to last at least 20 years since it will be used beginning with flight software development, continuing through Observatory integration and test (I&T) and through operations. Also, a database tool kit has been provided to the 18 various flight software development laboratories located in the United States, Europe, and Canada that allows the local users to create their own databases. Recently the JWST Project has been working with the Jet Propulsion Laboratory (JPL) and Object Management Group (OMG) XML Telemetry and Command Exchange (XTCE) personnel to provide all the information needed by JWST and JPL for exchanging database information using a XML standard structure. The lack of standardization requires custom ingest scripts for each ground system segment, increasing the cost of the total system. Providing a non-proprietary standard of the telemetry and command database definition formation will allow dissimilar systems to communicate without the need for expensive mission specific database tools and testing of the systems after the database translation. The various ground system components that would benefit from a standardized database are the telemetry and command systems, archives, simulators, and trending tools. JWST has exchanged the XML database with the Eclipse, EPOCH, ASIST ground systems, Portable spacecraft simulator (PSS), a front-end system, and Integrated Trending and Plotting System (ITPS) successfully. This paper will discuss how JWST decided to use XML, the barriers to a new concept, experiences utilizing the XML structure, exchanging databases with other users, and issues that have

  19. Adaptive clutter rejection for ultrasound color flow imaging based on recursive eigendecomposition.

    PubMed

    You, Wei; Wang, Yuanyuan

    2009-10-01

    In the conventional eigenfilter used to reject clutter components of ultrasound color flow imaging, input samples are required to be statistically stationary. However, clutter movements may vary over the depth of the imaged area, which makes the eigenfilter less efficient. In the current study, a novel clutter rejection method is proposed based on the recursive eigendecomposition algorithm. In this method, the current eigenvector matrix of the ultrasound echo correlation matrix, which will be used to construct the clutter subspace, is determined by previous eigenvector matrices and the current input. After the estimated clutter signal is obtained by projecting the original input into the clutter space, each filtered output is eventually obtained by subtracting the estimated clutter signal from the original input. This procedure is iterated for each sample volume along the depth. During the updating process, a forgetting factor is introduced to determine proper weights for different inputs. Simulated data in 3 situations and in vivo data collected from human carotid arteries are used to compare the proposed method with other popular clutter filters. Results show that the proposed method can achieve the most accurate velocity profiles in all simulation situations and introduces the fewest velocity artifacts in the tissue region in the in vivo experiment.

  20. Segmentation and tracking of the electro-encephalogram signal using an adaptive recursive bandpass filter.

    PubMed

    Gharieb, R R; Cichocki, A

    2001-03-01

    An adaptive filtering approach for the segmentation and tracking of electro-encephalogram (EEG) signal waves is described. In this approach, an adaptive recursive bandpass filter is employed for estimating and tracking the centre frequency associated with each EEG wave. The main advantage inherent in the approach is that the employed adaptive filter has only one unknown coefficient to be updated. This coefficient, having an absolute value less than 1, represents an efficient distinct feature for each EEG specific wave, and its time function reflects the non-stationarity behaviour of the EEG signal. Therefore the proposed approach is simple and accurate in comparison with existing multivariate adaptive approaches. The approach is examined using extensive computer simulations. It is applied to computer-generated EEG signals composed of different waves. The adaptive filter coefficient (i.e. the segmentation parameter) is -0.492 for the delta wave, -0.360 for the theta wave, -0.191 for the alpha wave, -0.027 for the sigma wave, 0.138 for the beta wave and 0.605 for the gamma wave. This implies that the segmentation parameter increases with the increase in the centre frequency of the EEG waves, which provides fast on-line information about the behaviour of the EEG signal. The approach is also applied to real-world EEG data for the detection of sleep spindles.

  1. Recursive Bayesian synthetic aperture geoacoustic inversion in the presence of motion dynamics.

    PubMed

    Tan, Bien Aik; Gerstoft, Peter; Yardim, Caglar; Hodgkiss, William S

    2014-09-01

    A low signal to noise ratio (SNR), single source/receiver, broadband, frequency-coherent matched-field inversion procedure recently has been proposed. It exploits coherently repeated transmissions to improve estimation of the geoacoustic parameters. The long observation time improves the SNR and creates a synthetic aperture due to relative source-receiver motion. To model constant velocity source/receiver horizontal motion, waveguide Doppler theory for normal modes is necessary. However, the inversion performance degrades when source/receiver acceleration exists. Furthermore processing a train of pulses all at once does not take advantage of the natural incremental acquisition of data along with the ability to assess the temporal evolution of parameter uncertainty. Here a recursive Bayesian estimation approach is developed that coherently processes the data pulse by pulse and incrementally updates estimates of parameter uncertainty. It also approximates source/receiver acceleration by assuming piecewise constant but linearly changing source/receiver velocities. When the source/receiver acceleration exists, it is shown that modeling acceleration can reduce further the parameter estimation biases and uncertainties. The method is demonstrated in simulation and in the analysis of low SNR, 100-900 Hz linear frequency modulated (LFM) pulses from the Shallow Water 2006 experiment.

  2. Discovery of a Recursive Principle: An Artificial Grammar Investigation of Human Learning of a Counting Recursion Language.

    PubMed

    Cho, Pyeong Whan; Szkudlarek, Emily; Tabor, Whitney

    2016-01-01

    Learning is typically understood as a process in which the behavior of an organism is progressively shaped until it closely approximates a target form. It is easy to comprehend how a motor skill or a vocabulary can be progressively learned-in each case, one can conceptualize a series of intermediate steps which lead to the formation of a proficient behavior. With grammar, it is more difficult to think in these terms. For example, center embedding recursive structures seem to involve a complex interplay between multiple symbolic rules which have to be in place simultaneously for the system to work at all, so it is not obvious how the mechanism could gradually come into being. Here, we offer empirical evidence from a new artificial language (or "artificial grammar") learning paradigm, Locus Prediction, that, despite the conceptual conundrum, recursion acquisition occurs gradually, at least for a simple formal language. In particular, we focus on a variant of the simplest recursive language, a (n) b (n) , and find evidence that (i) participants trained on two levels of structure (essentially ab and aabb) generalize to the next higher level (aaabbb) more readily than participants trained on one level of structure (ab) combined with a filler sentence; nevertheless, they do not generalize immediately; (ii) participants trained up to three levels (ab, aabb, aaabbb) generalize more readily to four levels than participants trained on two levels generalize to three; (iii) when we present the levels in succession, starting with the lower levels and including more and more of the higher levels, participants show evidence of transitioning between the levels gradually, exhibiting intermediate patterns of behavior on which they were not trained; (iv) the intermediate patterns of behavior are associated with perturbations of an attractor in the sense of dynamical systems theory. We argue that all of these behaviors indicate a theory of mental representation in which recursive

  3. Discovery of a Recursive Principle: An Artificial Grammar Investigation of Human Learning of a Counting Recursion Language

    PubMed Central

    Cho, Pyeong Whan; Szkudlarek, Emily; Tabor, Whitney

    2016-01-01

    Learning is typically understood as a process in which the behavior of an organism is progressively shaped until it closely approximates a target form. It is easy to comprehend how a motor skill or a vocabulary can be progressively learned—in each case, one can conceptualize a series of intermediate steps which lead to the formation of a proficient behavior. With grammar, it is more difficult to think in these terms. For example, center embedding recursive structures seem to involve a complex interplay between multiple symbolic rules which have to be in place simultaneously for the system to work at all, so it is not obvious how the mechanism could gradually come into being. Here, we offer empirical evidence from a new artificial language (or “artificial grammar”) learning paradigm, Locus Prediction, that, despite the conceptual conundrum, recursion acquisition occurs gradually, at least for a simple formal language. In particular, we focus on a variant of the simplest recursive language, anbn, and find evidence that (i) participants trained on two levels of structure (essentially ab and aabb) generalize to the next higher level (aaabbb) more readily than participants trained on one level of structure (ab) combined with a filler sentence; nevertheless, they do not generalize immediately; (ii) participants trained up to three levels (ab, aabb, aaabbb) generalize more readily to four levels than participants trained on two levels generalize to three; (iii) when we present the levels in succession, starting with the lower levels and including more and more of the higher levels, participants show evidence of transitioning between the levels gradually, exhibiting intermediate patterns of behavior on which they were not trained; (iv) the intermediate patterns of behavior are associated with perturbations of an attractor in the sense of dynamical systems theory. We argue that all of these behaviors indicate a theory of mental representation in which recursive

  4. Validation and Simplification of the Radiation Therapy Oncology Group Recursive Partitioning Analysis Classification for Glioblastoma

    SciTech Connect

    Li Jing; Wang Meihua; Won, Minhee; Shaw, Edward G.; Coughlin, Christopher; Curran, Walter J.; Mehta, Minesh P.

    2011-11-01

    Purpose: Previous recursive partitioning analysis (RPA) of patients with malignant glioma (glioblastoma multiforme [GBM] and anaplastic astrocytoma [AA]) produced six prognostic groups (I-VI) classified by six factors. We sought here to determine whether the classification for GBM could be improved by using an updated Radiation Therapy Oncology Group (RTOG) GBM database excluding AA and by considering additional baseline variables. Methods and Materials: The new analysis considered 42 baseline variables and 1,672 GBM patients from the expanded RTOG glioma database. Patients receiving radiation only were excluded such that all patients received radiation+carmustine. 'Radiation dose received' was replaced with 'radiation dose assigned.' The new RPA models were compared with the original model by applying them to a test dataset comprising 488 patients from six other RTOG trials. Fitness of the original and new models was evaluated using explained variation. Results: The original RPA model explained more variations in survival in the test dataset than did the new models (20% vs. 15%) and was therefore chosen for further analysis. It was reduced by combining Classes V and VI to produce three prognostic classes (Classes III, IV, and V+VI), as Classes V and VI had indistinguishable survival in the test dataset. The simplified model did not further improve performance (explained variation 18% vs. 20%) but is easier to apply because it involves only four variables: age, performance status, extent of resection, and neurologic function. Applying this simplified model to the updated GBM database resulted in three distinct classes with median survival times of 17.1, 11.2, and 7.5 months for Classes III, IV, and V+VI, respectively. Conclusions: The final model, the simplified original RPA model combining Classes V and VI, resulted in three distinct prognostic groups defined by age, performance status, extent of resection, and neurologic function. This classification will be used

  5. Alternatives to relational database: comparison of NoSQL and XML approaches for clinical data storage.

    PubMed

    Lee, Ken Ka-Yin; Tang, Wai-Choi; Choi, Kup-Sze

    2013-04-01

    Clinical data are dynamic in nature, often arranged hierarchically and stored as free text and numbers. Effective management of clinical data and the transformation of the data into structured format for data analysis are therefore challenging issues in electronic health records development. Despite the popularity of relational databases, the scalability of the NoSQL database model and the document-centric data structure of XML databases appear to be promising features for effective clinical data management. In this paper, three database approaches--NoSQL, XML-enabled and native XML--are investigated to evaluate their suitability for structured clinical data. The database query performance is reported, together with our experience in the databases development. The results show that NoSQL database is the best choice for query speed, whereas XML databases are advantageous in terms of scalability, flexibility and extensibility, which are essential to cope with the characteristics of clinical data. While NoSQL and XML technologies are relatively new compared to the conventional relational database, both of them demonstrate potential to become a key database technology for clinical data management as the technology further advances.

  6. Thinking about thinking - thinking about measurement: a Rasch analysis of recursive thinking.

    PubMed

    Müller, Ulrich; Overton, Willis F

    2010-01-01

    Two studies were conducted to examine the dimensionality and hierarchical organization of a measure of recursive thinking. In Study 1, Rasch analysis supported the claim that the recursive thinking task measures a single underlying dimension. Item difficulty, however, appeared to be influenced not only by level of embeddedness but also by syntactic features. In Study 2, this hypothesis was tested by adding new items to the recursive thinking measure. Rasch analysis of the modified recursive thinking task produced evidence for the undimensionality and segmentation. However, Study 2 did not support the idea that syntactic features influence item difficulty.

  7. Genome-wide identification of zero nucleotide recursive splicing in Drosophila.

    PubMed

    Duff, Michael O; Olson, Sara; Wei, Xintao; Garrett, Sandra C; Osman, Ahmad; Bolisetty, Mohan; Plocik, Alex; Celniker, Susan E; Graveley, Brenton R

    2015-05-21

    Recursive splicing is a process in which large introns are removed in multiple steps by re-splicing at ratchet points--5' splice sites recreated after splicing. Recursive splicing was first identified in the Drosophila Ultrabithorax (Ubx) gene and only three additional Drosophila genes have since been experimentally shown to undergo recursive splicing. Here we identify 197 zero nucleotide exon ratchet points in 130 introns of 115 Drosophila genes from total RNA sequencing data generated from developmental time points, dissected tissues and cultured cells. The sequential nature of recursive splicing was confirmed by identification of lariat introns generated by splicing to and from the ratchet points. We also show that recursive splicing is a constitutive process, that depletion of U2AF inhibits recursive splicing, and that the sequence and function of ratchet points are evolutionarily conserved in Drosophila. Finally, we identify four recursively spliced human genes, one of which is also recursively spliced in Drosophila. Together, these results indicate that recursive splicing is commonly used in Drosophila, occurs in humans, and provides insight into the mechanisms by which some large introns are removed.

  8. Interactive, Secure Web-enabled Aircraft Engine Simulation Using XML Databinding Integration

    NASA Technical Reports Server (NTRS)

    Lin, Risheng; Afjeh, Abdollah A.

    2003-01-01

    This paper discusses the detailed design of an XML databinding framework for aircraft engine simulation. The framework provides an object interface to access and use engine data. while at the same time preserving the meaning of the original data. The Language independent representation of engine component data enables users to move around XML data using HTTP through disparate networks. The application of this framework is demonstrated via a web-based turbofan propulsion system simulation using the World Wide Web (WWW). A Java Servlet based web component architecture is used for rendering XML engine data into HTML format and dealing with input events from the user, which allows users to interact with simulation data from a web browser. The simulation data can also be saved to a local disk for archiving or to restart the simulation at a later time.

  9. DataStaR: Bridging XML and OWL in Science Metadata Management

    NASA Astrophysics Data System (ADS)

    Lowe, Brian

    DataStaR is a science data “staging repository” developed by Albert R. Mann Library at Cornell University that produces semantic metadata while enabling the publication of data sets and accompanying metadata to discipline-specific data centers or to Cornell’s institutional repository. DataStaR, which employs OWL and RDF in its metadata store, serves as a Web-based platform for production and management of metadata and aims to reduce redundant manual input by reusing named ontology individuals. A key requirement of DataStaR is the ability to produce metadata records conforming to existing XML schemas that have been adopted by scientific communities. To facilitate this, DataStaR integrates ontologies that directly reflect XML schemas, generates HTML editing forms, and “lowers” ontology axioms into XML documents compliant with existing schemas. This paper describes our approach and implementation, and discusses the challenges involved.

  10. ForConX: A forcefield conversion tool based on XML.

    PubMed

    Lesch, Volker; Diddens, Diddo; Bernardes, Carlos E S; Golub, Benjamin; Dequidt, Alain; Zeindlhofer, Veronika; Sega, Marcello; Schröder, Christian

    2017-04-05

    The force field conversion from one MD program to another one is exhausting and error-prone. Although single conversion tools from one MD program to another exist not every combination and both directions of conversion are available for the favorite MD programs Amber, Charmm, Dl-Poly, Gromacs, and Lammps. We present here a general tool for the force field conversion on the basis of an XML document. The force field is converted to and from this XML structure facilitating the implementation of new MD programs for the conversion. Furthermore, the XML structure is human readable and can be manipulated before continuing the conversion. We report, as testcases, the conversions of topologies for acetonitrile, dimethylformamide, and 1-ethyl-3-methylimidazolium trifluoromethanesulfonate comprising also Urey-Bradley and Ryckaert-Bellemans potentials. © 2017 Wiley Periodicals, Inc.

  11. Standardization of XML Database Exchanges and the James Webb Space Telescope Experience

    NASA Technical Reports Server (NTRS)

    Gal-Edd, Jonathan; Detter, Ryan; Jones, Ron; Fatig, Curtis C.

    2007-01-01

    Personnel from the National Aeronautics and Space Administration (NASA) James Webb Space Telescope (JWST) Project have been working with various standard communities such the Object Management Group (OMG) and the Consultative Committee for Space Data Systems (CCSDS) to assist in the definition of a common extensible Markup Language (XML) for database exchange format. The CCSDS and OMG standards are intended for the exchange of core command and telemetry information, not for all database information needed to exercise a NASA space mission. The mission-specific database, containing all the information needed for a space mission, is translated from/to the standard using a translator. The standard is meant to provide a system that encompasses 90% of the information needed for command and telemetry processing. This paper will discuss standardization of the XML database exchange format, tools used, and the JWST experience, as well as future work with XML standard groups both commercial and government.

  12. At-sea demonstration of RF sensor tasking using XML over a worldwide network

    NASA Astrophysics Data System (ADS)

    Kellogg, Robert L.; Lee, Tom; Dumas, Diane; Raggo, Barbara

    2003-07-01

    As part of an At-Sea Demonstration for Space and Naval Warfare Command (SPAWAR, PMW-189), a prototype RF sensor for signal acquisition and direction finding queried and received tasking via a secure worldwide Automated Data Network System (ADNS). Using extended mark-up language (XML) constructs, both mission and signal tasking were available for push and pull Battlespace management. XML tasking was received by the USS Cape St George (CG-71) during an exercise along the Gulf Coast of the US from a test facility at SPAWAR, San Diego, CA. Although only one ship was used in the demonstration, the intent of the software initiative was to show that a network of different RF sensors on different platforms with different capabilitis could be tasked by a common web agent. A sensor software agent interpreted the XML task to match the sensor's capability. Future improvements will focus on enlarging the domain of mission tasking and incorporate report management.

  13. Recursive phase estimation with a spatial radar carrier

    NASA Astrophysics Data System (ADS)

    Garcia-Marquez, Jorge; Servin Guirado, Manuel; Paez, Gonzalo; Malacara-Hernandez, Daniel

    1999-08-01

    An interferogram can be demodulated to find the wavefront shape if a radial carrier is introduced. The phase determination is made in the space domain, but the low-pass filter characteristics must be properly chosen. One disadvantage of this method is the possible removal of some frequencies from the central lobe, resulting in a misinterpretation of the true phase. Nevertheless isolating the central order by using a recursive method when a radial carrier reference is used is possible. An example of a recovered phase from a simulated interferogram is shown.

  14. Recursive time-varying filter banks for subband image coding

    NASA Technical Reports Server (NTRS)

    Smith, Mark J. T.; Chung, Wilson C.

    1992-01-01

    Filter banks and wavelet decompositions that employ recursive filters have been considered previously and are recognized for their efficiency in partitioning the frequency spectrum. This paper presents an analysis of a new infinite impulse response (IIR) filter bank in which these computationally efficient filters may be changed adaptively in response to the input. The filter bank is presented and discussed in the context of finite-support signals with the intended application in subband image coding. In the absence of quantization errors, exact reconstruction can be achieved and by the proper choice of an adaptation scheme, it is shown that IIR time-varying filter banks can yield improvement over conventional ones.

  15. An accelerated procedure for recursive feature ranking on microarray data.

    PubMed

    Furlanello, C; Serafini, M; Merler, S; Jurman, G

    2003-01-01

    We describe a new wrapper algorithm for fast feature ranking in classification problems. The Entropy-based Recursive Feature Elimination (E-RFE) method eliminates chunks of uninteresting features according to the entropy of the weights distribution of a SVM classifier. With specific regard to DNA microarray datasets, the method is designed to support computationally intensive model selection in classification problems in which the number of features is much larger than the number of samples. We test E-RFE on synthetic and real data sets, comparing it with other SVM-based methods. The speed-up obtained with E-RFE supports predictive modeling on high dimensional microarray data.

  16. Recursive Model Identification for the Evaluation of Baroreflex Sensitivity.

    PubMed

    Le Rolle, Virginie; Beuchée, Alain; Praud, Jean-Paul; Samson, Nathalie; Pladys, Patrick; Hernández, Alfredo I

    2016-12-01

    A method for the recursive identification of physiological models of the cardiovascular baroreflex is proposed and applied to the time-varying analysis of vagal and sympathetic activities. The proposed method was evaluated with data from five newborn lambs, which were acquired during injection of vasodilator and vasoconstrictors and the results show a close match between experimental and simulated signals. The model-based estimation of vagal and sympathetic contributions were consistent with physiological knowledge and the obtained estimators of vagal and sympathetic activities were compared to traditional markers associated with baroreflex sensitivity. High correlations were observed between traditional markers and model-based indices.

  17. Clutter suppression using recursive and nonrecursive MTI filters

    NASA Astrophysics Data System (ADS)

    Liu, Bede

    1988-05-01

    The author calculates the effectiveness of clutter suppression of a moving-target-indicator (MTI) filter in tandem with a fast-Fourier-transform (FFT) Doppler filter bank, taking into account the transient response of the MTI filter. Both recursive and nonrecursive filters are considered. The analysis is extended to the high-pulse-repetition-frequency (PRF) case with clutter fold over. The results can be used to select key design parameters, including the MTI filter, the window size, and the initial transient segment to be discarded. Numerical examples are included.

  18. Vision-based recursive estimation of rotorcraft obstacle locations

    NASA Technical Reports Server (NTRS)

    Leblanc, D. J.; Mcclamroch, N. H.

    1992-01-01

    The authors address vision-based passive ranging during nap-of-the-earth (NOE) rotorcraft flight. They consider the problem of estimating the relative location of identifiable features on nearby obstacles, assuming a sequence of noisy camera images and imperfect measurements of the camera's translation and rotation. An iterated extended Kalman filter is used to provide recursive range estimation. The correspondence problem is simplified by predicting and tracking each feature's image within the Kalman filter framework. Simulation results are presented which show convergent estimates and generally successful feature point tracking. Estimation performance degrades for features near the optical axis and for accelerating motions. Image tracking is also sensitive to angular rate.

  19. Recursive partitioned inversion of large (1500 x 1500) symmetric matrices

    NASA Technical Reports Server (NTRS)

    Putney, B. H.; Brownd, J. E.; Gomez, R. A.

    1976-01-01

    A recursive algorithm was designed to invert large, dense, symmetric, positive definite matrices using small amounts of computer core, i.e., a small fraction of the core needed to store the complete matrix. The described algorithm is a generalized Gaussian elimination technique. Other algorithms are also discussed for the Cholesky decomposition and step inversion techniques. The purpose of the inversion algorithm is to solve large linear systems of normal equations generated by working geodetic problems. The algorithm was incorporated into a computer program called SOLVE. In the past the SOLVE program has been used in obtaining solutions published as the Goddard earth models.

  20. Description of an exact, recursive method to simplify shading calculations

    NASA Astrophysics Data System (ADS)

    Nawrocki, A. D.; Kammerud, R.

    An exact, recursive method called SHADE is described which attempts to simplify shading calculations as performed by a programmable calculator or microcomputer. Preliminary applications of SHADE using a Hewlett Packard HP-41C programmable calculator are outlined. For a given solar hour, SHADE is used to compute the following quantities for overhang and side fin combinations which shade various openings: the percentage of the total area of the opening which is shaded; the shaded area itself; the cosine of the angle of incidence between the Sun and glazing surface; the direct insolation at this surface, with and without shading; and the direct solar power at this surface, with and without shading.

  1. Recursive Ant Colony Global Optimization: a new technique for the inversion of geophysical data

    NASA Astrophysics Data System (ADS)

    Gupta, D. K.; Gupta, J. P.; Arora, Y.; Singh, U. K.

    2011-12-01

    We present a new method called Recursive Ant Colony Global Optimization (RACO) technique, a modified form of general ACO, which can be used to find the best solutions to inversion problems in geophysics. RACO simulates the social behaviour of ants to find the best path between the nest and the food source. A new term depth has been introduced, which controls the extent of recursion. A selective number of cities get qualified for the successive depth. The results of one depth are used to construct the models for the next depth and the range of values for each of the parameters is reduced without any change to the number of models. The three additional steps performed after each depth, are the pheromone tracking, pheromone updating and city selection. One of the advantages of RACO over ACO is that if a problem has multiple solutions, then pheromone accumulation will take place at more than one city thereby leading to formation of multiple nested ACO loops within the ACO loop of the previous depth. Also, while the convergence of ACO is almost linear, RACO shows exponential convergence and hence is faster than the ACO. RACO proves better over some other global optimization techniques, as it does not require any initial values to be assigned to the parameters function. The method has been tested on some mathematical functions, synthetic self-potential (SP) and synthetic gravity data. The obtained results reveal the efficiency and practicability of the method. The method is found to be efficient enough to solve the problems of SP and gravity anomalies due to a horizontal cylinder, a sphere, an inclined sheet and multiple idealized bodies buried inside the earth. These anomalies with and without noise were inverted using the RACO algorithm. The obtained results were compared with those obtained from the conventional methods and it was found that RACO results are more accurate. Finally this optimization technique was applied to real field data collected over the Surda

  2. RESRAD update

    SciTech Connect

    Yu, C.; Cheng, J.J.; Zielen, A.J.; Jones, L.G.; LePoire, D.J.; Wang, Y.Y. ); Yuan, Y.C. ); Loureiro, C.O. . Escola de Engenharia); Wallo, A. III; Peterson, H. . Offic

    1993-01-01

    A microcomputer program called RESRAD, which implements a pathway analysis method for radiological risk assessment, was developed by Argonne National Laboratory (ANL) in 1989. This program is used to derive allowable residual concentrations of radionuclides in soil and to predict effective dose equivalents and excess cancer incidence risks incurred by an individual exposed to radioactive materials. Since its development, the RESRAD code has been adopted by DOE in Order 5400.5 for the derivation of soil cleanup criteria and dose calculations, and it has been used widely by DOE, other agencies, and their contractors. The original models used by ANL to develop RESRAD were initially developed as part of a DOE effort that began in the early 1980s and involved most of the national laboratories and DOE program offices. The RESRAD code is continuously improved and updated to incorporate comments from users and new features that ease the interaction with users and increase the code's capability and flexibility. The DOE Offices of Environmental Guidance and Environmental Restoration also provide periodic guidance regarding any significant changes to the code. The RESRAD update, Version 5.0, has substantial improvements in many aspects compared with the last version released in 1989.

  3. RESRAD update

    SciTech Connect

    Yu, C.; Cheng, J.J.; Zielen, A.J.; Jones, L.G.; LePoire, D.J.; Wang, Y.Y.; Yuan, Y.C.; Loureiro, C.O.; Wallo, A. III; Peterson, H.; H Williams, W.A.

    1993-05-01

    A microcomputer program called RESRAD, which implements a pathway analysis method for radiological risk assessment, was developed by Argonne National Laboratory (ANL) in 1989. This program is used to derive allowable residual concentrations of radionuclides in soil and to predict effective dose equivalents and excess cancer incidence risks incurred by an individual exposed to radioactive materials. Since its development, the RESRAD code has been adopted by DOE in Order 5400.5 for the derivation of soil cleanup criteria and dose calculations, and it has been used widely by DOE, other agencies, and their contractors. The original models used by ANL to develop RESRAD were initially developed as part of a DOE effort that began in the early 1980s and involved most of the national laboratories and DOE program offices. The RESRAD code is continuously improved and updated to incorporate comments from users and new features that ease the interaction with users and increase the code`s capability and flexibility. The DOE Offices of Environmental Guidance and Environmental Restoration also provide periodic guidance regarding any significant changes to the code. The RESRAD update, Version 5.0, has substantial improvements in many aspects compared with the last version released in 1989.

  4. Dynamic XML-based exchange of relational data: application to the Human Brain Project.

    PubMed

    Tang, Zhengming; Kadiyska, Yana; Li, Hao; Suciu, Dan; Brinkley, James F

    2003-01-01

    This paper discusses an approach to exporting relational data in XML format for data exchange over the web. We describe the first real-world application of SilkRoute, a middleware program that dynamically converts existing relational data to a user-defined XML DTD. The application, called XBrain, wraps SilkRoute in a Java Server Pages framework, thus permitting a web-based XQuery interface to a legacy relational database. The application is demonstrated as a query interface to the University of Washington Brain Project's Language Map Experiment Management System, which is used to manage data about language organization in the brain.

  5. Research on heterogeneous data query and sharing mode based on XML

    NASA Astrophysics Data System (ADS)

    Sang, Yaqun; Mu, Qi

    2013-07-01

    At present, XML language has been regarded as a standard of data exchange due to such attributes as being unrelevant to platform, auto-description of itself, easy extention, separation of content and form, and so on. Therefore, in the course of study of share of heterogeneous-data in coal enterprises, we design a heterogeneous-data query system based on XML to provide a unified data integration platform and share of information resource with high quality and speed, which will solve a problem with information islands produced in coal enterprises.

  6. Chemical markup, XML, and the World Wide Web. 4. CML schema.

    PubMed

    Murray-Rust, Peter; Rzepa, Henry S

    2003-01-01

    A revision to Chemical Markup Language (CML) is presented as a XML Schema compliant form, modularized into nonchemical and chemical components. STMML contains generic concepts for numeric data and scientific units, while CMLCore retains most of the chemical functionality of the original CML 1.0 and extends it by adding handlers for chemical substances, extended bonding models and names. We propose extension via new namespaced components for chemical queries, reactions, spectra, and computational chemistry. The conformance with XML schemas allows much greater control over datatyping, document validation, and structure.

  7. The Surgical Simulation and Training Markup Language (SSTML): an XML-based language for medical simulation.

    PubMed

    Bacon, James; Tardella, Neil; Pratt, Janey; Hu, John; English, James

    2006-01-01

    Under contract with the Telemedicine & Advanced Technology Research Center (TATRC), Energid Technologies is developing a new XML-based language for describing surgical training exercises, the Surgical Simulation and Training Markup Language (SSTML). SSTML must represent everything from organ models (including tissue properties) to surgical procedures. SSTML is an open language (i.e., freely downloadable) that defines surgical training data through an XML schema. This article focuses on the data representation of the surgical procedures and organ modeling, as they highlight the need for a standard language and illustrate the features of SSTML. Integration of SSTML with software is also discussed.

  8. A new XML-aware compression technique for improving performance of healthcare information systems over hospital networks.

    PubMed

    Al-Shammary, Dhiah; Khalil, Ibrahim

    2010-01-01

    Most organizations exchange, collect, store and process data over the Internet. Many hospital networks deploy Web services to send and receive patient information. SOAP (Simple Object Access Protocol) is the most usable communication protocol for Web services. XML is the standard encoding language of SOAP messages. However, the major drawback of XML messages is the high network traffic caused by large overheads. In this paper, two XML-aware compressors are suggested to compress patient messages stemming from any data transactions between Web clients and servers. The proposed compression techniques are based on the XML structure concepts and use both fixed-length and Huffman encoding methods for translating the XML message tree. Experiments show that they outperform all the conventional compression methods and can save tremendous amount of network bandwidth.

  9. Stochastic and recursive calibration for operational, large-scale, agricultural land and water use management models

    NASA Astrophysics Data System (ADS)

    Maneta, M. P.; Kimball, J. S.; Jencso, K. G.

    2015-12-01

    Managing the impact of climatic cycles on agricultural production, on land allocation, and on the state of active and projected water sources is challenging. This is because in addition to the uncertainties associated with climate projections, it is difficult to anticipate how farmers will respond to climatic change or to economic and policy incentives. Some sophisticated decision support systems available to water managers consider farmers' adaptive behavior but they are data intensive and difficult to apply operationally over large regions. Satellite-based observational technologies, in conjunction with models and assimilation methods, create an opportunity for new, cost-effective analysis tools to support policy and decision-making over large spatial extents at seasonal scales.We present an integrated modeling framework that can be driven by satellite remote sensing to enable robust regional assessment and prediction of climatic and policy impacts on agricultural production, water resources, and management decisions. The core of this framework is a widely used model of agricultural production and resource allocation adapted to be used in conjunction with remote sensing inputs to quantify the amount of land and water farmers allocate for each crop they choose to grow on a seasonal basis in response to reduced or enhanced access to water due to climatic or policy restrictions. A recursive Bayesian update method is used to adjust the model parameters by assimilating information on crop acreage, production, and crop evapotranspiration as a proxy for water use that can be estimated from high spatial resolution satellite remote sensing. The data assimilation framework blends new and old information to avoid over-calibration to the specific conditions of a single year and permits the updating of parameters to track gradual changes in the agricultural system.This integrated framework provides an operational means of monitoring and forecasting what crops will be grown

  10. Recursive mentalizing and common knowledge in the bystander effect.

    PubMed

    Thomas, Kyle A; De Freitas, Julian; DeScioli, Peter; Pinker, Steven

    2016-05-01

    The more potential helpers there are, the less likely any individual is to help. A traditional explanation for this bystander effect is that responsibility diffuses across the multiple bystanders, diluting the responsibility of each. We investigate an alternative, which combines the volunteer's dilemma (each bystander is best off if another responds) with recursive theory of mind (each infers what the others know about what he knows) to predict that actors will strategically shirk when they think others feel compelled to help. In 3 experiments, participants responded to a (fictional) person who needed help from at least 1 volunteer. Participants were in groups of 2 or 5 and had varying information about whether other group members knew that help was needed. As predicted, people's decision to help zigzagged with the depth of their asymmetric, recursive knowledge (e.g., "John knows that Michael knows that John knows help is needed"), and replicated the classic bystander effect when they had common knowledge (everyone knowing what everyone knows). The results demonstrate that the bystander effect may result not from a mere diffusion of responsibility but specifically from actors' strategic computations.

  11. Efficient Execution of Recursive Programs on Commodity Vector Hardware

    SciTech Connect

    Ren, Bin; Jo, Youngjoon; Krishnamoorthy, Sriram; Agrawal, Kunal; Kulkarni, Milind

    2015-06-13

    The pursuit of computational efficiency has led to the proliferation of throughput-oriented hardware, from GPUs to increasingly-wide vector units on commodity processors and accelerators. This hardware is designed to efficiently execute data-parallel computations in a vectorized manner. However, many algorithms are more naturally expressed as divide-and-conquer, recursive, task-parallel computations; in the absence of data parallelism, it seems that such algorithms are not well-suited to throughput-oriented architectures. This paper presents a set of novel code transformations that expose the data-parallelism latent in recursive, task-parallel programs. These transformations facilitate straightforward vectorization of task-parallel programs on commodity hardware. We also present scheduling policies that maintain high utilization of vector resources while limiting space usage. Across several task-parallel benchmarks, we demonstrate both efficient vector resource utilization and substantial speedup on chips using Intel's SSE4.2 vector units as well as accelerators using Intel's AVX512 units.

  12. REQUEST: A Recursive QUEST Algorithm for Sequential Attitude Determination

    NASA Technical Reports Server (NTRS)

    Bar-Itzhack, Itzhack Y.

    1996-01-01

    In order to find the attitude of a spacecraft with respect to a reference coordinate system, vector measurements are taken. The vectors are pairs of measurements of the same generalized vector, taken in the spacecraft body coordinates, as well as in the reference coordinate system. We are interested in finding the best estimate of the transformation between these coordinate system.s The algorithm called QUEST yields that estimate where attitude is expressed by a quarternion. Quest is an efficient algorithm which provides a least squares fit of the quaternion of rotation to the vector measurements. Quest however, is a single time point (single frame) batch algorithm, thus measurements that were taken at previous time points are discarded. The algorithm presented in this work provides a recursive routine which considers all past measurements. The algorithm is based on on the fact that the, so called, K matrix, one of whose eigenvectors is the sought quaternion, is linerly related to the measured pairs, and on the ability to propagate K. The extraction of the appropriate eigenvector is done according to the classical QUEST algorithm. This stage, however, can be eliminated, and the computation simplified, if a standard eigenvalue-eigenvector solver algorithm is used. The development of the recursive algorithm is presented and illustrated via a numerical example.

  13. Quicksort: Combining Concurrency, Recursion, and Mutable Data Structures

    NASA Astrophysics Data System (ADS)

    Kitchin, David; Quark, Adrian; Misra, Jayadev

    Quicksort (Commun. ACM 4(7):321-322, 1961) remains one of the most studied algorithms in computer science. It is important not only as a practical sorting method, but also as a splendid teaching aid for introducing recursion and systematic algorithm development. The algorithm has been studied extensively; so, it is natural to assume that everything that needs to be said about it has already been said. Yet, in attempting to code it using a recent programming language of our design, we discovered that its structure is more clearly expressed as a concurrent program that manipulates a shared mutable store, without any locking or explicit synchronization. In this paper, we describe the essential aspects of our programming language Orc (Proceedings of FMOODS/FORTE, vol. 5522 of LNCS, pp. 1-25. Springer 2009), show a number of examples that combine its features in various forms, and then develop a concise description of Quicksort. We hope to highlight the importance of including concurrency, recursion and mutability within a single theory.

  14. Recursive architecture for large-scale adaptive system

    NASA Astrophysics Data System (ADS)

    Hanahara, Kazuyuki; Sugiyama, Yoshihiko

    1994-09-01

    'Large scale' is one of major trends in the research and development of recent engineering, especially in the field of aerospace structural system. This term expresses the large scale of an artifact in general, however, it also implies the large number of the components which make up the artifact in usual. Considering a large scale system which is especially used in remote space or deep-sea, such a system should be adaptive as well as robust by itself, because its control as well as maintenance by human operators are not easy due to the remoteness. An approach to realizing this large scale, adaptive and robust system is to build the system as an assemblage of components which are respectively adaptive by themselves. In this case, the robustness of the system can be achieved by using a large number of such components and suitable adaptation as well as maintenance strategies. Such a system gathers many research's interest and their studies such as decentralized motion control, configurating algorithm and characteristics of structural elements are reported. In this article, a recursive architecture concept is developed and discussed towards the realization of large scale system which consists of a number of uniform adaptive components. We propose an adaptation strategy based on the architecture and its implementation by means of hierarchically connected processing units. The robustness and the restoration from degeneration of the processing unit are also discussed. Two- and three-dimensional adaptive truss structures are conceptually designed based on the recursive architecture.

  15. Adaptive model reduction for continuous systems via recursive rational interpolation

    NASA Technical Reports Server (NTRS)

    Lilly, John H.

    1994-01-01

    A method for adaptive identification of reduced-order models for continuous stable SISO and MIMO plants is presented. The method recursively finds a model whose transfer function (matrix) matches that of the plant on a set of frequencies chosen by the designer. The algorithm utilizes the Moving Discrete Fourier Transform (MDFT) to continuously monitor the frequency-domain profile of the system input and output signals. The MDFT is an efficient method of monitoring discrete points in the frequency domain of an evolving function of time. The model parameters are estimated from MDFT data using standard recursive parameter estimation techniques. The algorithm has been shown in simulations to be quite robust to additive noise in the inputs and outputs. A significant advantage of the method is that it enables a type of on-line model validation. This is accomplished by simultaneously identifying a number of models and comparing each with the plant in the frequency domain. Simulations of the method applied to an 8th-order SISO plant and a 10-state 2-input 2-output plant are presented. An example of on-line model validation applied to the SISO plant is also presented.

  16. Fine tuning points of generating function construction for linear recursions

    NASA Astrophysics Data System (ADS)

    Yolcu, Bahar; Demiralp, Metin

    2014-10-01

    Recursions are quite important mathematical tools since many systems are mathematically modelled to ultimately take us to these equations because of their rather easy algebraic natures. They fit computer programming needs quite well in many circumstances to produce solutions. However, it is generally desired to find the asymptotic behaviour of the general term in the relevant sequence for convergence and therefore practicality issues. One of the general tendencies to find the general term asymptotic behaviour, when its ordering number grows unboundedly, is the integral representation over a generating function which does not depend on individual sequence elements. This is tried to be done almost for all types of recursions, even though the linear cases gain more importance than the others because they can be more effectively investigated by using many linear algebraic tools. Despite this may seem somehow to be rather trivial, there are a lot of theoretical fine tuning issues in the construction of true integral representations over true intervals on real axis or paths in complex domains. This work is devoted to focus on this issue starting from scratch for better understanding of the matter. The example cases are chosen to best illuminate the situations to get information for future generalization even though the work can be considered at somehow introductory level.

  17. Gibbs Recursive Sampler: finding transcription factor binding sites.

    PubMed

    Thompson, William; Rouchka, Eric C; Lawrence, Charles E

    2003-07-01

    The Gibbs Motif Sampler is a software package for locating common elements in collections of biopolymer sequences. In this paper we describe a new variation of the Gibbs Motif Sampler, the Gibbs Recursive Sampler, which has been developed specifically for locating multiple transcription factor binding sites for multiple transcription factors simultaneously in unaligned DNA sequences that may be heterogeneous in DNA composition. Here we describe the basic operation of the web-based version of this sampler. The sampler may be acces-sed at http://bayesweb.wadsworth.org/gibbs/gibbs.html and at http://www.bioinfo.rpi.edu/applications/bayesian/gibbs/gibbs.html. An online user guide is available at http://bayesweb.wadsworth.org/gibbs/bernoulli.html and at http://www.bioinfo.rpi.edu/applications/bayesian/gibbs/manual/bernoulli.html. Solaris, Solaris.x86 and Linux versions of the sampler are available as stand-alone programs for academic and not-for-profit users. Commercial licenses are also available. The Gibbs Recursive Sampler is distributed in accordance with the ISCB level 0 guidelines and a requirement for citation of use in scientific publications.

  18. Applications of recursive segmentation to the analysis of DNA sequences.

    PubMed

    Li, Wentian; Bernaola-Galván, Pedro; Haghighi, Fatameh; Grosse, Ivo

    2002-07-01

    Recursive segmentation is a procedure that partitions a DNA sequence into domains with a homogeneous composition of the four nucleotides A, C, G and T. This procedure can also be applied to any sequence converted from a DNA sequence, such as to a binary strong(G + C)/weak(A + T) sequence, to a binary sequence indicating the presence or absence of the dinucleotide CpG, or to a sequence indicating both the base and the codon position information. We apply various conversion schemes in order to address the following five DNA sequence analysis problems: isochore mapping, CpG island detection, locating the origin and terminus of replication in bacterial genomes, finding complex repeats in telomere sequences, and delineating coding and noncoding regions. We find that the recursive segmentation procedure can successfully detect isochore borders, CpG islands, and the origin and terminus of replication, but it needs improvement for detecting complex repeats as well as borders between coding and noncoding regions.

  19. Recursive feature selection with significant variables of support vectors.

    PubMed

    Tsai, Chen-An; Huang, Chien-Hsun; Chang, Ching-Wei; Chen, Chun-Houh

    2012-01-01

    The development of DNA microarray makes researchers screen thousands of genes simultaneously and it also helps determine high- and low-expression level genes in normal and disease tissues. Selecting relevant genes for cancer classification is an important issue. Most of the gene selection methods use univariate ranking criteria and arbitrarily choose a threshold to choose genes. However, the parameter setting may not be compatible to the selected classification algorithms. In this paper, we propose a new gene selection method (SVM-t) based on the use of t-statistics embedded in support vector machine. We compared the performance to two similar SVM-based methods: SVM recursive feature elimination (SVMRFE) and recursive support vector machine (RSVM). The three methods were compared based on extensive simulation experiments and analyses of two published microarray datasets. In the simulation experiments, we found that the proposed method is more robust in selecting informative genes than SVMRFE and RSVM and capable to attain good classification performance when the variations of informative and noninformative genes are different. In the analysis of two microarray datasets, the proposed method yields better performance in identifying fewer genes with good prediction accuracy, compared to SVMRFE and RSVM.

  20. Cutting a Long Intron Short: Recursive Splicing and Its Implications

    PubMed Central

    Georgomanolis, Theodore; Sofiadis, Konstantinos; Papantonis, Argyris

    2016-01-01

    Over time eukaryotic genomes have evolved to host genes carrying multiple exons separated by increasingly larger intronic, mostly non-protein-coding, sequences. Initially, little attention was paid to these intronic sequences, as they were considered not to contain regulatory information. However, advances in molecular biology, sequencing, and computational tools uncovered that numerous segments within these genomic elements do contribute to the regulation of gene expression. Introns are differentially removed in a cell type-specific manner to produce a range of alternatively-spliced transcripts, and many span tens to hundreds of kilobases. Recent work in human and fruitfly tissues revealed that long introns are extensively processed cotranscriptionally and in a stepwise manner, before their two flanking exons are spliced together. This process, called “recursive splicing,” often involves non-canonical splicing elements positioned deep within introns, and different mechanisms for its deployment have been proposed. Still, the very existence and widespread nature of recursive splicing offers a new regulatory layer in the transcript maturation pathway, which may also have implications in human disease. PMID:27965595

  1. A general, recursive, and open-ended response code.

    PubMed

    Ringholm, Magnus; Jonsson, Dan; Ruud, Kenneth

    2014-03-30

    We present a new implementation of a recent open-ended response theory formulation for time- and perturbation-dependent basis sets (Thorvaldsen et al., J. Chem. Phys. 2008, 129, 214108) at the Hartree-Fock and density functional levels of theory. A novel feature of the new implementation is the use of recursive programming techniques, making it possible to write highly compact code for the analytic calculation of any response property at any valid choice of rule for the order of perturbation at which to include perturbed density matrices. The formalism is expressed in terms of the density matrix in the atomic orbital basis, allowing the recursive scheme presented here to be used in linear-scaling formulations of response theory as well as with two- and four-component relativistic wave functions. To demonstrate the new code, we present calculations of the third geometrical derivatives of the frequency-dependent second hyperpolarizability for HSOH at the Hartree-Fock level of theory, a seventh-order energy derivative involving basis sets that are both time and perturbation dependent.

  2. Optimal Recursive Digital Filters for Active Bending Stabilization

    NASA Technical Reports Server (NTRS)

    Orr, Jeb S.

    2013-01-01

    In the design of flight control systems for large flexible boosters, it is common practice to utilize active feedback control of the first lateral structural bending mode so as to suppress transients and reduce gust loading. Typically, active stabilization or phase stabilization is achieved by carefully shaping the loop transfer function in the frequency domain via the use of compensating filters combined with the frequency response characteristics of the nozzle/actuator system. In this paper we present a new approach for parameterizing and determining optimal low-order recursive linear digital filters so as to satisfy phase shaping constraints for bending and sloshing dynamics while simultaneously maximizing attenuation in other frequency bands of interest, e.g. near higher frequency parasitic structural modes. By parameterizing the filter directly in the z-plane with certain restrictions, the search space of candidate filter designs that satisfy the constraints is restricted to stable, minimum phase recursive low-pass filters with well-conditioned coefficients. Combined with optimal output feedback blending from multiple rate gyros, the present approach enables rapid and robust parametrization of autopilot bending filters to attain flight control performance objectives. Numerical results are presented that illustrate the application of the present technique to the development of rate gyro filters for an exploration-class multi-engined space launch vehicle.

  3. Document Update and Compare

    NASA Technical Reports Server (NTRS)

    Knoch, C. F.; Caldwell, D. C.; Caldwell, D. L.

    1983-01-01

    Document Update and Compare programs provide simple computerized documentmaintenance system on Data General NOVA 840 computer. Document Update program allows user to update document either by batch or terminal input. Documents are modified and lists of modifications printed out.

  4. Cognitive representation of "musical fractals": Processing hierarchy and recursion in the auditory domain.

    PubMed

    Martins, Mauricio Dias; Gingras, Bruno; Puig-Waldmueller, Estela; Fitch, W Tecumseh

    2017-04-01

    The human ability to process hierarchical structures has been a longstanding research topic. However, the nature of the cognitive machinery underlying this faculty remains controversial. Recursion, the ability to embed structures within structures of the same kind, has been proposed as a key component of our ability to parse and generate complex hierarchies. Here, we investigated the cognitive representation of both recursive and iterative processes in the auditory domain. The experiment used a two-alternative forced-choice paradigm: participants were exposed to three-step processes in which pure-tone sequences were built either through recursive or iterative processes, and had to choose the correct completion. Foils were constructed according to generative processes that did not match the previous steps. Both musicians and non-musicians were able to represent recursion in the auditory domain, although musicians performed better. We also observed that general 'musical' aptitudes played a role in both recursion and iteration, although the influence of musical training was somehow independent from melodic memory. Moreover, unlike iteration, recursion in audition was well correlated with its non-auditory (recursive) analogues in the visual and action sequencing domains. These results suggest that the cognitive machinery involved in establishing recursive representations is domain-general, even though this machinery requires access to information resulting from domain-specific processes.

  5. Round-off error propagation in four generally applicable, recursive, least-squares-estimation schemes

    NASA Technical Reports Server (NTRS)

    Verhaegen, M. H.

    1987-01-01

    The numerical robustness of four generally applicable, recursive, least-squares-estimation schemes is analyzed by means of a theoretical round-off propagation study. This study highlights a number of practical, interesting insights of widely used recursive least-squares schemes. These insights have been confirmed in an experimental study as well.

  6. The Paradigm Recursion: Is It More Accessible When Introduced in Middle School?

    ERIC Educational Resources Information Center

    Gunion, Katherine; Milford, Todd; Stege, Ulrike

    2009-01-01

    Recursion is a programming paradigm as well as a problem solving strategy thought to be very challenging to grasp for university students. This article outlines a pilot study, which expands the age range of students exposed to the concept of recursion in computer science through instruction in a series of interesting and engaging activities. In…

  7. Exploration Update

    NASA Technical Reports Server (NTRS)

    2006-01-01

    Delores Beasley, NASA Public Affairs, introduces the panel who consist of: Scott "Doc" Horowitz, Associate Administrator of Exploration Systems from NASA Headquarters; Jeff Henley, Constellation Program Manager from NASA Johnson Space Flight Center; and Steve Cook, Manager Exploration Launch Office at NASA Marshall Space Flight Center. Scott Horowitz presents a short video entitled, "Ares Launching the Future". He further explains how NASA personnel came up with the name of Ares and where the name Ares was derived. Jeff Henley, updates the Constellation program and Steve Cook presents two slide presentations detailing the Ares l crew launch vehicle and Ares 5 cargo launch vehicle. A short question and answer period from the news media follows.

  8. Using Web Services and XML Harvesting to Achieve a Dynamic Web Site. Computers in Small Libraries

    ERIC Educational Resources Information Center

    Roberts, Gary

    2005-01-01

    Exploiting and contextualizing free information is a natural part of library culture. In this column, Gary Roberts, the information systems and reference librarian at Herrick Library, Alfred University in Alfred, NY, describes how to use XML content on a Web site to link to hundreds of free and useful resources. He gives a general overview of the…

  9. Strategic Industrial Alliances in Paper Industry: XML- vs Ontology-Based Integration Platforms

    ERIC Educational Resources Information Center

    Naumenko, Anton; Nikitin, Sergiy; Terziyan, Vagan; Zharko, Andriy

    2005-01-01

    Purpose: To identify cases related to design of ICT platforms for industrial alliances, where the use of Ontology-driven architectures based on Semantic web standards is more advantageous than application of conventional modeling together with XML standards. Design/methodology/approach: A comparative analysis of the two latest and the most obvious…

  10. Applying XML-Based Technologies to Developing Online Courses: The Case of a Prototype Learning Environment

    ERIC Educational Resources Information Center

    Jedrzejowicz, Joanna; Neumann, Jakub

    2007-01-01

    Purpose: This paper seeks to describe XML technologies and to show how they can be applied for developing web-based courses and supporting authors who do not have much experience with the preparation of web-based courses. Design/methodology/approach: When developing online courses the academic staff has to address the following problem--how to…

  11. Integration of HTML documents into an XML-based knowledge repository.

    PubMed

    Roemer, Lorrie K; Rocha, Roberto A; Del Fiol, Guilherme

    2005-01-01

    The Emergency Patient Instruction Generator (EPIG) is an electronic content compiler / viewer / editor developed by Intermountain Health Care. The content is vendor-licensed HTML patient discharge instructions. This work describes the process by which discharge instructions where converted from ASCII-encoded HTML to XML, then loaded to a database for use by EPIG.

  12. A distributed computing system for magnetic resonance imaging: Java-based processing and binding of XML.

    PubMed

    de Beer, R; Graveron-Demilly, D; Nastase, S; van Ormondt, D

    2004-03-01

    Recently we have developed a Java-based heterogeneous distributed computing system for the field of magnetic resonance imaging (MRI). It is a software system for embedding the various image reconstruction algorithms that we have created for handling MRI data sets with sparse sampling distributions. Since these data sets may result from multi-dimensional MRI measurements our system has to control the storage and manipulation of large amounts of data. In this paper we describe how we have employed the extensible markup language (XML) to realize this data handling in a highly structured way. To that end we have used Java packages, recently released by Sun Microsystems, to process XML documents and to compile pieces of XML code into Java classes. We have effectuated a flexible storage and manipulation approach for all kinds of data within the MRI system, such as data describing and containing multi-dimensional MRI measurements, data configuring image reconstruction methods and data representing and visualizing the various services of the system. We have found that the object-oriented approach, possible with the Java programming environment, combined with the XML technology is a convenient way of describing and handling various data streams in heterogeneous distributed computing systems.

  13. Association Rule Extraction from XML Stream Data for Wireless Sensor Networks

    PubMed Central

    Paik, Juryon; Nam, Junghyun; Kim, Ung Mo; Won, Dongho

    2014-01-01

    With the advances of wireless sensor networks, they yield massive volumes of disparate, dynamic and geographically-distributed and heterogeneous data. The data mining community has attempted to extract knowledge from the huge amount of data that they generate. However, previous mining work in WSNs has focused on supporting simple relational data structures, like one table per network, while there is a need for more complex data structures. This deficiency motivates XML, which is the current de facto format for the data exchange and modeling of a wide variety of data sources over the web, to be used in WSNs in order to encourage the interchangeability of heterogeneous types of sensors and systems. However, mining XML data for WSNs has two challenging issues: one is the endless data flow; and the other is the complex tree structure. In this paper, we present several new definitions and techniques related to association rule mining over XML data streams in WSNs. To the best of our knowledge, this work provides the first approach to mining XML stream data that generates frequent tree items without any redundancy. PMID:25046017

  14. An Electronic Finding Aid Using Extensible Markup Language (XML) and Encoded Archival Description (EAD).

    ERIC Educational Resources Information Center

    Chang, May

    2000-01-01

    Describes the development of electronic finding aids for archives at the University of Illinois, Urbana-Champaign that used XML (extensible markup language) and EAD (encoded archival description) to enable more flexible information management and retrieval than using MARC or a relational database management system. EAD template is appended.…

  15. A Semantic Analysis of XML Schema Matching for B2B Systems Integration

    ERIC Educational Resources Information Center

    Kim, Jaewook

    2011-01-01

    One of the most critical steps to integrating heterogeneous e-Business applications using different XML schemas is schema matching, which is known to be costly and error-prone. Many automatic schema matching approaches have been proposed, but the challenge is still daunting because of the complexity of schemas and immaturity of technologies in…

  16. Enhancing the Effectiveness of Description with Innovative Tools--The Example of MidosaXML, Germany

    ERIC Educational Resources Information Center

    Menne-Haritz, Angelika

    2005-01-01

    Online description using new Internet technologies, especially XML, is an interesting way to enhance the effectiveness and productivity of archival processing. Users and archivists benefit from the developments, making the achievement of their aims easier. The provenance-based structured presentation of finding aids can be supported with many new…

  17. XML3D and Xflow: combining declarative 3D for the Web with generic data flows.

    PubMed

    Klein, Felix; Sons, Kristian; Rubinstein, Dmitri; Slusallek, Philipp

    2013-01-01

    Researchers have combined XML3D, which provides declarative, interactive 3D scene descriptions based on HTML5, with Xflow, a language for declarative, high-performance data processing. The result lets Web developers combine a 3D scene graph with data flows for dynamic meshes, animations, image processing, and postprocessing.

  18. Recursive dynamics for flexible multibody systems using spatial operators

    NASA Technical Reports Server (NTRS)

    Jain, A.; Rodriguez, G.

    1990-01-01

    Due to their structural flexibility, spacecraft and space manipulators are multibody systems with complex dynamics and possess a large number of degrees of freedom. Here the spatial operator algebra methodology is used to develop a new dynamics formulation and spatially recursive algorithms for such flexible multibody systems. A key feature of the formulation is that the operator description of the flexible system dynamics is identical in form to the corresponding operator description of the dynamics of rigid multibody systems. A significant advantage of this unifying approach is that it allows ideas and techniques for rigid multibody systems to be easily applied to flexible multibody systems. The algorithms use standard finite-element and assumed modes models for the individual body deformation. A Newton-Euler Operator Factorization of the mass matrix of the multibody system is first developed. It forms the basis for recursive algorithms such as for the inverse dynamics, the computation of the mass matrix, and the composite body forward dynamics for the system. Subsequently, an alternative Innovations Operator Factorization of the mass matrix, each of whose factors is invertible, is developed. It leads to an operator expression for the inverse of the mass matrix, and forms the basis for the recursive articulated body forward dynamics algorithm for the flexible multibody system. For simplicity, most of the development here focuses on serial chain multibody systems. However, extensions of the algorithms to general topology flexible multibody systems are described. While the computational cost of the algorithms depends on factors such as the topology and the amount of flexibility in the multibody system, in general, it appears that in contrast to the rigid multibody case, the articulated body forward dynamics algorithm is the more efficient algorithm for flexible multibody systems containing even a small number of flexible bodies. The variety of algorithms described

  19. XML-Based Generator of C++ Code for Integration With GUIs

    NASA Technical Reports Server (NTRS)

    Hua, Hook; Oyafuso, Fabiano; Klimeck, Gerhard

    2003-01-01

    An open source computer program has been developed to satisfy a need for simplified organization of structured input data for scientific simulation programs. Typically, such input data are parsed in from a flat American Standard Code for Information Interchange (ASCII) text file into computational data structures. Also typically, when a graphical user interface (GUI) is used, there is a need to completely duplicate the input information while providing it to a user in a more structured form. Heretofore, the duplication of the input information has entailed duplication of software efforts and increases in susceptibility to software errors because of the concomitant need to maintain two independent input-handling mechanisms. The present program implements a method in which the input data for a simulation program are completely specified in an Extensible Markup Language (XML)-based text file. The key benefit for XML is storing input data in a structured manner. More importantly, XML allows not just storing of data but also describing what each of the data items are. That XML file contains information useful for rendering the data by other applications. It also then generates data structures in the C++ language that are to be used in the simulation program. In this method, all input data are specified in one place only, and it is easy to integrate the data structures into both the simulation program and the GUI. XML-to-C is useful in two ways: 1. As an executable, it generates the corresponding C++ classes and 2. As a library, it automatically fills the objects with the input data values.

  20. Recursive Newton-Euler formulation of manipulator dynamics

    NASA Technical Reports Server (NTRS)

    Nasser, M. G.

    1989-01-01

    A recursive Newton-Euler procedure is presented for the formulation and solution of manipulator dynamical equations. The procedure includes rotational and translational joints and a topological tree. This model was verified analytically using a planar two-link manipulator. Also, the model was tested numerically against the Walker-Orin model using the Shuttle Remote Manipulator System data. The hinge accelerations obtained from both models were identical. The computational requirements of the model vary linearly with the number of joints. The computational efficiency of this method exceeds that of Walker-Orin methods. This procedure may be viewed as a considerable generalization of Armstrong's method. A six-by-six formulation is adopted which enhances both the computational efficiency and simplicity of the model.

  1. Spatially recursive filtering and smoothing for multibody dynamics

    NASA Technical Reports Server (NTRS)

    Rodriguez, G.

    1988-01-01

    Methods developed recently by the author to solve the problem of forward dynamics for nonlinear joint-connected multibody systems are summarized. Solution of this problem is of interest in such application areas as robotics, deploying structures, ground vehicles, and pointing of antennas and instrumented platforms. The problem is solved by the recursive filtering and smoothing techniques of state estimation theory. The filtering stage takes the applied joint moments as inputs to produce a sequence of spatial constraint forces acting at the joints of the system. The smoothing stage takes the innovations process resulting from the filter as an input and produces a set of spatial accelerations and a corresponding set of joint-angle accelerations.

  2. Structure damage detection based on random forest recursive feature elimination

    NASA Astrophysics Data System (ADS)

    Zhou, Qifeng; Zhou, Hao; Zhou, Qingqing; Yang, Fan; Luo, Linkai

    2014-05-01

    Feature extraction is a key former step in structural damage detection. In this paper, a structural damage detection method based on wavelet packet decomposition (WPD) and random forest recursive feature elimination (RF-RFE) is proposed. In order to gain the most effective feature subset and to improve the identification accuracy a two-stage feature selection method is adopted after WPD. First, the damage features are sorted according to original random forest variable importance analysis. Second, using RF-RFE to eliminate the least important feature and reorder the feature list each time, then get the new feature importance sequence. Finally, k-nearest neighbor (KNN) algorithm, as a benchmark classifier, is used to evaluate the extracted feature subset. A four-storey steel shear building model is chosen as an example in method verification. The experimental results show that using the fewer features got from proposed method can achieve higher identification accuracy and reduce the detection time cost.

  3. Direct recursive identification of the Preisach hysteresis density function

    NASA Astrophysics Data System (ADS)

    Ruderman, Michael

    2013-12-01

    In this paper, a novel direct method of recursive identification of the Preisach hysteresis density function is proposed. Using the discrete dynamic Preisach model, which is a state-space realization of the classical scalar Preisach model, the method is designed based on the output increment error. After giving the general formulation, the identification scheme implemented for a discretized Preisach plane is introduced and evaluated through the use of numerical simulations. Two cases of Gaussian mixtures are considered for mapping the hysteresis system to be identified. The parameter convergence is shown for a low-pass filtered white-noise input. Further, the proposed identification method is applied to a magnetism-related application example, where the flux linkage hysteresis of a proportional solenoid is assumed from the measurements, and then the inverse of a standard demagnetization procedure is utilized as the identification sequence.

  4. Adaptive control of large space structures using recursive lattice filters

    NASA Technical Reports Server (NTRS)

    Sundararajan, N.; Goglia, G. L.

    1985-01-01

    The use of recursive lattice filters for identification and adaptive control of large space structures is studied. Lattice filters were used to identify the structural dynamics model of the flexible structures. This identification model is then used for adaptive control. Before the identified model and control laws are integrated, the identified model is passed through a series of validation procedures and only when the model passes these validation procedures is control engaged. This type of validation scheme prevents instability when the overall loop is closed. Another important area of research, namely that of robust controller synthesis, was investigated using frequency domain multivariable controller synthesis methods. The method uses the Linear Quadratic Guassian/Loop Transfer Recovery (LQG/LTR) approach to ensure stability against unmodeled higher frequency modes and achieves the desired performance.

  5. Virasoro constraints and polynomial recursion for the linear Hodge integrals

    NASA Astrophysics Data System (ADS)

    Guo, Shuai; Wang, Gehao

    2016-11-01

    The Hodge tau-function is a generating function for the linear Hodge integrals. It is also a tau-function of the KP hierarchy. In this paper, we first present the Virasoro constraints for the Hodge tau-function in the explicit form of the Virasoro equations. The expression of our Virasoro constraints is simply a linear combination of the Virasoro operators, where the coefficients are restored from a power series for the Lambert W function. Then, using this result, we deduce a simple version of the Virasoro constraints for the linear Hodge partition function, where the coefficients are restored from the Gamma function. Finally, we establish the equivalence relation between the Virasoro constraints and polynomial recursion formula for the linear Hodge integrals.

  6. Adaptive control of large space structures using recursive lattice filters

    NASA Technical Reports Server (NTRS)

    Goglia, G. L.

    1985-01-01

    The use of recursive lattice filters for identification and adaptive control of large space structures was studied. Lattice filters are used widely in the areas of speech and signal processing. Herein, they are used to identify the structural dynamics model of the flexible structures. This identified model is then used for adaptive control. Before the identified model and control laws are integrated, the identified model is passed through a series of validation procedures and only when the model passes these validation procedures control is engaged. This type of validation scheme prevents instability when the overall loop is closed. The results obtained from simulation were compared to those obtained from experiments. In this regard, the flexible beam and grid apparatus at the Aerospace Control Research Lab (ACRL) of NASA Langley Research Center were used as the principal candidates for carrying out the above tasks. Another important area of research, namely that of robust controller synthesis, was investigated using frequency domain multivariable controller synthesis methods.

  7. Lanczos and Recursion Techniques for Multiscale Kinetic Monte Carlo Simulations

    SciTech Connect

    Rudd, R E; Mason, D R; Sutton, A P

    2006-03-13

    We review an approach to the simulation of the class of microstructural and morphological evolution involving both relatively short-ranged chemical and interfacial interactions and long-ranged elastic interactions. The calculation of the anharmonic elastic energy is facilitated with Lanczos recursion. The elastic energy changes affect the rate of vacancy hopping, and hence the rate of microstructural evolution due to vacancy mediated diffusion. The elastically informed hopping rates are used to construct the event catalog for kinetic Monte Carlo simulation. The simulation is accelerated using a second order residence time algorithm. The effect of elasticity on the microstructural development has been assessed. This article is related to a talk given in honor of David Pettifor at the DGP60 Workshop in Oxford.

  8. Virasoro constraints and polynomial recursion for the linear Hodge integrals

    NASA Astrophysics Data System (ADS)

    Guo, Shuai; Wang, Gehao

    2017-04-01

    The Hodge tau-function is a generating function for the linear Hodge integrals. It is also a tau-function of the KP hierarchy. In this paper, we first present the Virasoro constraints for the Hodge tau-function in the explicit form of the Virasoro equations. The expression of our Virasoro constraints is simply a linear combination of the Virasoro operators, where the coefficients are restored from a power series for the Lambert W function. Then, using this result, we deduce a simple version of the Virasoro constraints for the linear Hodge partition function, where the coefficients are restored from the Gamma function. Finally, we establish the equivalence relation between the Virasoro constraints and polynomial recursion formula for the linear Hodge integrals.

  9. Binary recursive partitioning: background, methods, and application to psychology.

    PubMed

    Merkle, Edgar C; Shaffer, Victoria A

    2011-02-01

    Binary recursive partitioning (BRP) is a computationally intensive statistical method that can be used in situations where linear models are often used. Instead of imposing many assumptions to arrive at a tractable statistical model, BRP simply seeks to accurately predict a response variable based on values of predictor variables. The method outputs a decision tree depicting the predictor variables that were related to the response variable, along with the nature of the variables' relationships. No significance tests are involved, and the tree's 'goodness' is judged based on its predictive accuracy. In this paper, we describe BRP methods in a detailed manner and illustrate their use in psychological research. We also provide R code for carrying out the methods.

  10. Source localization using recursively applied and projected (RAP) MUSIC

    SciTech Connect

    Mosher, J.C.; Leahy, R.M.

    1998-03-01

    A new method for source localization is described that is based on a modification of the well known multiple signal classification (MUSIC) algorithm. In classical MUSIC, the array manifold vector is projected onto an estimate of the signal subspace, but errors in the estimate can make location of multiple sources difficult. Recursively applied and projected (RAP) MUSIC uses each successively located source to form an intermediate array gain matrix, and projects both the array manifold and the signal subspace estimate into its orthogonal complement. The MUSIC projection is then performed in this reduced subspace. Using the metric of principal angles, the authors describe a general form of the RAP-MUSIC algorithm for the case of diversely polarized sources. Through a uniform linear array simulation, the authors demonstrate the improved Monte Carlo performance of RAP-MUSIC relative to MUSIC and two other sequential subspace methods, S and IES-MUSIC.

  11. A recursive-partitioning model for blood brain barrier permeation

    NASA Astrophysics Data System (ADS)

    Mente*, S. R.; Lombardo, F.

    2005-07-01

    A series of bagged recursive partitioning models for log(BB) is presented. Using a LGO-CV, three sets of physical property descriptors are evaluated and found to have Q2 values of 0.51 (CPSA), 0.53 (Ro5x) and 0.53 (MOE). Extrapolating these models to Pfizer chemical space is difficult due to P-glycoprotein (P-gp) mediated efflux. Low correlation coefficients for this test set are improved ( R 2=0.39) when compounds known to be P-gp substrates or statistical extrapolations are removed. The use of simple linear models for specific chemical series is also found to improve the correlation over a limited chemical space.

  12. Recurrent networks with recursive processing elements: paradigm for dynamical computing

    NASA Astrophysics Data System (ADS)

    Farhat, Nabil H.; del Moral Hernandez, Emilio

    1996-11-01

    It was shown earlier that models of cortical neurons can, under certain conditions of coherence in their input, behave as recursive processing elements (PEs) that are characterized by an iterative map on the phase interval and by bifurcation diagrams that demonstrate the complex encoding cortical neurons might be able to perform on their input. Here we present results of numerical experiments carried on a recurrent network of such recursive PEs modeled by the logistic map. Network behavior is studied under a novel scheme for generating complex spatio-temporal input patterns that could range from being coherent to partially coherent to being completely incoherent. A nontraditional nonlinear coupling scheme between neurons is employed to incorporate recent findings in brain science, namely that neurons use more than one kind of neurotransmitter in their chemical signaling. It is shown that such network shave the capacity to 'self-anneal' or collapse into period-m attractors that are uniquely related to the stimulus pattern following a transient 'chaotic' period during which the network searches it state-space for the associated dynamic attractor. The network accepts naturally both dynamical or stationary input patterns. Moreover we find that the use of quantized coupling strengths, introduced to reflect recent molecular biology and neurophysiological reports on synapse dynamics, endows the network with clustering ability wherein, depending ont eh stimulus pattern, PEs in the network with clustering ability wherein, depending on the stimulus pattern, PEs in the network divide into phase- locked groups with the PEs in each group being synchronized in period-m orbits. The value of m is found to be the same for all clusters and the number of clusters gives the dimension of the periodic attractor. The implications of these findings for higher-level processing such as feature- binding and for the development of novel learning algorithms are briefly discussed.

  13. Recursive random forest algorithm for constructing multilayered hierarchical gene regulatory networks that govern biological pathways

    PubMed Central

    Zhang, Kui; Busov, Victor; Wei, Hairong

    2017-01-01

    Background Present knowledge indicates a multilayered hierarchical gene regulatory network (ML-hGRN) often operates above a biological pathway. Although the ML-hGRN is very important for understanding how a pathway is regulated, there is almost no computational algorithm for directly constructing ML-hGRNs. Results A backward elimination random forest (BWERF) algorithm was developed for constructing the ML-hGRN operating above a biological pathway. For each pathway gene, the BWERF used a random forest model to calculate the importance values of all transcription factors (TFs) to this pathway gene recursively with a portion (e.g. 1/10) of least important TFs being excluded in each round of modeling, during which, the importance values of all TFs to the pathway gene were updated and ranked until only one TF was remained in the list. The above procedure, termed BWERF. After that, the importance values of a TF to all pathway genes were aggregated and fitted to a Gaussian mixture model to determine the TF retention for the regulatory layer immediately above the pathway layer. The acquired TFs at the secondary layer were then set to be the new bottom layer to infer the next upper layer, and this process was repeated until a ML-hGRN with the expected layers was obtained. Conclusions BWERF improved the accuracy for constructing ML-hGRNs because it used backward elimination to exclude the noise genes, and aggregated the individual importance values for determining the TFs retention. We validated the BWERF by using it for constructing ML-hGRNs operating above mouse pluripotency maintenance pathway and Arabidopsis lignocellulosic pathway. Compared to GENIE3, BWERF showed an improvement in recognizing authentic TFs regulating a pathway. Compared to the bottom-up Gaussian graphical model algorithm we developed for constructing ML-hGRNs, the BWERF can construct ML-hGRNs with significantly reduced edges that enable biologists to choose the implicit edges for experimental

  14. Developing an eXtensible Markup Language (XML) Application for DFO Marine Data Exchange via the Web

    DTIC Science & Technology

    2003-05-01

    XML). Le langage XML accepte le format ajouté et les restrictions de contenu dans ces éléments. Un schéma XML complet a été créé pour la structure...de contenu pour les séquences. Ces dernières étaient combinées en structures que l’on utilisait pour loger les données de profil des océans. Les...physical, chemical or biological analyses carried out on the data availability Pure Stores information about the possible release of the data to the

  15. Theory of Mind Development in Adolescence and Early Adulthood: The Growing Complexity of Recursive Thinking Ability.

    PubMed

    Valle, Annalisa; Massaro, Davide; Castelli, Ilaria; Marchetti, Antonella

    2015-02-01

    This study explores the development of theory of mind, operationalized as recursive thinking ability, from adolescence to early adulthood (N = 110; young adolescents = 47; adolescents = 43; young adults = 20). The construct of theory of mind has been operationalized in two different ways: as the ability to recognize the correct mental state of a character, and as the ability to attribute the correct mental state in order to predict the character's behaviour. The Imposing Memory Task, with five recursive thinking levels, and a third-order false-belief task with three recursive thinking levels (devised for this study) have been used. The relationship among working memory, executive functions, and linguistic skills are also analysed. Results show that subjects exhibit less understanding of elevated recursive thinking levels (third, fourth, and fifth) compared to the first and second levels. Working memory is correlated with total recursive thinking, whereas performance on the linguistic comprehension task is related to third level recursive thinking in both theory of mind tasks. An effect of age on third-order false-belief task performance was also found. A key finding of the present study is that the third-order false-belief task shows significant age differences in the application of recursive thinking that involves the prediction of others' behaviour. In contrast, such an age effect is not observed in the Imposing Memory Task. These results may support the extension of the investigation of the third order false belief after childhood.

  16. Data rate management and real time operation: recursive adaptive frame integration of limited data

    NASA Astrophysics Data System (ADS)

    Rafailov, Michael K.

    2006-08-01

    Recursive Limited Frame Integration was proposed as a way to improve frame integration performance and mitigate issues related to high data rate needed to support conventional frame integration. The technique uses two thresholds -one tuned for optimum probability of detection, the other to manage required false alarm rate, and places integration process between those thresholds. This configuration allows a non-linear integration process that, along with Signal-to-Noise Ratio (SNR) gain, provides system designers more capability where cost, weight, or power considerations limit system data rate, processing, or memory capability. However, Recursive Frame Integration Limited may have performance issues when single-frame SNR is really low. Recursive Adaptive Limited Frame Integration was proposed as a means to improve limited integration performance with really low single-frame SNR. It combines the benefits of nonlinear recursive limited frame integration and adaptive thresholds with a kind of conventional frame integration. Adding the third threshold may help in managing real time operations. In the paper the Recursive Frame Integration is presented in form of multiple parallel recursive integration. Such an approach can help not only in data rate management but in mitigation of low single frame SNR issue for Recursive Integration as well as in real time operations with frame integration.

  17. Theory of Mind Development in Adolescence and Early Adulthood: The Growing Complexity of Recursive Thinking Ability

    PubMed Central

    Valle, Annalisa; Massaro, Davide; Castelli, Ilaria; Marchetti, Antonella

    2015-01-01

    This study explores the development of theory of mind, operationalized as recursive thinking ability, from adolescence to early adulthood (N = 110; young adolescents = 47; adolescents = 43; young adults = 20). The construct of theory of mind has been operationalized in two different ways: as the ability to recognize the correct mental state of a character, and as the ability to attribute the correct mental state in order to predict the character’s behaviour. The Imposing Memory Task, with five recursive thinking levels, and a third-order false-belief task with three recursive thinking levels (devised for this study) have been used. The relationship among working memory, executive functions, and linguistic skills are also analysed. Results show that subjects exhibit less understanding of elevated recursive thinking levels (third, fourth, and fifth) compared to the first and second levels. Working memory is correlated with total recursive thinking, whereas performance on the linguistic comprehension task is related to third level recursive thinking in both theory of mind tasks. An effect of age on third-order false-belief task performance was also found. A key finding of the present study is that the third-order false-belief task shows significant age differences in the application of recursive thinking that involves the prediction of others’ behaviour. In contrast, such an age effect is not observed in the Imposing Memory Task. These results may support the extension of the investigation of the third order false belief after childhood. PMID:27247645

  18. Efficient design of two-dimensional recursive digital filters. Final report

    SciTech Connect

    Twogood, R.E.; Mitra, S.K.

    1980-01-01

    This report outlines the research progress during the period August 1978 to July 1979. This work can be divided into seven basic project areas. Project 1 deals with a comparative study of 2-D recursive and nonrecursive digital filters. The second project addresses a new design technique for 2-D half-plane recursive filters, and Projects 3 thru 5 deal with implementation issues. The sixth project presents our recent study of the applicability of array processors to 2-D digital signal processing. The final project involves our investigation into techniques for incorporating symmetry constraints on 2-D recursive filters in order to yield more efficient implementations.

  19. Development of the Plate Tectonics and Seismology markup languages with XML

    NASA Astrophysics Data System (ADS)

    Babaie, H.; Babaei, A.

    2003-04-01

    The Extensible Markup Language (XML) and its specifications such as the XSD Schema, allow geologists to design discipline-specific vocabularies such as Seismology Markup Language (SeismML) or Plate Tectonics Markup Language (TectML). These languages make it possible to store and interchange structured geological information over the Web. Development of a geological markup language requires mapping geological concepts, such as "Earthquake" or "Plate" into a UML object model, applying a modeling and design environment. We have selected four inter-related geological concepts: earthquake, fault, plate, and orogeny, and developed four XML Schema Definitions (XSD), that define the relationships, cardinalities, hierarchies, and semantics of these concepts. In such a geological concept model, the UML object "Earthquake" is related to one or more "Wave" objects, each arriving to a seismic station at a specific "DateTime", and relating to a specific "Epicenter" object that lies at a unique "Location". The "Earthquake" object occurs along a "Segment" of a "Fault" object, which is related to a specific "Plate" object. The "Fault" has its own associations with such things as "Bend", "Step", and "Segment", and could be of any kind (e.g., "Thrust", "Transform'). The "Plate" is related to many other objects such as "MOR", "Subduction", and "Forearc", and is associated with an "Orogeny" object that relates to "Deformation" and "Strain" and several other objects. These UML objects were mapped into XML Metadata Interchange (XMI) formats, which were then converted into four XSD Schemas. The schemas were used to create and validate the XML instance documents, and to create a relational database hosting the plate tectonics and seismological data in the Microsoft Access format. The SeismML and TectML allow seismologists and structural geologists, among others, to submit and retrieve structured geological data on the Internet. A seismologist, for example, can submit peer-reviewed and

  20. Recursive least squares background prediction of univariate syndromic surveillance data

    PubMed Central

    2009-01-01

    Background Surveillance of univariate syndromic data as a means of potential indicator of developing public health conditions has been used extensively. This paper aims to improve the performance of detecting outbreaks by using a background forecasting algorithm based on the adaptive recursive least squares method combined with a novel treatment of the Day of the Week effect. Methods Previous work by the first author has suggested that univariate recursive least squares analysis of syndromic data can be used to characterize the background upon which a prediction and detection component of a biosurvellance system may be built. An adaptive implementation is used to deal with data non-stationarity. In this paper we develop and implement the RLS method for background estimation of univariate data. The distinctly dissimilar distribution of data for different days of the week, however, can affect filter implementations adversely, and so a novel procedure based on linear transformations of the sorted values of the daily counts is introduced. Seven-days ahead daily predicted counts are used as background estimates. A signal injection procedure is used to examine the integrated algorithm's ability to detect synthetic anomalies in real syndromic time series. We compare the method to a baseline CDC forecasting algorithm known as the W2 method. Results We present detection results in the form of Receiver Operating Characteristic curve values for four different injected signal to noise ratios using 16 sets of syndromic data. We find improvements in the false alarm probabilities when compared to the baseline W2 background forecasts. Conclusion The current paper introduces a prediction approach for city-level biosurveillance data streams such as time series of outpatient clinic visits and sales of over-the-counter remedies. This approach uses RLS filters modified by a correction for the weekly patterns often seen in these data series, and a threshold detection algorithm from the

  1. Multi-Resolution Seismic Tomography Based on Recursive Tessellation Hierarchy

    SciTech Connect

    Simmons, N A; Myers, S C; Ramirez, A

    2009-07-01

    A 3-D global tomographic model that reconstructs velocity structure at multiple scales and incorporates laterally variable seismic discontinuities is currently being developed. The model parameterization is node-based where nodes are placed along vertices defined by triangular tessellations of a spheroidal surface. The triangular tessellation framework is hierarchical. Starting with a tetrahexahedron representing the whole globe (1st level of the hierarchy, 24 faces), they divide each triangle of the tessellation into daughter triangles. The collection of all daughter triangles comprises the 2nd level of the tessellation hierarchy and further recursion produces an arbitrary number of tessellation levels and arbitrarily fine node-spacing. They have developed an inversion procedure that takes advantage of the recursive properties of the tessellation hierarchies by progressively solving for shorter wavelength heterogeneities. In this procedure, we first perform the tomographic inversion using a tessellation level with coarse node spacing. They find that a coarse node spacing of approximately 8{sup o} is adequate to capture bulk regional properties. They then conduct the tomographic inversion on a 4{sup o} tessellation level using the residuals and inversion results from the 8{sup o} run. In practice they find that the progressive tomography approach is robust, providing an intrinsic regularization for inversion stability and avoids the issue of predefining resolution levels. Further, determining average regional properties with coarser tessellation levels enables long-wavelength heterogeneities to account for sparsely sampled regions (or regions of the mantle where longer wavelength patterns of heterogeneity suffice) while allowing shorter length-scale heterogeneities to emerge where necessary. They demonstrate the inversion approach with a set of synthetic test cases that mimic the complex nature of data arrangements (mixed-determined inversion) common to most

  2. A Parallel Implementation of Multilevel Recursive Spectral Bisection for Application to Adaptive Unstructured Meshes. Chapter 1

    NASA Technical Reports Server (NTRS)

    Barnard, Stephen T.; Simon, Horst; Lasinski, T. A. (Technical Monitor)

    1994-01-01

    The design of a parallel implementation of multilevel recursive spectral bisection is described. The goal is to implement a code that is fast enough to enable dynamic repartitioning of adaptive meshes.

  3. Kalman filtering, smoothing and recursive robot arm forward and inverse dynamics

    NASA Technical Reports Server (NTRS)

    Rodriguez, G.

    1986-01-01

    The inverse and forward dynamics problems for multi-link serial manipulators are solved by using recursive techniques from linear filtering and smoothing theory. The pivotal step is to cast the system dynamics and kinematics as a two-point boundary-value problem. Solution of this problem leads to filtering and smoothing techniques identical to the equations of Kalman filtering and Bryson-Frazier fixed time-interval smoothing. The solutions prescribe an inward filtering recursion to compute a sequence of constraint moments and forces followed by an outward recursion to determine a corresponding sequence of angular and linear accelerations. In addition to providing techniques to compute joint accelerations from applied joint moments (and vice versa), the report provides an approach to evaluate recursively the composite multi-link system inertia matrix and its inverse. The report lays the foundation for the potential use of filtering and smoothing techniques in robot inverse and forward dynamics and in robot control design.

  4. Transparent XML Binding using the ALMA Common Software (ACS) Container/Component Framework

    NASA Astrophysics Data System (ADS)

    Sommer, H.; Chiozzi, G.; Fugate, D.; Sekoranja, M.

    2004-07-01

    ALMA software, from high-level data flow applications down to instrument control, is built using the ACS framework. The common architecture and infrastructure used for the whole ALMA software is presented at this conference in (Schwarz, Farris, & Sommer 2004). ACS offers a CORBA-based container/component model and supports the exchange and persistence of XML data. For the Java programming language, the container integrates transparently the use of type-safe Java binding classes to let applications conveniently work with XML transfer objects without having to parse or serialize them. This paper will show how the ACS container/component architecture serves to pass complex data structures, such as observation meta-data, between heterogeneous applications.

  5. Representation of thermal infrared imaging data in the DICOM using XML configuration files.

    PubMed

    Ruminski, Jacek

    2007-01-01

    The DICOM standard has become a widely accepted and implemented format for the exchange and storage of medical imaging data. Different imaging modalities are supported however there is not a dedicated solution for thermal infrared imaging in medicine. In this article we propose new ideas and improvements to final proposal of the new DICOM Thermal Infrared Imaging structures and services. Additionally, we designed, implemented and tested software packages for universal conversion of existing thermal imaging files to the DICOM format using XML configuration files. The proposed solution works fast and requires minimal number of user interactions. The XML configuration file enables to compose a set of attributes for any source file format of thermal imaging camera.

  6. Enhancement of the Work in Scia Engineer's Environment by Employment of XML Programming Language

    NASA Astrophysics Data System (ADS)

    Kortiš, Ján

    2015-12-01

    The productivity of the work of engineers in the design of building structures by applying the rules of technical standards [1] has been increasing by using different software products for recent years. The software products offer engineers new possibilities to design different structures. However, there are problems especially for design of structures with similar static schemes as it is needed to follow the same work-steps. This can be more effective if the steps are done automatically by using a programming language for leading the processes that are done by software. The design process of timber structure which is done in the environment of Scia Engineer software is presented in the article. XML Programming Language is used for automatization of the design and the XML code is modified in the Excel environment by using VBA Programming language [2], [3].

  7. Using XML Configuration-Driven Development to Create a Customizable Ground Data System

    NASA Technical Reports Server (NTRS)

    Nash, Brent; DeMore, Martha

    2009-01-01

    The Mission data Processing and Control Subsystem (MPCS) is being developed as a multi-mission Ground Data System with the Mars Science Laboratory (MSL) as the first fully supported mission. MPCS is a fully featured, Java-based Ground Data System (GDS) for telecommand and telemetry processing based on Configuration-Driven Development (CDD). The eXtensible Markup Language (XML) is the ideal language for CDD because it is easily readable and editable by all levels of users and is also backed by a World Wide Web Consortium (W3C) standard and numerous powerful processing tools that make it uniquely flexible. The CDD approach adopted by MPCS minimizes changes to compiled code by using XML to create a series of configuration files that provide both coarse and fine grained control over all aspects of GDS operation.

  8. Light at Night Markup Language (LANML): XML Technology for Light at Night Monitoring Data

    NASA Astrophysics Data System (ADS)

    Craine, B. L.; Craine, E. R.; Craine, E. M.; Crawford, D. L.

    2013-05-01

    Light at Night Markup Language (LANML) is a standard, based upon XML, useful in acquiring, validating, transporting, archiving and analyzing multi-dimensional light at night (LAN) datasets of any size. The LANML standard can accommodate a variety of measurement scenarios including single spot measures, static time-series, web based monitoring networks, mobile measurements, and airborne measurements. LANML is human-readable, machine-readable, and does not require a dedicated parser. In addition LANML is flexible; ensuring future extensions of the format will remain backward compatible with analysis software. The XML technology is at the heart of communicating over the internet and can be equally useful at the desktop level, making this standard particularly attractive for web based applications, educational outreach and efficient collaboration between research groups.

  9. XTCE and XML Database Evolution and Lessons from JWST, LandSat, and Constellation

    NASA Technical Reports Server (NTRS)

    Gal-Edd, Jonathan; Kreistle, Steven; Fatig. Cirtos; Jones, Ronald

    2008-01-01

    The database organizations within three different NASA projects have advanced current practices by creating database synergy between the various spacecraft life cycle stakeholders and educating users in the benefits of the Consultative Committee for Space Data Systems (CCSDS) XML Telemetry and Command Exchange (XTCE) format. The combination of XML for managing program data and CCSDS XTCE for exchange is a robust approach that will meet all user requirements using Standards and Non proprietary tools. COTS tools for XTCEKML are very wide and varied. To combine together various low cost and free tools can be more expensive in the long run than choosing a more expensive COTS tool that meets all the needs. This was especially important when deploying in 32 remote sites with no need for licenses. A common mission XTCEKML format between dissimilar systems is possible and is not difficult. Command XMLKTCE is more complex than telemetry and the use of XTCEKML metadata to describe pages and scripts is needed due to the proprietary nature of most current ground systems. Other mission and science products such as spacecraft loads, science image catalogs, and mission operation procedures can all be described with XML as well to increase there flexibility as systems evolve and change. Figure 10 is an example of a spacecraft table load. The word is out and the XTCE community is growing, The f sXt TCE user group was held in October and in addition to ESAESOC, SC02000, and CNES identified several systems based on XTCE. The second XTCE user group is scheduled for March 10, 2008 with LDMC and others joining. As the experience with XTCE grows and the user community receives the promised benefits of using XTCE and XML the interest is growing fast.

  10. Scalable 3D GIS environment managed by 3D-XML-based modeling

    NASA Astrophysics Data System (ADS)

    Shi, Beiqi; Rui, Jianxun; Chen, Neng

    2008-10-01

    Nowadays, the namely 3D GIS technologies become a key factor in establishing and maintaining large-scale 3D geoinformation services. However, with the rapidly increasing size and complexity of the 3D models being acquired, a pressing needed for suitable data management solutions has become apparent. This paper outlines that storage and exchange of geospatial data between databases and different front ends like 3D models, GIS or internet browsers require a standardized format which is capable to represent instances of 3D GIS models, to minimize loss of information during data transfer and to reduce interface development efforts. After a review of previous methods for spatial 3D data management, a universal lightweight XML-based format for quick and easy sharing of 3D GIS data is presented. 3D data management based on XML is a solution meeting the requirements as stated, which can provide an efficient means for opening a new standard way to create an arbitrary data structure and share it over the Internet. To manage reality-based 3D models, this paper uses 3DXML produced by Dassault Systemes. 3DXML uses opening XML schemas to communicate product geometry, structure and graphical display properties. It can be read, written and enriched by standard tools; and allows users to add extensions based on their own specific requirements. The paper concludes with the presentation of projects from application areas which will benefit from the functionality presented above.

  11. WaterML: an XML Language for Communicating Water Observations Data

    NASA Astrophysics Data System (ADS)

    Maidment, D. R.; Zaslavsky, I.; Valentine, D.

    2007-12-01

    One of the great impediments to the synthesis of water information is the plethora of formats used to publish such data. Each water agency uses its own approach. XML (eXtended Markup Languages) are generalizations of Hypertext Markup Language to communicate specific kinds of information via the internet. WaterML is an XML language for water observations data - streamflow, water quality, groundwater levels, climate, precipitation and aquatic biology data, recorded at fixed, point locations as a function of time. The Hydrologic Information System project of the Consortium of Universities for the Advancement of Hydrologic Science, Inc (CUAHSI) has defined WaterML and prepared a set of web service functions called WaterOneFLow that use WaterML to provide information about observation sites, the variables measured there and the values of those measurments. WaterML has been submitted to the Open GIS Consortium for harmonization with its standards for XML languages. Academic investigators at a number of testbed locations in the WATERS network are providing data in WaterML format using WaterOneFlow web services. The USGS and other federal agencies are also working with CUAHSI to similarly provide access to their data in WaterML through WaterOneFlow services.

  12. Semi-automated XML markup of biosystematic legacy literature with the GoldenGATE editor.

    PubMed

    Sautter, Guido; Böhm, Klemens; Agosti, Donat

    2007-01-01

    Today, digitization of legacy literature is a big issue. This also applies to the domain of biosystematics, where this process has just started. Digitized biosystematics literature requires a very precise and fine grained markup in order to be useful for detailed search, data linkage and mining. However, manual markup on sentence level and below is cumbersome and time consuming. In this paper, we present and evaluate the GoldenGATE editor, which is designed for the special needs of marking up OCR output with XML. It is built in order to support the user in this process as far as possible: Its functionality ranges from easy, intuitive tagging through markup conversion to dynamic binding of configurable plug-ins provided by third parties. Our evaluation shows that marking up an OCR document using GoldenGATE is three to four times faster than with an off-the-shelf XML editor like XML-Spy. Using domain-specific NLP-based plug-ins, these numbers are even higher.

  13. Transformation of HDF-EOS metadata from the ECS model to ISO 19115-based XML

    NASA Astrophysics Data System (ADS)

    Wei, Yaxing; Di, Liping; Zhao, Baohua; Liao, Guangxuan; Chen, Aijun

    2007-02-01

    Nowadays, geographic data, such as NASA's Earth Observation System (EOS) data, are playing an increasing role in many areas, including academic research, government decisions and even in people's every lives. As the quantity of geographic data becomes increasingly large, a major problem is how to fully make use of such data in a distributed, heterogeneous network environment. In order for a user to effectively discover and retrieve the specific information that is useful, the geographic metadata should be described and managed properly. Fortunately, the emergence of XML and Web Services technologies greatly promotes information distribution across the Internet. The research effort discussed in this paper presents a method and its implementation for transforming Hierarchical Data Format (HDF)-EOS metadata from the NASA ECS model to ISO 19115-based XML, which will be managed by the Open Geospatial Consortium (OGC) Catalogue Services—Web Profile (CSW). Using XML and international standards rather than domain-specific models to describe the metadata of those HDF-EOS data, and further using CSW to manage the metadata, can allow metadata information to be searched and interchanged more widely and easily, thus promoting the sharing of HDF-EOS data.

  14. Subdivision of large introns in Drosophila by recursive splicing at nonexonic elements.

    PubMed

    Burnette, James M; Miyamoto-Sato, Etsuko; Schaub, Marc A; Conklin, Jamie; Lopez, A Javier

    2005-06-01

    Many genes with important roles in development and disease contain exceptionally long introns, but special mechanisms for their expression have not been investigated. We present bioinformatic, phylogenetic, and experimental evidence in Drosophila for a mechanism that subdivides many large introns by recursive splicing at nonexonic elements and alternative exons. Recursive splice sites predicted with highly stringent criteria are found at much higher frequency than expected in the sense strands of introns >20 kb, but they are found only at the expected frequency on the antisense strands, and they are underrepresented within introns <10 kb. The predicted sites in long introns are highly conserved between Drosophila melanogaster and Drosophila pseudoobscura, despite extensive divergence of other sequences within the same introns. These patterns of enrichment and conservation indicate that recursive splice sites are advantageous in the context of long introns. Experimental analyses of in vivo processing intermediates and lariat products from four large introns in the unrelated genes kuzbanian, outspread, and Ultrabithorax confirmed that these introns are removed by a series of recursive splicing steps using the predicted nonexonic sites. Mutation of nonexonic site RP3 within Ultrabithorax also confirmed that recursive splicing is the predominant processing pathway even with a shortened version of the intron. We discuss currently known and potential roles for recursive splicing.

  15. Some error bounds for K-iterated Gaussian recursive filters

    NASA Astrophysics Data System (ADS)

    Cuomo, Salvatore; Galletti, Ardelio; Giunta, Giulio; Marcellino, Livia

    2016-10-01

    Recursive filters (RFs) have achieved a central role in several research fields over the last few years. For example, they are used in image processing, in data assimilation and in electrocardiogram denoising. More in particular, among RFs, the Gaussian RFs are an efficient computational tool for approximating Gaussian-based convolutions and are suitable for digital image processing and applications of the scale-space theory. As is a common knowledge, the Gaussian RFs, applied to signals with support in a finite domain, generate distortions and artifacts, mostly localized at the boundaries. Heuristic and theoretical improvements have been proposed in literature to deal with this issue (namely boundary conditions). They include the case in which a Gaussian RF is applied more than once, i.e. the so called K-iterated Gaussian RFs. In this paper, starting from a summary of the comprehensive mathematical background, we consider the case of the K-iterated first-order Gaussian RF and provide the study of its numerical stability and some component-wise theoretical error bounds.

  16. Semantics boosts syntax in artificial grammar learning tasks with recursion.

    PubMed

    Fedor, Anna; Varga, Máté; Szathmáry, Eörs

    2012-05-01

    Center-embedded recursion (CER) in natural language is exemplified by sentences such as "The malt that the rat ate lay in the house." Parsing center-embedded structures is in the focus of attention because this could be one of the cognitive capacities that make humans distinct from all other animals. The ability to parse CER is usually tested by means of artificial grammar learning (AGL) tasks, during which participants have to infer the rule from a set of artificial sentences. One of the surprising results of previous AGL experiments is that learning CER is not as easy as had been thought. We hypothesized that because artificial sentences lack semantic content, semantics could help humans learn the syntax of center-embedded sentences. To test this, we composed sentences from 4 vocabularies of different degrees of semantic content due to 3 factors (familiarity, meaning of words, and semantic relationship between words). According to our results, these factors have no effect one by one but they make learning significantly faster when combined. This leads to the assumption that there were different mechanisms at work when CER was parsed in natural and in artificial languages. This finding questions the suitability of AGL tasks with artificial vocabularies for studying the learning and processing of linguistic CER.

  17. Recursive fuzzy granulation for gene subsets extraction and cancer classification.

    PubMed

    Tang, Yuchun; Zhang, Yan-Qing; Huang, Zhen; Hu, Xiaohua; Zhao, Yichuan

    2008-11-01

    A typical microarray gene expression dataset is usually both extremely sparse and imbalanced. To select multiple highly informative gene subsets for cancer classification and diagnosis, a new Fuzzy Granular Support Vector Machine---Recursive Feature Elimination algorithm (FGSVM-RFE) is designed in this paper. As a hybrid algorithm of statistical learning, fuzzy clustering, and granular computing, the FGSVM-RFE separately eliminates irrelevant, redundant, or noisy genes in different granules at different stages and selects highly informative genes with potentially different biological functions in balance. Empirical studies on three public datasets demonstrate that the FGSVM-RFE outperforms state-of-the-art approaches. Moreover, the FGSVM-RFE can extract multiple gene subsets on each of which a classifier can be modeled with 100% accuracy. Specifically, the independent testing accuracy for the prostate cancer dataset is significantly improved. The previous best result is 86% with 16 genes and our best result is 100% with only eight genes. The identified genes are annotated by Onto-Express to be biologically meaningful.

  18. Fragment-based prediction of skin sensitization using recursive partitioning.

    PubMed

    Lu, Jing; Zheng, Mingyue; Wang, Yong; Shen, Qiancheng; Luo, Xiaomin; Jiang, Hualiang; Chen, Kaixian

    2011-09-01

    Skin sensitization is an important toxic endpoint in the risk assessment of chemicals. In this paper, structure-activity relationships analysis was performed on the skin sensitization potential of 357 compounds with local lymph node assay data. Structural fragments were extracted by GASTON (GrAph/Sequence/Tree extractiON) from the training set. Eight fragments with accuracy significantly higher than 0.73 (p<0.1) were retained to make up an indicator descriptor fragment. The fragment descriptor and eight other physicochemical descriptors closely related to the endpoint were calculated to construct the recursive partitioning tree (RP tree) for classification. The balanced accuracy of the training set, test set I, and test set II in the leave-one-out model were 0.846, 0.800, and 0.809, respectively. The results highlight that fragment-based RP tree is a preferable method for identifying skin sensitizers. Moreover, the selected fragments provide useful structural information for exploring sensitization mechanisms, and RP tree creates a graphic tree to identify the most important properties associated with skin sensitization. They can provide some guidance for designing of drugs with lower sensitization level.

  19. Recursive recovery of Markov transition probabilities from boundary value data

    SciTech Connect

    Patch, Sarah Kathyrn

    1994-04-01

    In an effort to mathematically describe the anisotropic diffusion of infrared radiation in biological tissue Gruenbaum posed an anisotropic diffusion boundary value problem in 1989. In order to accommodate anisotropy, he discretized the temporal as well as the spatial domain. The probabilistic interpretation of the diffusion equation is retained; radiation is assumed to travel according to a random walk (of sorts). In this random walk the probabilities with which photons change direction depend upon their previous as well as present location. The forward problem gives boundary value data as a function of the Markov transition probabilities. The inverse problem requires finding the transition probabilities from boundary value data. Problems in the plane are studied carefully in this thesis. Consistency conditions amongst the data are derived. These conditions have two effects: they prohibit inversion of the forward map but permit smoothing of noisy data. Next, a recursive algorithm which yields a family of solutions to the inverse problem is detailed. This algorithm takes advantage of all independent data and generates a system of highly nonlinear algebraic equations. Pluecker-Grassmann relations are instrumental in simplifying the equations. The algorithm is used to solve the 4 x 4 problem. Finally, the smallest nontrivial problem in three dimensions, the 2 x 2 x 2 problem, is solved.

  20. Fragment-based prediction of skin sensitization using recursive partitioning

    NASA Astrophysics Data System (ADS)

    Lu, Jing; Zheng, Mingyue; Wang, Yong; Shen, Qiancheng; Luo, Xiaomin; Jiang, Hualiang; Chen, Kaixian

    2011-09-01

    Skin sensitization is an important toxic endpoint in the risk assessment of chemicals. In this paper, structure-activity relationships analysis was performed on the skin sensitization potential of 357 compounds with local lymph node assay data. Structural fragments were extracted by GASTON (GrAph/Sequence/Tree extractiON) from the training set. Eight fragments with accuracy significantly higher than 0.73 ( p < 0.1) were retained to make up an indicator descriptor fragment. The fragment descriptor and eight other physicochemical descriptors closely related to the endpoint were calculated to construct the recursive partitioning tree (RP tree) for classification. The balanced accuracy of the training set, test set I, and test set II in the leave-one-out model were 0.846, 0.800, and 0.809, respectively. The results highlight that fragment-based RP tree is a preferable method for identifying skin sensitizers. Moreover, the selected fragments provide useful structural information for exploring sensitization mechanisms, and RP tree creates a graphic tree to identify the most important properties associated with skin sensitization. They can provide some guidance for designing of drugs with lower sensitization level.

  1. Recursive dynamic programming for adaptive sequence and structure alignment

    SciTech Connect

    Thiele, R.; Zimmer, R.; Lengauer, T.

    1995-12-31

    We propose a new alignment procedure that is capable of aligning protein sequences and structures in a unified manner. Recursive dynamic programming (RDP) is a hierarchical method which, on each level of the hierarchy, identifies locally optimal solutions and assembles them into partial alignments of sequences and/or structures. In contrast to classical dynamic programming, RDP can also handle alignment problems that use objective functions not obeying the principle of prefix optimality, e.g. scoring schemes derived from energy potentials of mean force. For such alignment problems, RDP aims at computing solutions that are near-optimal with respect to the involved cost function and biologically meaningful at the same time. Towards this goal, RDP maintains a dynamic balance between different factors governing alignment fitness such as evolutionary relationships and structural preferences. As in the RDP method gaps are not scored explicitly, the problematic assignment of gap cost parameters is circumvented. In order to evaluate the RDP approach we analyse whether known and accepted multiple alignments based on structural information can be reproduced with the RDP method.

  2. Implicit Learning of Recursive Context-Free Grammars

    PubMed Central

    Rohrmeier, Martin; Fu, Qiufang; Dienes, Zoltan

    2012-01-01

    Context-free grammars are fundamental for the description of linguistic syntax. However, most artificial grammar learning experiments have explored learning of simpler finite-state grammars, while studies exploring context-free grammars have not assessed awareness and implicitness. This paper explores the implicit learning of context-free grammars employing features of hierarchical organization, recursive embedding and long-distance dependencies. The grammars also featured the distinction between left- and right-branching structures, as well as between centre- and tail-embedding, both distinctions found in natural languages. People acquired unconscious knowledge of relations between grammatical classes even for dependencies over long distances, in ways that went beyond learning simpler relations (e.g. n-grams) between individual words. The structural distinctions drawn from linguistics also proved important as performance was greater for tail-embedding than centre-embedding structures. The results suggest the plausibility of implicit learning of complex context-free structures, which model some features of natural languages. They support the relevance of artificial grammar learning for probing mechanisms of language learning and challenge existing theories and computational models of implicit learning. PMID:23094021

  3. Implicit learning of recursive context-free grammars.

    PubMed

    Rohrmeier, Martin; Fu, Qiufang; Dienes, Zoltan

    2012-01-01

    Context-free grammars are fundamental for the description of linguistic syntax. However, most artificial grammar learning experiments have explored learning of simpler finite-state grammars, while studies exploring context-free grammars have not assessed awareness and implicitness. This paper explores the implicit learning of context-free grammars employing features of hierarchical organization, recursive embedding and long-distance dependencies. The grammars also featured the distinction between left- and right-branching structures, as well as between centre- and tail-embedding, both distinctions found in natural languages. People acquired unconscious knowledge of relations between grammatical classes even for dependencies over long distances, in ways that went beyond learning simpler relations (e.g. n-grams) between individual words. The structural distinctions drawn from linguistics also proved important as performance was greater for tail-embedding than centre-embedding structures. The results suggest the plausibility of implicit learning of complex context-free structures, which model some features of natural languages. They support the relevance of artificial grammar learning for probing mechanisms of language learning and challenge existing theories and computational models of implicit learning.

  4. Description of an exact recursive method to simplify shading calculations

    SciTech Connect

    Nawrocki, A.D.; Kammerud, R.

    1981-04-01

    An exact, recursive method called SHADE is described which attempts to simplify shading calculations as performed by a programmable calculator or microcomputer. Preliminary applications of SHADE using a Hewlett Packard HP-41C programmable calculation are outlined. In particular, for a given solar hour, SHADE is used to compute the following quantities for overhang and side fin combinations which shade various openings: the percentage of the total area of the opening which is shaded; the shaded area itself; the cosine of the angle of incidence between sun and glazing surface; the direct insolation at this surface, with and without shading; and the direct solar power at this surface, with and without shading. Hence, in its present HP-41C application, SHADE can be used in preliminary design and comparative analyses of shading devices on an hourly, daily, or seasonal basis, provided that: (1) the fins and overhangs be square or rectangular, and lie in planes perpendicular to the plane of the opening; and (2) the opening itself be vertical and rectangular with arbitrary building azimuths. Design candidates include conventional overhangs and side fins, porches, and reveals. In principle, SHADE can be extended to awnings, slatted sun screens, and bevelled recesses; in addition, its HP-41C application can be extended to calculations of direct solar gain through vertical and non-vertical glazings, thereby providing a more useful tool in building heating and cooling load calculations.

  5. Updating Situation Models

    ERIC Educational Resources Information Center

    Zwaan, Rolf A.; Madden, Carol J.

    2004-01-01

    The authors examined how situation models are updated during text comprehension. If comprehenders keep track of the evolving situation, they should update their models such that the most current information, the here and now, is more available than outdated information. Contrary to this updating hypothesis, E. J. O'Brien, M. L. Rizzella, J. E.…

  6. Recursive Subsystems in Aphasia and Alzheimer's Disease: Case Studies in Syntax and Theory of Mind

    PubMed Central

    Bánréti, Zoltán; Hoffmann, Ildikó; Vincze, Veronika

    2016-01-01

    The relationship between recursive sentence embedding and theory-of-mind (ToM) inference is investigated in three persons with Broca's aphasia, two persons with Wernicke's aphasia, and six persons with mild and moderate Alzheimer's disease (AD). We asked questions of four types about photographs of various real-life situations. Type 4 questions asked participants about intentions, thoughts, or utterances of the characters in the pictures (“What may X be thinking/asking Y to do?”). The expected answers typically involved subordinate clauses introduced by conjunctions or direct quotations of the characters' utterances. Broca's aphasics did not produce answers with recursive sentence embedding. Rather, they projected themselves into the characters' mental states and gave direct answers in the first person singular, with relevant ToM content. We call such replies “situative statements.” Where the question concerned the mental state of the character but did not require an answer with sentence embedding (“What does X hate?”), aphasics gave descriptive answers rather than situative statements. Most replies given by persons with AD to Type 4 questions were grammatical instances of recursive sentence embedding. They also gave a few situative statements but the ToM content of these was irrelevant. In more than one third of their well-formed sentence embeddings, too, they conveyed irrelevant ToM contents. Persons with moderate AD were unable to pass secondary false belief tests. The results reveal double dissociation: Broca's aphasics are unable to access recursive sentence embedding but they can make appropriate ToM inferences; moderate AD persons make the wrong ToM inferences but they are able to access recursive sentence embedding. The double dissociation may be relevant for the nature of the relationship between the two recursive capacities. Broca's aphasics compensated for the lack of recursive sentence embedding by recursive ToM reasoning represented in very

  7. Multi-Layer and Recursive Neural Networks for Metagenomic Classification.

    PubMed

    Ditzler, Gregory; Polikar, Robi; Rosen, Gail

    2015-09-01

    Recent advances in machine learning, specifically in deep learning with neural networks, has made a profound impact on fields such as natural language processing, image classification, and language modeling; however, feasibility and potential benefits of the approaches to metagenomic data analysis has been largely under-explored. Deep learning exploits many layers of learning nonlinear feature representations, typically in an unsupervised fashion, and recent results have shown outstanding generalization performance on previously unseen data. Furthermore, some deep learning methods can also represent the structure in a data set. Consequently, deep learning and neural networks may prove to be an appropriate approach for metagenomic data. To determine whether such approaches are indeed appropriate for metagenomics, we experiment with two deep learning methods: i) a deep belief network, and ii) a recursive neural network, the latter of which provides a tree representing the structure of the data. We compare these approaches to the standard multi-layer perceptron, which has been well-established in the machine learning community as a powerful prediction algorithm, though its presence is largely missing in metagenomics literature. We find that traditional neural networks can be quite powerful classifiers on metagenomic data compared to baseline methods, such as random forests. On the other hand, while the deep learning approaches did not result in improvements to the classification accuracy, they do provide the ability to learn hierarchical representations of a data set that standard classification methods do not allow. Our goal in this effort is not to determine the best algorithm in terms accuracy-as that depends on the specific application-but rather to highlight the benefits and drawbacks of each of the approach we discuss and provide insight on how they can be improved for predictive metagenomic analysis.

  8. Development of Updated ABsorption SIMulation Software (ABSIM)

    SciTech Connect

    Yang, Zhiyao; Tang, Xin; Qu, Ming; Abdelaziz, Omar; Gluesenkamp, Kyle R

    2014-01-01

    ABsorption SIMulation, ABSIM, was developed for the simulation of absorption systems by The Oak Ridge National Laboratory during 1980s and 1990s. ABSIM provides a platform for users to investigate various cycle configurations and working fluids, to calculate their operating parameters, to predict their performance, and to compare them with each other on a uniform basis. ABSIM is indeed a very useful and accurate tool for researchers to investigate various absorption systems. However, it has not been well maintained: it is incompatible with recent operating systems; the interface needs improved user-friendliness, and the system needs better parameter setting and debugging tools to help achieve convergence. Therefore, it is highly needed to update and improve ABSIM. The paper presents recent efforts to improve ABSIM s compatibility with current operating systems, user interface, and analysis capabilities. The paper details the features and functions of the newly updated ABSIM software. The new ABSIM still uses the previously validated calculation engine of the old ABSIM. The new graphic user interfaces (GUI) were developed in Qt, which is an open source license GUI software based on C++. XML was used as the database for data storage in the new ABSIM. The new ABSIM has been designed to be easily learned and used. It has enhanced editing and construction functions, plus enhanced analysis features including parametric tables, plotting, property plots, and master panels for debugging. A single effect water/LiBr absorption system is used as a case study in this paper to illustrate the features, capabilities, and functions of the new ABSIM. This case study was actually an example system available in the old ABSIM. The new version of ABSIM will be continuously developed to include additional subroutines for the components in liquid desiccant systems. The new ABSIM will be available to public for free. The ultimate goal of the new ABSIM is to allow it to become a simulation

  9. Web-Based Delivery System for Disaster Prevention Information Using a New Jma Dpi Xml Format and Amedas Data

    NASA Astrophysics Data System (ADS)

    Nishio, M.; Mori, M.

    2012-07-01

    The Automated Meteorological Data Acquisition System (AMeDAS) Data is used along with compound disaster information for a geographic information system (GIS) by integration into the Japan Meteorological Agency (JMA) disaster prevention information XML data. A JMA XML format is a next generation format that contains weather warnings, tsunami warnings, and earthquake information, etc. However, it is not possible to process it by reading disaster prevention information XML Data and AMeDAS Data directly to the GIS system. Therefore, development of a program that converts the data structure is important to consolidate a variety of disaster prevention information on the GIS system. Information on escape routes and evacuation sites, etc. were given as points for regional meteorological observation forecasts using AMeDAS Data by disaster prevention information XML data and integrating it where the disaster was generated, giving a range of expansion of damage and a damage level. There are two main aims; the first is to deliver these compound data of disaster prevention information XML data and AMeDAS Data via the Internet. The second aim is to provide GIS files (shapefile format) of these data to such as local governments for their individual analysis. This was furthermore confirmed to enable the construction of a system using WebGIS (Google Maps) and Open Source Software GIS to monitor disaster information at low cost.

  10. Mining IP to Domain Name Interactions to Detect DNS Flood Attacks on Recursive DNS Servers.

    PubMed

    Alonso, Roberto; Monroy, Raúl; Trejo, Luis A

    2016-08-17

    The Domain Name System (DNS) is a critical infrastructure of any network, and, not surprisingly a common target of cybercrime. There are numerous works that analyse higher level DNS traffic to detect anomalies in the DNS or any other network service. By contrast, few efforts have been made to study and protect the recursive DNS level. In this paper, we introduce a novel abstraction of the recursive DNS traffic to detect a flooding attack, a kind of Distributed Denial of Service (DDoS). The crux of our abstraction lies on a simple observation: Recursive DNS queries, from IP addresses to domain names, form social groups; hence, a DDoS attack should result in drastic changes on DNS social structure. We have built an anomaly-based detection mechanism, which, given a time window of DNS usage, makes use of features that attempt to capture the DNS social structure, including a heuristic that estimates group composition. Our detection mechanism has been successfully validated (in a simulated and controlled setting) and with it the suitability of our abstraction to detect flooding attacks. To the best of our knowledge, this is the first time that work is successful in using this abstraction to detect these kinds of attacks at the recursive level. Before concluding the paper, we motivate further research directions considering this new abstraction, so we have designed and tested two additional experiments which exhibit promising results to detect other types of anomalies in recursive DNS servers.

  11. The language faculty that wasn't: a usage-based account of natural language recursion

    PubMed Central

    Christiansen, Morten H.; Chater, Nick

    2015-01-01

    In the generative tradition, the language faculty has been shrinking—perhaps to include only the mechanism of recursion. This paper argues that even this view of the language faculty is too expansive. We first argue that a language faculty is difficult to reconcile with evolutionary considerations. We then focus on recursion as a detailed case study, arguing that our ability to process recursive structure does not rely on recursion as a property of the grammar, but instead emerges gradually by piggybacking on domain-general sequence learning abilities. Evidence from genetics, comparative work on non-human primates, and cognitive neuroscience suggests that humans have evolved complex sequence learning skills, which were subsequently pressed into service to accommodate language. Constraints on sequence learning therefore have played an important role in shaping the cultural evolution of linguistic structure, including our limited abilities for processing recursive structure. Finally, we re-evaluate some of the key considerations that have often been taken to require the postulation of a language faculty. PMID:26379567

  12. Recursive construction and error correction of DNA molecules and libraries from synthetic and natural DNA.

    PubMed

    Yehezkel, Tuval Ben; Linshiz, Gregory; Kaplan, Shai; Gronau, Ilan; Ravid, Sivan; Adar, Rivka; Shapiro, Ehud

    2011-01-01

    Making error-free, custom DNA assemblies from potentially faulty building blocks is a fundamental challenge in synthetic biology. Here, we show how recursion can be used to address this challenge using a recursive procedure that constructs error-free DNA molecules and their libraries from error-prone synthetic oligonucleotides and naturally existing DNA. Specifically, we describe how divide and conquer (D&C), the quintessential recursive problem-solving technique, is applied in silico to divide target DNA sequences into overlapping, albeit error prone, oligonucleotides, and how recursive construction is applied in vitro to combine them to form error-prone DNA molecules. To correct DNA sequence errors, error-free fragments of these molecules are then identified, extracted, and used as new, typically longer and more accurate, inputs to another iteration of the recursive construction procedure; the entire process repeats until an error-free target molecule is formed. The method allows combining synthetic and natural DNA fragments into error-free designer DNA libraries, thus providing a foundation for the design and construction of complex synthetic DNA assemblies.

  13. Thermodynamics and Phase Transitions of Ising Model on Inhomogeneous Stochastic Recursive Lattice

    NASA Astrophysics Data System (ADS)

    Huang, Ran

    As one of the few exactly solvable thermodynamic models, the Ising model on recursive lattice is featured by its impressive advantages and successful applications in various thermodynamic and statistical researches. However this model was considered that, since the recursive calculation demands homogeneous structure, it can only describe the bulk and even systems with narrow utilization. In this work we figured out a practical methodology to extend the conventional homogeneous structure of single-unit Husimi lattice to be random inhomogeneous lattices with variable units and structures, while keeping the feature of exact calculation. Three designs of inhomogeneous recursive lattices: the random-angled rhombus lattice, the Husimi lattice of variable units, and the randomly multi-branched Husimi square lattice; and the corresponding exact recursive calculations based on the partial partition function algorithm, which is derived from the Bethe Cavity method, have been investigated and developed. With the ``total-symmetry assumption'' and the ``iterative-replica trick'' we were able to exactly solve the classical ferromagnetic spin-1 Ising models on these lattices, to describe the complex systems that can only be solved by approximations or simulations on regular lattices. Our work may enhance the application of the exact calculation on recursive lattices in various fields of materials science and applied physics, especially it may serve as a powerful tool to explore the cross-dimensional thermodynamics and phase transitions. National Natural Science Foundation of China (Grant No. 11505110).

  14. Mining IP to Domain Name Interactions to Detect DNS Flood Attacks on Recursive DNS Servers

    PubMed Central

    Alonso, Roberto; Monroy, Raúl; Trejo, Luis A.

    2016-01-01

    The Domain Name System (DNS) is a critical infrastructure of any network, and, not surprisingly a common target of cybercrime. There are numerous works that analyse higher level DNS traffic to detect anomalies in the DNS or any other network service. By contrast, few efforts have been made to study and protect the recursive DNS level. In this paper, we introduce a novel abstraction of the recursive DNS traffic to detect a flooding attack, a kind of Distributed Denial of Service (DDoS). The crux of our abstraction lies on a simple observation: Recursive DNS queries, from IP addresses to domain names, form social groups; hence, a DDoS attack should result in drastic changes on DNS social structure. We have built an anomaly-based detection mechanism, which, given a time window of DNS usage, makes use of features that attempt to capture the DNS social structure, including a heuristic that estimates group composition. Our detection mechanism has been successfully validated (in a simulated and controlled setting) and with it the suitability of our abstraction to detect flooding attacks. To the best of our knowledge, this is the first time that work is successful in using this abstraction to detect these kinds of attacks at the recursive level. Before concluding the paper, we motivate further research directions considering this new abstraction, so we have designed and tested two additional experiments which exhibit promising results to detect other types of anomalies in recursive DNS servers. PMID:27548169

  15. Recursive least squares estimation and Kalman filtering by systolic arrays

    NASA Technical Reports Server (NTRS)

    Chen, M. J.; Yao, K.

    1988-01-01

    One of the most promising new directions for high-throughput-rate problems is that based on systolic arrays. In this paper, using the matrix-decomposition approach, a systolic Kalman filter is formulated as a modified square-root information filter consisting of a whitening filter followed by a simple least-squares operation based on the systolic QR algorithm. By proper skewing of the input data, a fully pipelined time and measurement update systolic Kalman filter can be achieved with O(n squared) processing cells, resulting in a system throughput rate of O (n).

  16. Event Compression Using Recursive Least Squares Signal Processing.

    DTIC Science & Technology

    1980-07-01

    addition, he requires knowledge of the data’s noise statistics . From these assumptions he derives a likelihood ratio for the beginning of an event at each... statistics are not known, detailed information about the events is not available or the number of available exponentials from which they could be composed...combining eqs. (2.23), (2.24) and (2.25b) and it shows how to update the state matrix V. T V = V - r (2.28) Tl+r Yr 23 - The full RLS iteration procedes

  17. Lapin Data Interchange Among Database, Analysis and Display Programs Using XML-Based Text Files

    NASA Technical Reports Server (NTRS)

    2005-01-01

    The purpose of grant NCC3-966 was to investigate and evaluate the interchange of application-specific data among multiple programs each carrying out part of the analysis and design task. This has been carried out previously by creating a custom program to read data produced by one application and then write that data to a file whose format is specific to the second application that needs all or part of that data. In this investigation, data of interest is described using the XML markup language that allows the data to be stored in a text-string. Software to transform output data of a task into an XML-string and software to read an XML string and extract all or a portion of the data needed for another application is used to link two independent applications together as part of an overall design effort. This approach was initially used with a standard analysis program, Lapin, along with standard applications a standard spreadsheet program, a relational database program, and a conventional dialog and display program to demonstrate the successful sharing of data among independent programs. Most of the effort beyond that demonstration has been concentrated on the inclusion of more complex display programs. Specifically, a custom-written windowing program organized around dialogs to control the interactions have been combined with an independent CAD program (Open Cascade) that supports sophisticated display of CAD elements such as lines, spline curves, and surfaces and turbine-blade data produced by an independent blade design program (UD0300).

  18. Lapin Data Interchange Among Database, Analysis and Display Programs Using XML-Based Text Files

    NASA Technical Reports Server (NTRS)

    2004-01-01

    The purpose was to investigate and evaluate the interchange of application- specific data among multiple programs each carrying out part of the analysis and design task. This has been carried out previously by creating a custom program to read data produced by one application and then write that data to a file whose format is specific to the second application that needs all or part of that data. In this investigation, data of interest is described using the XML markup language that allows the data to be stored in a text-string. Software to transform output data of a task into an XML-string and software to read an XML string and extract all or a portion of the data needed for another application is used to link two independent applications together as part of an overall design effort. This approach was initially used with a standard analysis program, Lapin, along with standard applications a standard spreadsheet program, a relational database program, and a conventional dialog and display program to demonstrate the successful sharing of data among independent programs. See Engineering Analysis Using a Web-Based Protocol by J.D. Schoeffler and R.W. Claus, NASA TM-2002-211981, October 2002. Most of the effort beyond that demonstration has been concentrated on the inclusion of more complex display programs. Specifically, a custom-written windowing program organized around dialogs to control the interactions have been combined with an independent CAD program (Open Cascade) that supports sophisticated display of CAD elements such as lines, spline curves, and surfaces and turbine-blade data produced by an independent blade design program (UD0300).

  19. Agent Based Manufacturing Capability Assessment in the Extended Enterprise Using Step AP224 and XML

    NASA Astrophysics Data System (ADS)

    Ratchev, Svetan; Medani, Omar

    Data exchange in the extended enterprise is one of the most critical tasks in supporting collaborative decision-making. Companies often rely on geographically distributed suppliers for efficient product design and manufacture. Early design assessment can substantially reduce the cost of product development and production. This research proposes a new STEP AP224 EXPRESS based data model to facilitate the exchange of part and process data between distributed key agents in the early design process. The approach is illustrated using a prototype XML/CORBA environment to support the information exchange between collaborating design and manufacturing agents.

  20. A Transcription System from MusicXML Format to Braille Music Notation

    NASA Astrophysics Data System (ADS)

    Goto, D.; Gotoh, T.; Minamikawa-Tachino, R.; Tamura, N.

    2006-12-01

    The Internet enables us to freely access music as recorded sound and even music scores. For the visually impaired, music scores must be transcribed from computer-based musical formats to Braille music notation. This paper proposes a transcription system from the MusicXML format to Braille music notation using a structural model of Braille music notation. The resultant Braille scores inspected by volunteer transcribers are up to the international standard. Using this simple and efficient transcription system, it should be possible to provide Braille music scores via the Internet to the visually impaired.

  1. Mapping XML documents into databases: a Data-Driven Framework for bioinformatic data interchange.

    PubMed

    Canfield, K; Sorace, J

    2000-01-01

    The Data-Driven Framework (DDF) described here addresses two major problems for healthcare Electronic Data Interchange, data formats and software development costs. The use of a standard XML Document Type Definition (DTD) allows robust representation in any application area and leverages industry-standard tools and development directions. The DDF allows reduced software development and maintenance costs since all data-entry and database tools are generated from the DTD. The DTD can change and the tools can be regenerated. The case-study below uses the DDF for reporting cell assays to determine the roles of factors influencing cellular gene expression and regulation.

  2. [Applications of tree-shaped structure in DICOM based on XML].

    PubMed

    Yao, Zhi-Hong; Pan, Min-Hui

    2006-03-01

    In order to improve the sharing of DICOM, a proposal is offered to introduce XML into DICOM on the SPIE 2005 and in this way, better readability and expansibility of DICOM files have been achieved. On this basis, the DICOM data structure is improved further by replacing the one-dimension linear data structure with multilayer tree-shaped data structure. And thus the nesting and dependence relationships between information object and data elements are much clearer and it is more convenient and easy to browse DICOM image files.

  3. Real time hybrid simulation with online model updating: An analysis of accuracy

    NASA Astrophysics Data System (ADS)

    Ou, Ge; Dyke, Shirley J.; Prakash, Arun

    2017-02-01

    In conventional hybrid simulation (HS) and real time hybrid simulation (RTHS) applications, the information exchanged between the experimental substructure and numerical substructure is typically restricted to the interface boundary conditions (force, displacement, acceleration, etc.). With additional demands being placed on RTHS and recent advances in recursive system identification techniques, an opportunity arises to improve the fidelity by extracting information from the experimental substructure. Online model updating algorithms enable the numerical model of components (herein named the target model), that are similar to the physical specimen to be modified accordingly. This manuscript demonstrates the power of integrating a model updating algorithm into RTHS (RTHSMU) and explores the possible challenges of this approach through a practical simulation. Two Bouc-Wen models with varying levels of complexity are used as target models to validate the concept and evaluate the performance of this approach. The constrained unscented Kalman filter (CUKF) is selected for using in the model updating algorithm. The accuracy of RTHSMU is evaluated through an estimation output error indicator, a model updating output error indicator, and a system identification error indicator. The results illustrate that, under applicable constraints, by integrating model updating into RTHS, the global response accuracy can be improved when the target model is unknown. A discussion on model updating parameter sensitivity to updating accuracy is also presented to provide guidance for potential users.

  4. Diesel Engine Technology Update

    DTIC Science & Technology

    1987-07-01

    AFWAL-TR-87-20 54 83-021-DET DIESEL ENGINE TECHNOLOGY UPDATE Kaupert, Andrew W., Lt. Col. USAFR Air Force Reserves Detroit Detachment 2 Ann Arbor, MI...nn AFR OH 45433-6563 63723F 3139 1 01 01 11. TITLE (Include Security Classification) DIESEL ENGINE TECHNOLOGY UPDATE 12. PERSONAL AUTHOR(S) Kaupert...methodology for technology prediction. The objective of the present report is to update the technology transfer/ 0 development status of diesel engine

  5. Puzzles in modern biology. II. Language, cancer and the recursive processes of evolutionary innovation.

    PubMed

    Frank, Steven A

    2016-01-01

    Human language emerged abruptly. Diverse body forms evolved suddenly. Seed-bearing plants spread rapidly. How do complex evolutionary innovations arise so quickly? Resolving alternative claims remains difficult. The great events of the past happened a long time ago. Cancer provides a model to study evolutionary innovation. A tumor must evolve many novel traits to become an aggressive cancer. I use what we know or could study about cancer to describe the key processes of innovation. In general, evolutionary systems form a hierarchy of recursive processes. Those recursive processes determine the rates at which innovations are generated, spread and transmitted. I relate the recursive processes to abrupt evolutionary innovation.

  6. Puzzles in modern biology. II. Language, cancer and the recursive processes of evolutionary innovation

    PubMed Central

    Frank, Steven A.

    2016-01-01

    Human language emerged abruptly. Diverse body forms evolved suddenly. Seed-bearing plants spread rapidly. How do complex evolutionary innovations arise so quickly? Resolving alternative claims remains difficult. The great events of the past happened a long time ago. Cancer provides a model to study evolutionary innovation. A tumor must evolve many novel traits to become an aggressive cancer. I use what we know or could study about cancer to describe the key processes of innovation. In general, evolutionary systems form a hierarchy of recursive processes. Those recursive processes determine the rates at which innovations are generated, spread and transmitted. I relate the recursive processes to abrupt evolutionary innovation. PMID:28184282

  7. CliniViewer: a tool for viewing electronic medical records based on natural language processing and XML.

    PubMed

    Liu, Hongfang; Friedman, Carol

    2004-01-01

    With the evolving use of computers in healthcare, the electronic medical record (EMR) is becoming more and more popular. A tool is needed that would enable physicians to accurately and efficiently access clinical information in multiple medical records associated with a particular patient. Both natural language processing (NLP) and the eXtensible Markup Language (XML) have been used in the clinical domain for capturing, representing, and utilizing clinical information and both have shown great potential. In this paper, we demonstrate another use of XML and NLP through CliniViewer, a tool that organizes and presents the clinical information in multiple records. We also describe the flexibility and capability provided when combining XML and NLP to summarize, navigate, and conceptualize structured information. The tool has been fully implemented and tested using patients with multiple discharge summaries.

  8. ISS Update: Suitport

    NASA Video Gallery

    ISS Update commentator Lynnette Madison interviews Mallory Jennings, Suitport Human Testing Lead, about making spacewalks easier and more efficient with the Suitport. Questions? Ask us on Twitter @...

  9. Methods for assessing movement path recursion with application to African buffalo in South Africa

    USGS Publications Warehouse

    Bar-David, S.; Bar-David, I.; Cross, P.C.; Ryan, S.J.; Knechtel, C.U.; Getz, W.M.

    2009-01-01

    Recent developments of automated methods for monitoring animal movement, e.g., global positioning systems (GPS) technology, yield high-resolution spatiotemporal data. To gain insights into the processes creating movement patterns, we present two new techniques for extracting information from these data on repeated visits to a particular site or patch ("recursions"). Identification of such patches and quantification of recursion pathways, when combined with patch-related ecological data, should contribute to our understanding of the habitat requirements of large herbivores, of factors governing their space-use patterns, and their interactions with the ecosystem. We begin by presenting output from a simple spatial model that simulates movements of large-herbivore groups based on minimal parameters: resource availability and rates of resource recovery after a local depletion. We then present the details of our new techniques of analyses (recursion analysis and circle analysis) and apply them to data generated by our model, as well as two sets of empirical data on movements of African buffalo (Syncerus coffer): the first collected in Klaserie Private Nature Reserve and the second in Kruger National Park, South Africa. Our recursion analyses of model outputs provide us with a basis for inferring aspects of the processes governing the production of buffalo recursion patterns, particularly the potential influence of resource recovery rate. Although the focus of our simulations was a comparison of movement patterns produced by different resource recovery rates, we conclude our paper with a comprehensive discussion of how recursion analyses can be used when appropriate ecological data are available to elucidate various factors influencing movement. Inter alia, these include the various limiting and preferred resources, parasites, and topographical and landscape factors. ?? 2009 by the Ecological Society of America.

  10. Direct phase-domain calculation of transmission line transients using two-sided recursions

    SciTech Connect

    Angelidis, G.; Semlyen, A.

    1995-04-01

    This paper presents a new method for the simulation of electromagnetic transients on transmission lines. Instead of using convolutions of the input variables only, the authors perform short convolutions with both input and output variables. The result is a method of Two-Sided Recursions (TSR), which is comparable in efficiency with the existing recursive convolutions or with their equivalent state variable formulations. It is, however, conceptually simpler and can be applied, in addition to fast modal-domain solutions, to the direct phase-domain calculation of transmission line transients with very accurate results.

  11. Kalman filtering, smoothing, and recursive robot arm forward and inverse dynamics

    NASA Technical Reports Server (NTRS)

    Rodriguez, Guillermo

    1987-01-01

    The recursive difference equations of Kalman filtering and Bryson-Frazier fixed time-interval smoothing, arising in the state estimation theory for linear state space systems, are used here to solve problems of serial manipulator inverse and forward dynamics. The configuration analyzed is that of a joint connected N-link serial manipulator attached to an immobile base. The equivalence between the filtering and smoothing techniques from state estimation theory and recursive robot dynamics methods is demonstrated. Several areas for future research are suggested.

  12. On the Hosoya index of a family of deterministic recursive trees

    NASA Astrophysics Data System (ADS)

    Chen, Xufeng; Zhang, Jingyuan; Sun, Weigang

    2017-01-01

    In this paper, we calculate the Hosoya index in a family of deterministic recursive trees with a special feature that includes new nodes which are connected to existing nodes with a certain rule. We then obtain a recursive solution of the Hosoya index based on the operations of a determinant. The computational complexity of our proposed algorithm is O(log2 n) with n being the network size, which is lower than that of the existing numerical methods. Finally, we give a weighted tree shrinking method as a graphical interpretation of the recurrence formula for the Hosoya index.

  13. Recursive analytical solution describing artificial satellite motion perturbed by an arbitrary number of zonal terms

    NASA Technical Reports Server (NTRS)

    Mueller, A. C.

    1977-01-01

    An analytical first order solution has been developed which describes the motion of an artificial satellite perturbed by an arbitrary number of zonal harmonics of the geopotential. A set of recursive relations for the solution, which was deduced from recursive relations of the geopotential, was derived. The method of solution is based on Von-Zeipel's technique applied to a canonical set of two-body elements in the extended phase space which incorporates the true anomaly as a canonical element. The elements are of Poincare type, that is, they are regular for vanishing eccentricities and inclinations. Numerical results show that this solution is accurate to within a few meters after 500 revolutions.

  14. High-rate systematic recursive convolutional encoders: minimal trellis and code search

    NASA Astrophysics Data System (ADS)

    Benchimol, Isaac; Pimentel, Cecilio; Souza, Richard Demo; Uchôa-Filho, Bartolomeu F.

    2012-12-01

    We consider high-rate systematic recursive convolutional encoders to be adopted as constituent encoders in turbo schemes. Douillard and Berrou showed that, despite its complexity, the construction of high-rate turbo codes by means of high-rate constituent encoders is advantageous over the construction based on puncturing rate-1/2 constituent encoders. To reduce the decoding complexity of high-rate codes, we introduce the construction of the minimal trellis for a systematic recursive convolutional encoding matrix. A code search is conducted and examples are provided which indicate that a more finely grained decoding complexity-error performance trade-off is obtained.

  15. Development of a Bayesian recursive algorithm to find free-spaces for an intelligent wheelchair.

    PubMed

    Nguyen, Anh V; Su, Steven; Nguyen, Hung T

    2011-01-01

    This paper introduces a new shared control strategy for an intelligent wheelchair using a Bayesian recursive algorithm. Using the local environment information gathered by a laser range finder sensor and commands acquired through a user interface, a Bayesian recursive algorithm has been developed to find the most appropriate free-space, which corresponds to the highest posterior probability value. Then, an autonomous navigation algorithm will assist to manoeuvre the wheelchair in the chosen free-space. Experiment results demonstrate that the new method provides excellent performance with great flexibility and fast response.

  16. Student Monks--Teaching Recursion in an IS or CS Programming Course Using the Towers of Hanoi

    ERIC Educational Resources Information Center

    Benander, Alan C.; Benander, Barbara A.

    2008-01-01

    Educators have been using the Towers of Hanoi problem for many years as an example of a problem that has a very elegant recursive solution. However, the elegance and conciseness of this solution can make it difficult for students to understand the amount of computer time required in the execution of this solution. And, like many recursive computer…

  17. Use of XML and Java for collaborative petroleum reservoir modeling on the Internet

    USGS Publications Warehouse

    Victorine, J.; Watney, W.L.; Bhattacharya, S.

    2005-01-01

    The GEMINI (Geo-Engineering Modeling through INternet Informatics) is a public-domain, web-based freeware that is made up of an integrated suite of 14 Java-based software tools to accomplish on-line, real-time geologic and engineering reservoir modeling. GEMINI facilitates distant collaborations for small company and academic clients, negotiating analyses of both single and multiple wells. The system operates on a single server and an enterprise database. External data sets must be uploaded into this database. Feedback from GEMINI users provided the impetus to develop Stand Alone Web Start Applications of GEMINI modules that reside in and operate from the user's PC. In this version, the GEMINI modules run as applets, which may reside in local user PCs, on the server, or Java Web Start. In this enhanced version, XML-based data handling procedures are used to access data from remote and local databases and save results for later access and analyses. The XML data handling process also integrates different stand-alone GEMINI modules enabling the user(s) to access multiple databases. It provides flexibility to the user to customize analytical approach, database location, and level of collaboration. An example integrated field-study using GEMINI modules and Stand Alone Web Start Applications is provided to demonstrate the versatile applicability of this freeware for cost-effective reservoir modeling. ?? 2005 Elsevier Ltd. All rights reserved.

  18. Use of XML and Java for collaborative petroleum reservoir modeling on the Internet

    NASA Astrophysics Data System (ADS)

    Victorine, John; Watney, W. Lynn; Bhattacharya, Saibal

    2005-11-01

    The GEMINI (Geo-Engineering Modeling through INternet Informatics) is a public-domain, web-based freeware that is made up of an integrated suite of 14 Java-based software tools to accomplish on-line, real-time geologic and engineering reservoir modeling. GEMINI facilitates distant collaborations for small company and academic clients, negotiating analyses of both single and multiple wells. The system operates on a single server and an enterprise database. External data sets must be uploaded into this database. Feedback from GEMINI users provided the impetus to develop Stand Alone Web Start Applications of GEMINI modules that reside in and operate from the user's PC. In this version, the GEMINI modules run as applets, which may reside in local user PCs, on the server, or Java Web Start. In this enhanced version, XML-based data handling procedures are used to access data from remote and local databases and save results for later access and analyses. The XML data handling process also integrates different stand-alone GEMINI modules enabling the user(s) to access multiple databases. It provides flexibility to the user to customize analytical approach, database location, and level of collaboration. An example integrated field-study using GEMINI modules and Stand Alone Web Start Applications is provided to demonstrate the versatile applicability of this freeware for cost-effective reservoir modeling.

  19. Ontology aided modeling of organic reaction mechanisms with flexible and fragment based XML markup procedures.

    PubMed

    Sankar, Punnaivanam; Aghila, Gnanasekaran

    2007-01-01

    The mechanism models for primary organic reactions encoding the structural fragments undergoing substitution, addition, elimination, and rearrangements are developed. In the proposed models, each and every structural component of mechanistic pathways is represented with flexible and fragment based markup technique in XML syntax. A significant feature of the system is the encoding of the electron movements along with the other components like charges, partial charges, half bonded species, lone pair electrons, free radicals, reaction arrows, etc. needed for a complete representation of reaction mechanism. The rendering of reaction schemes described with the proposed methodology is achieved with a concise XML extension language interoperating with the structure markup. The reaction scheme is visualized as 2D graphics in a browser by converting them into SVG documents enabling the desired layouts normally perceived by the chemists conventionally. An automatic representation of the complex patterns of the reaction mechanism is achieved by reusing the knowledge in chemical ontologies and developing artificial intelligence components in terms of axioms.

  20. MASCOT HTML and XML parser: an implementation of a novel object model for protein identification data.

    PubMed

    Yang, Chunguang G; Granite, Stephen J; Van Eyk, Jennifer E; Winslow, Raimond L

    2006-11-01

    Protein identification using MS is an important technique in proteomics as well as a major generator of proteomics data. We have designed the protein identification data object model (PDOM) and developed a parser based on this model to facilitate the analysis and storage of these data. The parser works with HTML or XML files saved or exported from MASCOT MS/MS ions search in peptide summary report or MASCOT PMF search in protein summary report. The program creates PDOM objects, eliminates redundancy in the input file, and has the capability to output any PDOM object to a relational database. This program facilitates additional analysis of MASCOT search results and aids the storage of protein identification information. The implementation is extensible and can serve as a template to develop parsers for other search engines. The parser can be used as a stand-alone application or can be driven by other Java programs. It is currently being used as the front end for a system that loads HTML and XML result files of MASCOT searches into a relational database. The source code is freely available at http://www.ccbm.jhu.edu and the program uses only free and open-source Java libraries.

  1. Should Iteration and Recursion Be Part of the Secondary Student's Mathematics Toolbox?

    ERIC Educational Resources Information Center

    Quesada, Antonio

    1999-01-01

    Presents examples to illustrate that iteration and recursion are useful pedagogical tools that can be used in many areas of mathematics, make models traditionally taught in upper levels accessible in grades 9-14, and provide alternative problem solving approaches less dependent upon ready-made formulas. (Author/MM)

  2. Recursive inverse kinematics for robot arms via Kalman filtering and Bryson-Frazier smoothing

    NASA Technical Reports Server (NTRS)

    Rodriguez, G.; Scheid, R. E., Jr.

    1987-01-01

    This paper applies linear filtering and smoothing theory to solve recursively the inverse kinematics problem for serial multilink manipulators. This problem is to find a set of joint angles that achieve a prescribed tip position and/or orientation. A widely applicable numerical search solution is presented. The approach finds the minimum of a generalized distance between the desired and the actual manipulator tip position and/or orientation. Both a first-order steepest-descent gradient search and a second-order Newton-Raphson search are developed. The optimal relaxation factor required for the steepest descent method is computed recursively using an outward/inward procedure similar to those used typically for recursive inverse dynamics calculations. The second-order search requires evaluation of a gradient and an approximate Hessian. A Gauss-Markov approach is used to approximate the Hessian matrix in terms of products of first-order derivatives. This matrix is inverted recursively using a two-stage process of inward Kalman filtering followed by outward smoothing. This two-stage process is analogous to that recently developed by the author to solve by means of spatial filtering and smoothing the forward dynamics problem for serial manipulators.

  3. A Cognitive Processing Account of Individual Differences in Novice Logo Programmers' Conceptualisation and Use of Recursion.

    ERIC Educational Resources Information Center

    Gibbons, Pamela

    1995-01-01

    Describes a study that investigated individual differences in the construction of mental models of recursion in LOGO programming. The learning process was investigated from the perspective of Norman's mental models theory and employed diSessa's ontology regarding distributed, functional, and surrogate mental models, and the Luria model of brain…

  4. The recursive maximum likelihood proportion estimator: User's guide and test results

    NASA Technical Reports Server (NTRS)

    Vanrooy, D. L.

    1976-01-01

    Implementation of the recursive maximum likelihood proportion estimator is described. A user's guide to programs as they currently exist on the IBM 360/67 at LARS, Purdue is included, and test results on LANDSAT data are described. On Hill County data, the algorithm yields results comparable to the standard maximum likelihood proportion estimator.

  5. User's Guide for the Precision Recursive Estimator for Ephemeris Refinement (PREFER)

    NASA Technical Reports Server (NTRS)

    Gibbs, B. P.

    1982-01-01

    PREFER is a recursive orbit determination program which is used to refine the ephemerides produced by a batch least squares program (e.g., GTDS). It is intended to be used primarily with GTDS and, thus, is compatible with some of the GTDS input/output files.

  6. Split-remerge method for eliminating processing window artifacts in recursive hierarchical segmentation

    NASA Technical Reports Server (NTRS)

    Tilton, James C. (Inventor)

    2010-01-01

    A method, computer readable storage, and apparatus for implementing recursive segmentation of data with spatial characteristics into regions including splitting-remerging of pixels with contagious region designations and a user controlled parameter for providing a preference for merging adjacent regions to eliminate window artifacts.

  7. Analysis of litter size and average litter weight in pigs using a recursive model.

    PubMed

    Varona, Luis; Sorensen, Daniel; Thompson, Robin

    2007-11-01

    An analysis of litter size and average piglet weight at birth in Landrace and Yorkshire using a standard two-trait mixed model (SMM) and a recursive mixed model (RMM) is presented. The RMM establishes a one-way link from litter size to average piglet weight. It is shown that there is a one-to-one correspondence between the parameters of SMM and RMM and that they generate equivalent likelihoods. As parameterized in this work, the RMM tests for the presence of a recursive relationship between additive genetic values, permanent environmental effects, and specific environmental effects of litter size, on average piglet weight. The equivalent standard mixed model tests whether or not the covariance matrices of the random effects have a diagonal structure. In Landrace, posterior predictive model checking supports a model without any form of recursion or, alternatively, a SMM with diagonal covariance matrices of the three random effects. In Yorkshire, the same criterion favors a model with recursion at the level of specific environmental effects only, or, in terms of the SMM, the association between traits is shown to be exclusively due to an environmental (negative) correlation. It is argued that the choice between a SMM or a RMM should be guided by the availability of software, by ease of interpretation, or by the need to test a particular theory or hypothesis that may best be formulated under one parameterization and not the other.

  8. Recursive Frame Analysis: Reflections on the Development of a Qualitative Research Method

    ERIC Educational Resources Information Center

    Keeney, Hillary; Keeney, Bradford

    2012-01-01

    The origin of recursive frame analysis (RFA) is revisited and discussed as a postmodern alternative to modernist therapeutic models and research methods that foster hegemony of a preferred therapeutic metaphor, narrative, or strategy. It encourages improvisational performance while enabling a means of scoring the change and movement of the…

  9. XML technologies for the Omaha System: a data model, a Java tool and several case studies supporting home healthcare.

    PubMed

    Vittorini, Pierpaolo; Tarquinio, Antonietta; di Orio, Ferdinando

    2009-03-01

    The eXtensible markup language (XML) is a metalanguage which is useful to represent and exchange data between heterogeneous systems. XML may enable healthcare practitioners to document, monitor, evaluate, and archive medical information and services into distributed computer environments. Therefore, the most recent proposals on electronic health records (EHRs) are usually based on XML documents. Since none of the existing nomenclatures were specifically developed for use in automated clinical information systems, but were adapted to such use, numerous current EHRs are organized as a sequence of events, each represented through codes taken from international classification systems. In nursing, a hierarchically organized problem-solving approach is followed, which hardly couples with the sequential organization of such EHRs. Therefore, the paper presents an XML data model for the Omaha System taxonomy, which is one of the most important international nomenclatures used in the home healthcare nursing context. Such a data model represents the formal definition of EHRs specifically developed for nursing practice. Furthermore, the paper delineates a Java application prototype which is able to manage such documents, shows the possibility to transform such documents into readable web pages, and reports several case studies, one currently managed by the home care service of a Health Center in Central Italy.

  10. Integrating XQuery-Enabled SCORM XML Metadata Repositories into an RDF-Based E-Learning P2P Network

    ERIC Educational Resources Information Center

    Qu, Changtao; Nejdl, Wolfgang

    2004-01-01

    Edutella is an RDF-based E-Learning P2P network that is aimed to accommodate heterogeneous learning resource metadata repositories in a P2P manner and further facilitate the exchange of metadata between these repositories based on RDF. Whereas Edutella provides RDF metadata repositories with a quite natural integration approach, XML metadata…

  11. Using Extensible Markup Language (XML) for the Single Source Delivery of Educational Resources by Print and Online: A Case Study

    ERIC Educational Resources Information Center

    Walsh, Lucas

    2007-01-01

    This article seeks to provide an introduction to Extensible Markup Language (XML) by looking at its use in a single source publishing approach to the provision of teaching resources in both hardcopy and online. Using the development of the International Baccalaureate Organisation's online Economics Subject Guide as a practical example, this…

  12. XML Storage for Magnetotelluric Transfer Functions: Towards a Comprehensive Online Reference Database

    NASA Astrophysics Data System (ADS)

    Kelbert, A.; Blum, C.

    2015-12-01

    Magnetotelluric Transfer Functions (MT TFs) represent most of the information about Earth electrical conductivity found in the raw electromagnetic data, providing inputs for further inversion and interpretation. To be useful for scientific interpretation, they must also contain carefully recorded metadata. Making these data available in a discoverable and citable fashion would provide the most benefit to the scientific community, but such a development requires that the metadata is not only present in the file but is also searchable. The most commonly used MT TF format to date, the historical Society of Exploration Geophysicists Electromagnetic Data Interchange Standard 1987 (EDI), no longer supports some of the needs of modern magnetotellurics, most notably accurate error bars recording. Moreover, the inherent heterogeneity of EDI's and other historic MT TF formats has mostly kept the community away from healthy data sharing practices. Recently, the MT team at Oregon State University in collaboration with IRIS Data Management Center developed a new, XML-based format for MT transfer functions, and an online system for long-term storage, discovery and sharing of MT TF data worldwide (IRIS SPUD; www.iris.edu/spud/emtf). The system provides a query page where all of the MT transfer functions collected within the USArray MT experiment and other field campaigns can be searched for and downloaded; an automatic on-the-fly conversion to the historic EDI format is also included. To facilitate conversion to the new, more comprehensive and sustainable, XML format for MT TFs, and to streamline inclusion of historic data into the online database, we developed a set of open source format conversion tools, which can be used for rotation of MT TFs as well as a general XML <-> EDI converter (https://seiscode.iris.washington.edu/projects/emtf-fcu). Here, we report on the newly established collaboration between the USGS Geomagnetism Program and the Oregon State University to gather

  13. Evaluation of ISO EN 13606 as a result of its implementation in XML.

    PubMed

    Austin, Tony; Sun, Shanghua; Hassan, Taher; Kalra, Dipak

    2013-12-01

    The five parts of the ISO EN 13606 standard define a means by which health-care records can be exchanged between computer systems. Starting within the European standardisation process, it has now become internationally ratified in ISO. However, ISO standards do not require that a reference implementation be provided, and in order for ISO EN 13606 to deliver the expected benefits, it must be provided not as a document, but as an operational system that is not vendor specific. This article describes the evolution of an Extensible Markup Language (XML) Schema through three iterations, each of which emphasised one particular approach to delivering an executable equivalent to the printed standard. Developing these operational versions and incorporating feedback from users of these demonstrated where implementation compromises were needed and exposed defects in the standard. These are discussed herein. They may require a future technical revision to ISO EN 13606 to resolve the issues identified.

  14. Evaluation of ISO EN 13606 as a result of its implementation in XML

    PubMed Central

    Sun, Shanghua; Hassan, Taher; Kalra, Dipak

    2013-01-01

    The five parts of the ISO EN 13606 standard define a means by which health-care records can be exchanged between computer systems. Starting within the European standardisation process, it has now become internationally ratified in ISO. However, ISO standards do not require that a reference implementation be provided, and in order for ISO EN 13606 to deliver the expected benefits, it must be provided not as a document, but as an operational system that is not vendor specific. This article describes the evolution of an Extensible Markup Language (XML) Schema through three iterations, each of which emphasised one particular approach to delivering an executable equivalent to the printed standard. Developing these operational versions and incorporating feedback from users of these demonstrated where implementation compromises were needed and exposed defects in the standard. These are discussed herein. They may require a future technical revision to ISO EN 13606 to resolve the issues identified. PMID:23995217

  15. Encoding of Fundamental Chemical Entities of Organic Reactivity Interest using chemical ontology and XML.

    PubMed

    Durairaj, Vijayasarathi; Punnaivanam, Sankar

    2015-09-01

    Fundamental chemical entities are identified in the context of organic reactivity and classified as appropriate concept classes namely ElectronEntity, AtomEntity, AtomGroupEntity, FunctionalGroupEntity and MolecularEntity. The entity classes and their subclasses are organized into a chemical ontology named "ChemEnt" for the purpose of assertion, restriction and modification of properties through entity relations. Individual instances of entity classes are defined and encoded as a library of chemical entities in XML. The instances of entity classes are distinguished with a unique notation and identification values in order to map them with the ontology definitions. A model GUI named Entity Table is created to view graphical representations of all the entity instances. The detection of chemical entities in chemical structures is achieved through suitable algorithms. The possibility of asserting properties to the entities at different levels and the mechanism of property flow within the hierarchical entity levels is outlined.

  16. QuakeML: XML for Seismological Data Exchange and Resource Metadata Description

    NASA Astrophysics Data System (ADS)

    Euchner, F.; Schorlemmer, D.; Becker, J.; Heinloo, A.; Kästli, P.; Saul, J.; Weber, B.; QuakeML Working Group

    2007-12-01

    QuakeML is an XML-based data exchange format for seismology that is under development. Current collaborators are from ETH, GFZ, USC, USGS, IRIS DMC, EMSC, ORFEUS, and ISTI. QuakeML development was motivated by the lack of a widely accepted and well-documented data format that is applicable to a broad range of fields in seismology. The development team brings together expertise from communities dealing with analysis and creation of earthquake catalogs, distribution of seismic bulletins, and real-time processing of seismic data. Efforts to merge QuakeML with existing XML dialects are under way. The first release of QuakeML will cover a basic description of seismic events including picks, arrivals, amplitudes, magnitudes, origins, focal mechanisms, and moment tensors. Further extensions are in progress or planned, e.g., for macroseismic information, location probability density functions, slip distributions, and ground motion information. The QuakeML language definition is supplemented by a concept to provide resource metadata and facilitate metadata exchange between distributed data providers. For that purpose, we introduce unique, location-independent identifiers of seismological resources. As an application of QuakeML, ETH Zurich currently develops a Python-based seismicity analysis toolkit as a contribution to CSEP (Collaboratory for the Study of Earthquake Predictability). We follow a collaborative and transparent development approach along the lines of the procedures of the World Wide Web Consortium (W3C). QuakeML currently is in working draft status. The standard description will be subjected to a public Request for Comments (RFC) process and eventually reach the status of a recommendation. QuakeML can be found at http://www.quakeml.org.

  17. Veterinary medicines update.

    PubMed

    2017-03-11

    The following information has been produced for Veterinary Record by the Veterinary Medicines Directorate (VMD) to provide an update for veterinary surgeons on recent changes to marketing authorisations for veterinary medicines in the UK and on other relevant issues.

  18. ISS Update: NEEMO 16

    NASA Video Gallery

    ISS Update commentator Josh Byerly interviews astronaut Stan Love about the NEEMO 16 mission from Aquarius Base. Questions? Ask us on Twitter @NASA_Johnson and include the hashtag #askStation. For ...

  19. Veterinary medicines: product update.

    PubMed

    2014-04-05

    The following information has been produced for Veterinary Record by the Veterinary Medicines Directorate (VMD) to provide an update for veterinary surgeons on recent changes to marketing authorisations for veterinary medicines in the UK and on other relevant issues.

  20. Veterinary medicines: product update.

    PubMed

    2014-03-01

    The following information has been produced for Veterinary Record by the Veterinary Medicines Directorate (VMD) to provide an update for veterinary surgeons on recent changes to marketing authorisations for veterinary medicines in the UK and on other relevant issues.

  1. Veterinary medicines: product update.

    PubMed

    2014-08-02

    The following information has been produced for Veterinary Record by the Veterinary Medicines Directorate (VMD) to provide an update for veterinary surgeons on recent changes to marketing authorisations for veterinary medicines in the UK and on other relevant issues.

  2. Veterinary medicines: product update.

    PubMed

    2014-11-01

    The following information has been produced for Veterinary Record by the Veterinary Medicines Directorate (VMD) to provide an update for veterinary surgeons on recent changes to marketing authorisations for veterinary medicines in the UK and on other relevant issues.

  3. Veterinary medicines: product update.

    PubMed

    2014-09-06

    The following information has been produced for Veterinary Record by the Veterinary Medicines Directorate (VMD) to provide an update for veterinary surgeons on recent changes to marketing authorisations for veterinary medicines in the UK, and on other relevant issues.

  4. Red Hill Updates

    EPA Pesticide Factsheets

    This and other periodic updates are intended to keep the public informed on major progress being made to protect public health and the environment at the Red Hill Underground Fuel Storage Facility in Hawaii.

  5. ISS Update: Suitport Testing

    NASA Video Gallery

    ISS Update commentator Lynnette Madison interviews Joel Maganza, Test Director, about thermal vacuum chambers and unmanned and human-testing with the Suitport. Questions? Ask us on Twitter @NASA_Jo...

  6. HepML, an XML-based format for describing simulated data in high energy physics

    NASA Astrophysics Data System (ADS)

    Belov, S.; Dudko, L.; Kekelidze, D.; Sherstnev, A.

    2010-10-01

    In this paper we describe a HepML format and a corresponding C++ library developed for keeping complete description of parton level events in a unified and flexible form. HepML tags contain enough information to understand what kind of physics the simulated events describe and how the events have been prepared. A HepML block can be included into event files in the LHEF format. The structure of the HepML block is described by means of several XML Schemas. The Schemas define necessary information for the HepML block and how this information should be located within the block. The library libhepml is a C++ library intended for parsing and serialization of HepML tags, and representing the HepML block in computer memory. The library is an API for external software. For example, Matrix Element Monte Carlo event generators can use the library for preparing and writing a header of an LHEF file in the form of HepML tags. In turn, Showering and Hadronization event generators can parse the HepML header and get the information in the form of C++ classes. libhepml can be used in C++, C, and Fortran programs. All necessary parts of HepML have been prepared and we present the project to the HEP community. Program summaryProgram title: libhepml Catalogue identifier: AEGL_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEGL_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: GNU GPLv3 No. of lines in distributed program, including test data, etc.: 138 866 No. of bytes in distributed program, including test data, etc.: 613 122 Distribution format: tar.gz Programming language: C++, C Computer: PCs and workstations Operating system: Scientific Linux CERN 4/5, Ubuntu 9.10 RAM: 1 073 741 824 bytes (1 Gb) Classification: 6.2, 11.1, 11.2 External routines: Xerces XML library ( http://xerces.apache.org/xerces-c/), Expat XML Parser ( http://expat.sourceforge.net/) Nature of problem: Monte Carlo simulation in high

  7. Non-abelian Z-theory: Berends-Giele recursion for the α '-expansion of disk integrals

    NASA Astrophysics Data System (ADS)

    Mafra, Carlos R.; Schlotterer, Oliver

    2017-01-01

    We present a recursive method to calculate the α ' -expansion of disk integrals arising in tree-level scattering of open strings which resembles the approach of Berends and Giele to gluon amplitudes. Following an earlier interpretation of disk integrals as doubly partial amplitudes of an effective theory of scalars dubbed as Z-theory, we pinpoint the equation of motion of Z-theory from the Berends-Giele recursion for its tree amplitudes. A computer implementation of this method including explicit results for the recursion up to order α '7 is made available on the website http://repo.or.cz/BGap.git.

  8. A recursive vesicle-based model protocell with a primitive model cell cycle

    PubMed Central

    Kurihara, Kensuke; Okura, Yusaku; Matsuo, Muneyuki; Toyota, Taro; Suzuki, Kentaro; Sugawara, Tadashi

    2015-01-01

    Self-organized lipid structures (protocells) have been proposed as an intermediate between nonliving material and cellular life. Synthetic production of model protocells can demonstrate the potential processes by which living cells first arose. While we have previously described a giant vesicle (GV)-based model protocell in which amplification of DNA was linked to self-reproduction, the ability of a protocell to recursively self-proliferate for multiple generations has not been demonstrated. Here we show that newborn daughter GVs can be restored to the status of their parental GVs by pH-induced vesicular fusion of daughter GVs with conveyer GVs filled with depleted substrates. We describe a primitive model cell cycle comprising four discrete phases (ingestion, replication, maturity and division), each of which is selectively activated by a specific external stimulus. The production of recursive self-proliferating model protocells represents a step towards eventual production of model protocells that are able to mimic evolution. PMID:26418735

  9. Memory efficient and constant time 2D-recursive spatial averaging filter for embedded implementations

    NASA Astrophysics Data System (ADS)

    Gan, Qifeng; Seoud, Lama; Ben Tahar, Houssem; Langlois, J. M. Pierre

    2016-04-01

    Spatial Averaging Filters (SAF) are extensively used in image processing for image smoothing and denoising. Their latest implementations have already achieved constant time computational complexity regardless of kernel size. However, all the existing O(1) algorithms require additional memory for temporary data storage. In order to minimize memory usage in embedded systems, we introduce a new two-dimensional recursive SAF. It uses previous resultant pixel values along both rows and columns to calculate the current one. It can achieve constant time computational complexity without using any additional memory usage. Experimental comparisons with previous SAF implementations shows that the proposed 2D-Recursive SAF does not require any additional memory while offering a computational time similar to the most efficient existing SAF algorithm. These features make it especially suitable for embedded systems with limited memory capacity.

  10. Recursive estimation of 3D motion and surface structure from local affine flow parameters.

    PubMed

    Calway, Andrew

    2005-04-01

    A recursive structure from motion algorithm based on optical flow measurements taken from an image sequence is described. It provides estimates of surface normals in addition to 3D motion and depth. The measurements are affine motion parameters which approximate the local flow fields associated with near-planar surface patches in the scene. These are integrated over time to give estimates of the 3D parameters using an extended Kalman filter. This also estimates the camera focal length and, so, the 3D estimates are metric. The use of parametric measurements means that the algorithm is computationally less demanding than previous optical flow approaches and the recursive filter builds in a degree of noise robustness. Results of experiments on synthetic and real image sequences demonstrate that the algorithm performs well.

  11. Learning the dynamics and time-recursive boundary detection of deformable objects.

    PubMed

    Sun, Walter; Cetin, Müjdat; Chan, Raymond; Willsky, Alan S

    2008-11-01

    We propose a principled framework for recursively segmenting deformable objects across a sequence of frames. We demonstrate the usefulness of this method on left ventricular segmentation across a cardiac cycle. The approach involves a technique for learning the system dynamics together with methods of particle-based smoothing as well as nonparametric belief propagation on a loopy graphical model capturing the temporal periodicity of the heart. The dynamic system state is a low-dimensional representation of the boundary, and the boundary estimation involves incorporating curve evolution into recursive state estimation. By formulating the problem as one of state estimation, the segmentation at each particular time is based not only on the data observed at that instant, but also on predictions based on past and future boundary estimates. Although this paper focuses on left ventricle segmentation, the method generalizes to temporally segmenting any deformable object.

  12. Attitude determination and calibration using a recursive maximum likelihood-based adaptive Kalman filter

    NASA Technical Reports Server (NTRS)

    Kelly, D. A.; Fermelia, A.; Lee, G. K. F.

    1990-01-01

    An adaptive Kalman filter design that utilizes recursive maximum likelihood parameter identification is discussed. At the center of this design is the Kalman filter itself, which has the responsibility for attitude determination. At the same time, the identification algorithm is continually identifying the system parameters. The approach is applicable to nonlinear, as well as linear systems. This adaptive Kalman filter design has much potential for real time implementation, especially considering the fast clock speeds, cache memory and internal RAM available today. The recursive maximum likelihood algorithm is discussed in detail, with special attention directed towards its unique matrix formulation. The procedure for using the algorithm is described along with comments on how this algorithm interacts with the Kalman filter.

  13. Recursive dynamics of topological trees of rigid bodies via Kalman filtering and Bryson-Frazier smoothing

    NASA Technical Reports Server (NTRS)

    Rodriguez, G.

    1988-01-01

    The inverse and forward dynamics problems for a set of rigid bodies connected by hinges to form a topological tree are solved by using recursive techniques from linear filtering and smoothing theory. An inward filtering sequence computes a set of constraint moments and forces. This is followed by an outward sequence to determine a corresponding set of angular and linear accelerations. An inward sequence begins at the tips of all of the terminal bodies of the tree and proceeds inwardly through all of the branches until it reaches the root. Similarly, an outward sequence begins at the root and propagates to all of the tree branches until it reaches the tips of the terminal bodies. The paper also provides an approach to evaluate recursively the composite multibody system inertia matrix and its inverse.

  14. Improved recursive Green's function formalism for quasi one-dimensional systems with realistic defects

    NASA Astrophysics Data System (ADS)

    Teichert, Fabian; Zienert, Andreas; Schuster, Jörg; Schreiber, Michael

    2017-04-01

    We derive an improved version of the recursive Green's function formalism (RGF), which is a standard tool in the quantum transport theory. We consider the case of disordered quasi one-dimensional materials where the disorder is applied in form of randomly distributed realistic defects, leading to partly periodic Hamiltonian matrices. The algorithm accelerates the common RGF in the recursive decimation scheme, using the iteration steps of the renormalization decimation algorithm. This leads to a smaller effective system, which is treated using the common forward iteration scheme. The computational complexity scales linearly with the number of defects, instead of linearly with the total system length for the conventional approach. We show that the scaling of the calculation time of the Green's function depends on the defect density of a random test system. Furthermore, we discuss the calculation time and the memory requirement of the whole transport formalism applied to defective carbon nanotubes.

  15. A novel thermodynamic state recursion method for description of nonideal nonlinear chromatographic process of frontal analysis.

    PubMed

    Liu, Qian; OuYang, Liangfei; Liang, Heng; Li, Nan; Geng, Xindu

    2012-06-01

    A novel thermodynamic state recursion (TSR) method, which is based on nonequilibrium thermodynamic path described by the Lagrangian-Eulerian representation, is presented to simulate the whole chromatographic process of frontal analysis using the spatial distribution of solute bands in time series like as a series of images. TSR differs from the current numerical methods using the partial differential equations in Eulerian representation. The novel method is used to simulate the nonideal, nonlinear hydrophobic interaction chromatography (HIC) processes of lysozyme and myoglobin under the discrete complex boundary conditions. The results show that the simulated breakthrough curves agree well with the experimental ones. The apparent diffusion coefficient and the Langmuir isotherm parameters of the two proteins in HIC are obtained by the state recursion inverse method. Due to its the time domain and Markov characteristics, TSR is applicable to the design and online control of the nonlinear multicolumn chromatographic systems.

  16. On the development and application of a continuous-discrete recursive prediction error algorithm.

    PubMed

    Stigter, J D; Beck, M B

    2004-10-01

    Recursive state and parameter reconstruction is a well-established field in control theory. In the current paper we derive a continuous-discrete version of recursive prediction error algorithm and apply the filter in an environmental and biological setting as a possible alternative to the well-known extended Kalman filter. The framework from which the derivation is started is the so-called 'innovations-format' of the (continuous time) system model, including (discrete time) measurements. After the algorithm has been motivated and derived, it is subsequently applied to hypothetical and 'real-life' case studies including reconstruction of biokinetic parameters and parameters characterizing the dynamics of a river in the United Kingdom. Advantages and characteristics of the method are discussed.

  17. Non-recursive augmented Lagrangian algorithms for the forward and inverse dynamics of constrained flexible multibodies

    NASA Technical Reports Server (NTRS)

    Bayo, Eduardo; Ledesma, Ragnar

    1993-01-01

    A technique is presented for solving the inverse dynamics of flexible planar multibody systems. This technique yields the non-causal joint efforts (inverse dynamics) as well as the internal states (inverse kinematics) that produce a prescribed nominal trajectory of the end effector. A non-recursive global Lagrangian approach is used in formulating the equations for motion as well as in solving the inverse dynamics equations. Contrary to the recursive method previously presented, the proposed method solves the inverse problem in a systematic and direct manner for both open-chain as well as closed-chain configurations. Numerical simulation shows that the proposed procedure provides an excellent tracking of the desired end effector trajectory.

  18. Prior estimation of motion using recursive perceptron with sEMG: a case of wrist angle.

    PubMed

    Kuroda, Yoshihiro; Tanaka, Takeshi; Imura, Masataka; Oshiro, Osamu

    2012-01-01

    Muscle activity is followed by myoelectric potentials. Prior estimation of motion by surface electromyography can be utilized to assist the physically impaired people as well as surgeon. In this paper, we proposed a real-time method for the prior estimation of motion from surface electromyography, especially in the case of wrist angle. The method was based on the recursive processing of multi-layer perceptron, which is trained quickly. A single layer perceptron calculates quasi tensional force of muscles from surface electromyography. A three-layer perceptron calculates the wrist's change in angle. In order to estimate a variety of motions properly, the perceptron was designed to estimate motion in a short time period, e.g. 1ms. Recursive processing enables the method to estimate motion in the target time period, e.g. 50ms. The results of the experiments showed statistical significance for the precedence of estimated angle to the measured one.

  19. An explicit algorithm for fully flexible unit cell simulation with recursive thermostat chains.

    PubMed

    Jung, Kwangsub; Cho, Maenghyo

    2008-10-28

    Through the combination of the recursive multiple thermostat (RMT) Nose-Poincare and Parrinello-Rahman methods, the recursive multiple thermostat chained fully flexible unit cell (RMT-NsigmaT) molecular dynamics method is proposed for isothermal-isobaric simulation. The RMT method is known to have the advantage of achieving the ergodicity that is required for canonical sampling of the harmonic oscillator. Thus, an explicit time integration algorithm is developed for RMT-NsigmaT. We examine the ergodicity for various parameters of RMT-NsigmaT using bulk and thin film structures with different numbers of copper atoms and thicknesses in various environments. Through the numerical simulations, we conclude that the RMT-NsigmaT method is advantageous in the cases of lower temperatures.

  20. A recursive vesicle-based model protocell with a primitive model cell cycle

    NASA Astrophysics Data System (ADS)

    Kurihara, Kensuke; Okura, Yusaku; Matsuo, Muneyuki; Toyota, Taro; Suzuki, Kentaro; Sugawara, Tadashi

    2015-09-01

    Self-organized lipid structures (protocells) have been proposed as an intermediate between nonliving material and cellular life. Synthetic production of model protocells can demonstrate the potential processes by which living cells first arose. While we have previously described a giant vesicle (GV)-based model protocell in which amplification of DNA was linked to self-reproduction, the ability of a protocell to recursively self-proliferate for multiple generations has not been demonstrated. Here we show that newborn daughter GVs can be restored to the status of their parental GVs by pH-induced vesicular fusion of daughter GVs with conveyer GVs filled with depleted substrates. We describe a primitive model cell cycle comprising four discrete phases (ingestion, replication, maturity and division), each of which is selectively activated by a specific external stimulus. The production of recursive self-proliferating model protocells represents a step towards eventual production of model protocells that are able to mimic evolution.

  1. A Note on Local Stability Conditions for Two Types of Monetary Models with Recursive Utility

    NASA Astrophysics Data System (ADS)

    Miyazaki, Kenji; Utsunomiya, Hitoshi

    2009-09-01

    This note explores local stability conditions for money-in-utility-function (MIUF) and transaction-costs (TC) models with recursive utility. Although Chen et al. [Chen, B.-L., M. Hsu, and C.-H. Lin, 2008, Inflation and growth: impatience and a qualitative equivalent, Journal of Money, Credit, and Banking, Vol. 40, No. 6, 1310-1323] investigated the relationship between inflation and growth in MIUF and TC models with recursive utility, they conducted only a comparative static analysis in a steady state. By establishing sufficient conditions for local stability, this note proves that impatience should be increasing in consumption and real balances. Increasing impatience, although less plausible from an empirical point of view, receives more support from a theoretical viewpoint.

  2. QCD amplitudes with 2 initial spacelike legs via generalised BCFW recursion

    NASA Astrophysics Data System (ADS)

    Kutak, Krzysztof; van Hameren, Andreas; Serino, Mirko

    2017-02-01

    We complete the generalisation of the BCFW recursion relation to the off-shell case, allowing for the computation of tree level scattering amplitudes for full High Energy Factorisation (HEF), i.e. with both incoming partons having a non-vanishing transverse momentum. We provide explicit results for color-ordered amplitudes with two off-shell legs in massless QCD up to 4 point, continuing the program begun in two previous papers. For the 4-fermion amplitudes, which are not BCFW-recursible, we perform a diagrammatic computation, so as to offer a complete set of expressions. We explicitly show and discuss some plots of the squared 2 → 2 matrix elements as functions of the differences in rapidity and azimuthal angle of the final state particles.

  3. A fast iterative recursive least squares algorithm for Wiener model identification of highly nonlinear systems.

    PubMed

    Kazemi, Mahdi; Arefi, Mohammad Mehdi

    2017-03-01

    In this paper, an online identification algorithm is presented for nonlinear systems in the presence of output colored noise. The proposed method is based on extended recursive least squares (ERLS) algorithm, where the identified system is in polynomial Wiener form. To this end, an unknown intermediate signal is estimated by using an inner iterative algorithm. The iterative recursive algorithm adaptively modifies the vector of parameters of the presented Wiener model when the system parameters vary. In addition, to increase the robustness of the proposed method against variations, a robust RLS algorithm is applied to the model. Simulation results are provided to show the effectiveness of the proposed approach. Results confirm that the proposed method has fast convergence rate with robust characteristics, which increases the efficiency of the proposed model and identification approach. For instance, the FIT criterion will be achieved 92% in CSTR process where about 400 data is used.

  4. Majorant recursions to determine eigenstate bounds of a symmetric exponential quantum anharmonic oscillator

    SciTech Connect

    Özdemir, Semra Bayat; Demiralp, Metin

    2015-12-31

    The determination of the energy states is highly studied issue in the quantum mechanics. Based on expectation values dynamics, energy states can be observed. But conditions and calculations vary depending on the created system. In this work, a symmetric exponential anharmonic oscillator is considered and development of a recursive approximation method is studied to find its ground energy state. The use of majorant values facilitates the approximate calculation of expectation values.

  5. Attitude estimation of earth orbiting satellites by decomposed linear recursive filters

    NASA Technical Reports Server (NTRS)

    Kou, S. R.

    1975-01-01

    Attitude estimation of earth orbiting satellites (including Large Space Telescope) subjected to environmental disturbances and noises was investigated. Modern control and estimation theory is used as a tool to design an efficient estimator for attitude estimation. Decomposed linear recursive filters for both continuous-time systems and discrete-time systems are derived. By using this accurate estimation of the attitude of spacecrafts, state variable feedback controller may be designed to achieve (or satisfy) high requirements of system performance.

  6. Modifications to Iterative Recursion Unfolding Algorithms and Computer Codes to Find More Appropriate Neutron Spectra.

    DTIC Science & Technology

    1984-06-06

    Iterative ReusinUnfodin * Algorithms And Computer Codes to Find More Apropriate Neutron Spectra L A. LOWRY AND T. L. JOHNSON Healt Plvwlcs S June 6, 1984...Classification) Modifications to Iterative Recursion Unfolding Algorithms and Computer Codes to Find More Appropriate Neutron Spectra 18. SUBJECT TERMS... TO FIND MORE APPROPRIATE NEUTRON SPECTRA INTRODUCTION The unfolding of neutron spectra using data from activation foils, Bonner spheres, or other

  7. An efficient and high performance linear recursive variable expansion implementation of the Smith-Waterman algorithm.

    PubMed

    Hasan, Laiq; Al-Ars, Zaid

    2009-01-01

    In this paper, we present an efficient and high performance linear recursive variable expansion (RVE) implementation of the Smith-Waterman (S-W) algorithm and compare it with a traditional linear systolic array implementation. The results demonstrate that the linear RVE implementation performs up to 2.33 times better than the traditional linear systolic array implementation, at the cost of utilizing 2 times more resources.

  8. Recursive Estimation of the Stein Center of SPD Matrices & its Applications.

    PubMed

    Salehian, Hesamoddin; Cheng, Guang; Vemuri, Baba C; Ho, Jeffrey

    2013-12-01

    Symmetric positive-definite (SPD) matrices are ubiquitous in Computer Vision, Machine Learning and Medical Image Analysis. Finding the center/average of a population of such matrices is a common theme in many algorithms such as clustering, segmentation, principal geodesic analysis, etc. The center of a population of such matrices can be defined using a variety of distance/divergence measures as the minimizer of the sum of squared distances/divergences from the unknown center to the members of the population. It is well known that the computation of the Karcher mean for the space of SPD matrices which is a negatively-curved Riemannian manifold is computationally expensive. Recently, the LogDet divergence-based center was shown to be a computationally attractive alternative. However, the LogDet-based mean of more than two matrices can not be computed in closed form, which makes it computationally less attractive for large populations. In this paper we present a novel recursive estimator for center based on the Stein distance - which is the square root of the LogDet divergence - that is significantly faster than the batch mode computation of this center. The key theoretical contribution is a closed-form solution for the weighted Stein center of two SPD matrices, which is used in the recursive computation of the Stein center for a population of SPD matrices. Additionally, we show experimental evidence of the convergence of our recursive Stein center estimator to the batch mode Stein center. We present applications of our recursive estimator to K-means clustering and image indexing depicting significant time gains over corresponding algorithms that use the batch mode computations. For the latter application, we develop novel hashing functions using the Stein distance and apply it to publicly available data sets, and experimental results have shown favorable comparisons to other competing methods.

  9. Systematic generation of multibody equations of motion suitable for recursive and parallel manipulation

    NASA Technical Reports Server (NTRS)

    Nikravesh, Parviz E.; Gim, Gwanghum; Arabyan, Ara; Rein, Udo

    1989-01-01

    The formulation of a method known as the joint coordinate method for automatic generation of the equations of motion for multibody systems is summarized. For systems containing open or closed kinematic loops, the equations of motion can be reduced systematically to a minimum number of second order differential equations. The application of recursive and nonrecursive algorithms to this formulation, computational considerations and the feasibility of implementing this formulation on multiprocessor computers are discussed.

  10. Evaluating Applicability of Four Recursive Algorithms for Computation of the Fully Normalized Associated Legendre Functions

    NASA Astrophysics Data System (ADS)

    Lei, Weiwei; Li, Kai

    2016-12-01

    There are four recursive algorithms used in the computation of the fully normalized associated Legendre functions (FNALFs): the standard forward column algorithm, the standard forward row algorithm, the recursive algorithm between every other degree, and the Belikov algorithm. These algorithms were evaluated in terms of their first relative numerical accuracy, second relative numerical accuracy, and computation speed and efficiency. The results show that when the degree n reaches 3000, both the recursive algorithm between every other degree and the Belikov algorithm are applicable for | cos θ | ∈[0, 1], with the latter better second relative numerical accuracy than the former at a slower computation speed. In terms of | cos θ | ∈[0, 1], the standard forward column algorithm, the recursive algorithm between every other degree, and the Belikov algorithm are applicable within degree n of 1900, and the standard forward column algorithm has the highest computation speed. The standard forward column algorithm is applicable for | cos θ | ∈[0, 1] within degree n of 1900. This algorithm's range of applicability decreases as the degree increases beyond 1900; however, it remains applicable within a minute range when | cos θ | is approximately equal to 1. The standard forward row algorithm has the smallest range of applicability: it is only applicable within degree n of 100 for | cos θ | ∈[0, 1], and its range of applicability decreases rapidly when the degree is greater than 100. The results of this research are expected to be useful to researchers in choosing the best algorithms for use in the computation of the FNALFs.

  11. Recursive Hierarchical Image Segmentation by Region Growing and Constrained Spectral Clustering

    NASA Technical Reports Server (NTRS)

    Tilton, James C.

    2002-01-01

    This paper describes an algorithm for hierarchical image segmentation (referred to as HSEG) and its recursive formulation (referred to as RHSEG). The HSEG algorithm is a hybrid of region growing and constrained spectral clustering that produces a hierarchical set of image segmentations based on detected convergence points. In the main, HSEG employs the hierarchical stepwise optimization (HS WO) approach to region growing, which seeks to produce segmentations that are more optimized than those produced by more classic approaches to region growing. In addition, HSEG optionally interjects between HSWO region growing iterations merges between spatially non-adjacent regions (i.e., spectrally based merging or clustering) constrained by a threshold derived from the previous HSWO region growing iteration. While the addition of constrained spectral clustering improves the segmentation results, especially for larger images, it also significantly increases HSEG's computational requirements. To counteract this, a computationally efficient recursive, divide-and-conquer, implementation of HSEG (RHSEG) has been devised and is described herein. Included in this description is special code that is required to avoid processing artifacts caused by RHSEG s recursive subdivision of the image data. Implementations for single processor and for multiple processor computer systems are described. Results with Landsat TM data are included comparing HSEG with classic region growing. Finally, an application to image information mining and knowledge discovery is discussed.

  12. Multiple Sclerosis: An Update.

    PubMed

    Faguy, Kathryn

    2016-05-01

    Multiple sclerosis (MS) is the most common disabling neurologic condition in young adults and imposes high financial and quality of life costs on patients, their families, and society. Yet, developments in the battle against MS include new treatments to slow its progression and updated diagnostic criteria that can accelerate diagnosis and effective treatment. This article offers a review and update on the disease, focusing on risk factors and possible causes, symptoms, forms of MS, diagnostic criteria and tools, and the expanding array of approved treatments. It also reports on the skyrocketing cost of MS drugs, misdiagnosis, and special patient populations with MS.

  13. Recursive causality in evolution: a model for epigenetic mechanisms in cancer development.

    PubMed

    Haslberger, A; Varga, F; Karlic, H

    2006-01-01

    Interactions between adaptative and selective processes are illustrated in the model of recursive causality as defined in Rupert Riedl's systems theory of evolution. One of the main features of this theory also termed as theory of evolving complexity is the centrality of the notion of 'recursive' or 'feedback' causality - 'the idea that every biological effect in living systems, in some way, feeds back to its own cause'. Our hypothesis is that "recursive" or "feedback" causality provides a model for explaining the consequences of interacting genetic and epigenetic mechanisms which are known to play a key role in development of cancer. Epigenetics includes any process that alters gene activity without changes of the DNA sequence. The most important epigenetic mechanisms are DNA-methylation and chromatin remodeling. Hypomethylation of so-called oncogenes and hypermethylation of tumor suppressor genes appear to be critical determinants of cancer. Folic acid, vitamin B12 and other nutrients influence the function of enzymes that participate in various methylation processes by affecting the supply of methyl groups into a variety of molecules which may be directly or indirectly associated with cancerogenesis. We present an example from our own studies by showing that vitamin D3 has the potential to de-methylate the osteocalcin-promoter in MG63 osteosarcoma cells. Consequently, a stimulation of osteocalcin synthesis can be observed. The above mentioned enzymes also play a role in development and differentiation of cells and organisms and thus illustrate the close association between evolutionary and developmental mechanisms. This enabled new ways to understand the interaction between the genome and environment and may improve biomedical concepts including environmental health aspects where epigenetic and genetic modifications are closely associated. Recent observations showed that methylated nucleotides in the gene promoter may serve as a target for solar UV

  14. Monte Carlo-Minimization and Monte Carlo Recursion Approaches to Structure and Free Energy.

    NASA Astrophysics Data System (ADS)

    Li, Zhenqin

    1990-08-01

    Biological systems are intrinsically "complex", involving many degrees of freedom, heterogeneity, and strong interactions among components. For the simplest of biological substances, e.g., biomolecules, which obey the laws of thermodynamics, we may attempt a statistical mechanical investigational approach. Even for these simplest many -body systems, assuming microscopic interactions are completely known, current computational methods in characterizing the overall structure and free energy face the fundamental challenge of an exponential amount of computation, with the rise in the number of degrees of freedom. As an attempt to surmount such problems, two computational procedures, the Monte Carlo-minimization and Monte Carlo recursion methods, have been developed as general approaches to the determination of structure and free energy of a complex thermodynamic system. We describe, in Chapter 2, the Monte Carlo-minimization method, which attempts to simulate natural protein folding processes and to overcome the multiple-minima problem. The Monte Carlo-minimization procedure has been applied to a pentapeptide, Met-enkephalin, leading consistently to the lowest energy structure, which is most likely to be the global minimum structure for Met-enkephalin in the absence of water, given the ECEPP energy parameters. In Chapter 3 of this thesis, we develop a Monte Carlo recursion method to compute the free energy of a given physical system with known interactions, which has been applied to a 32-particle Lennard-Jones fluid. In Chapter 4, we describe an efficient implementation of the recursion procedure, for the computation of the free energy of liquid water, with both MCY and TIP4P potential parameters for water. As a further demonstration of the power of the recursion method for calculating free energy, a general formalism of cluster formation from monatomic vapor is developed in Chapter 5. The Gibbs free energy of constrained clusters can be computed efficiently using the

  15. SEI: An update

    NASA Technical Reports Server (NTRS)

    Peach, Lewis L., Jr.

    1991-01-01

    An update on the Space Exploration Initiative (SEI) is given in viewgraph form. Topics covered include the key prerequisites of human exploration, project planning, Mars and lunar explorations, supporting technologies, near-term strategies for SEI, human support elements, and Space Station Freedom SEI accommodations.

  16. Cultural practices updates

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Cultural practice updates from 2013 included the effects of shredding in spring, residue management, periodic flooding, no-till fertilizer applications, and billet planting on cane tonnage and sugar yield. Shredding, whether high or low, had little impacts in 2013. However, burning following shreddi...

  17. Accountability Update, March 2000.

    ERIC Educational Resources Information Center

    Washington State Higher Education Coordinating Board, Olympia.

    This report provides the Washington State legislature, the Governor, and other interested parties with an update on the accountability performance of each of the state's public baccalaureate institutions (Central Washington University, Eastern Washington University, Evergreen State College, Washington State University, Western Washington…

  18. COPPER RESEARCH UPDATE

    EPA Science Inventory

    This presentation provides an update and overview of new research results and remaining research needs with respect to copper corrosion control issues. The topics to be covered include: occurrence of elevated copper release in systems that meet the Action Level; impact of water c...

  19. Community Update, 1999.

    ERIC Educational Resources Information Center

    Anderson, Julie, Ed.

    1999-01-01

    This document consists of nine issues (covering January through December 1999) of the newsletter "Community Update," containing articles on community and family involvement in education. Article topics include: new programs to help students prepare for college early; Vice President Al Gore announced the first-ever national Hispanic Education…

  20. Year 2000 Update.

    ERIC Educational Resources Information Center

    Berg, Frances M.

    1994-01-01

    Presents an update on the Year 2000 objectives for the nation that establish targets in 22 priority areas. The article offers information from a study on exercise by the President's Council on Physical Fitness and Sports, reviews Healthy People 2000 data, and lists Year 2000 physical activity and fitness objectives. (SM)

  1. Southern Horticultural Lab Update

    Technology Transfer Automated Retrieval System (TEKTRAN)

    This publication is a new quarterly update for members of the Mississippi Nursery and Landscape Association that is published quarterly in their association newsletter. Two other versions of this newsletter are being submitted to the Louisiana Nursery and Landscape Association (LALNLA) and the Alaba...

  2. Updating Older Fume Hoods.

    ERIC Educational Resources Information Center

    Saunders, G. Thomas

    1985-01-01

    Provides information on updating older fume hoods. Areas addressed include: (1) adjustment of the hood's back baffle; (2) hood air leakage; (3) light level; (4) hood location in relation to room traffic and room air; and (5) establishing and maintaining hood performance. (JN)

  3. Updating: Learning versus Supposing

    ERIC Educational Resources Information Center

    Zhao, Jiaying; Crupi, Vincenzo; Tentori, Katya; Fitelson, Branden; Osherson, Daniel

    2012-01-01

    Bayesian orthodoxy posits a tight relationship between conditional probability and updating. Namely, the probability of an event "A" after learning "B" should equal the conditional probability of "A" given "B" prior to learning "B". We examine whether ordinary judgment conforms to the orthodox view. In three experiments we found substantial…

  4. Special Education Law Update.

    ERIC Educational Resources Information Center

    Zirkel, Perry A.

    1989-01-01

    Due to an error made in printing Professor Zirkel's "Special Education Law; Recent Developments" in Volume 48 (October 13, 1988), these updates to the article are printed. Citations are organized under the following major sections: (1) diagnosis and placement; (2) treatment issues; (3) financial issues; and (4) gifted students. (MLF)

  5. Supreme Court Update

    ERIC Educational Resources Information Center

    Taylor, Kelley R.

    2009-01-01

    "Chief Justice Flubs Oath." "Justice Ginsburg Has Cancer Surgery." At the start of this year, those were the news headlines about the U.S. Supreme Court. But January 2009 also brought news about key education cases--one resolved and two others on the docket--of which school administrators should take particular note. The Supreme Court updates on…

  6. Community Update, 2001.

    ERIC Educational Resources Information Center

    Ashby, Nicole, Ed.

    2001-01-01

    This document consists of 10 issues (covering January through December 2000) of the newsletter, "Community Update," which features articles on community and family involvement in education. In addition to the articles, each issue (except the Special Issue) includes a preview of the month's Satellite Town Meeting; events and information…

  7. Integrating personalized medical test contents with XML and XSL-FO

    PubMed Central

    2011-01-01

    Background In 2004 the adoption of a modular curriculum at the medical faculty in Muenster led to the introduction of centralized examinations based on multiple-choice questions (MCQs). We report on how organizational challenges of realizing faculty-wide personalized tests were addressed by implementation of a specialized software module to automatically generate test sheets from individual test registrations and MCQ contents. Methods Key steps of the presented method for preparing personalized test sheets are (1) the compilation of relevant item contents and graphical media from a relational database with database queries, (2) the creation of Extensible Markup Language (XML) intermediates, and (3) the transformation into paginated documents. Results The software module by use of an open source print formatter consistently produced high-quality test sheets, while the blending of vectorized textual contents and pixel graphics resulted in efficient output file sizes. Concomitantly the module permitted an individual randomization of item sequences to prevent illicit collusion. Conclusions The automatic generation of personalized MCQ test sheets is feasible using freely available open source software libraries, and can be efficiently deployed on a faculty-wide scale. PMID:21362187

  8. Automatic Indexing for Content Analysis of Whale Recordings and XML Representation

    NASA Astrophysics Data System (ADS)

    Bénard, Frédéric; Glotin, Hervé

    2010-12-01

    This paper focuses on the robust indexing of sperm whale hydrophone recordings based on a set of features extracted from a real-time passive underwater acoustic tracking algorithm for multiple whales using four hydrophones. Acoustic localization permits the study of whale behavior in deep water without interfering with the environment. Given the position coordinates, we are able to generate different features such as the speed, energy of the clicks, Inter-Click-Interval (ICI), and so on. These features allow to construct different markers which allow us to index and structure the audio files. Thus, the behavior study is facilitated by choosing and accessing the corresponding index in the audio file. The complete indexing algorithm is processed on real data from the NUWC (Naval Undersea Warfare Center of the US Navy) and the AUTEC (Atlantic Undersea Test & Evaluation Center-Bahamas). Our model is validated by similar results from the US Navy (NUWC) and SOEST (School of Ocean and Earth Science and Technology) Hawaii university labs in a single whale case. Finally, as an illustration, we index a single whale sound file using the extracted whale's features provided by the tracking, and we present an example of an XML script structuring it.

  9. Development of XML Schema for Broadband Digital Seismograms and Data Center Portal

    NASA Astrophysics Data System (ADS)

    Takeuchi, N.; Tsuboi, S.; Ishihara, Y.; Nagao, H.; Yamagishi, Y.; Watanabe, T.; Yanaka, H.; Yamaji, H.

    2008-12-01

    There are a number of data centers around the globe, where the digital broadband seismograms are opened to researchers. Those centers use their own user interfaces and there are no standard to access and retrieve seismograms from different data centers using unified interface. One of the emergent technologies to realize unified user interface for different data centers is the concept of WebService and WebService portal. Here we have developed a prototype of data center portal for digital broadband seismograms. This WebService portal uses WSDL (Web Services Description Language) to accommodate differences among the different data centers. By using the WSDL, alteration and addition of data center user interfaces can be easily managed. This portal, called NINJA Portal, assumes three WebServices: (1) database Query service, (2) Seismic event data request service, and (3) Seismic continuous data request service. Current system supports both station search of database Query service and seismic continuous data request service. Data centers supported by this NINJA portal will be OHP data center in ERI and Pacific21 data center in IFREE/JAMSTEC in the beginning. We have developed metadata standard for seismological data based on QuakeML for parametric data, which has been developed by ETH Zurich, and XML-SEED for waveform data, which was developed by IFREE/JAMSTEC. The prototype of NINJA portal is now released through IFREE web page (http://www.jamstec.go.jp/pacific21/).

  10. The Use of XML to the Description of Devirse Observational Data

    NASA Astrophysics Data System (ADS)

    Vitkovskij, V. V.; Zhelenkova, O. P.; Kalinina, N. A.; Shergin, V. S.; Chernenkov, V. N.; Mal'Kova, G. A.

    The Internet using for scientific researches and formation of specialized environment including interoperable archives, catalogues, software is an urgent task in world astronomical community. "Virtual telescope", "virtual observatory" are uniting name for the innovation development on interoperability and scalability of on-line data. The FITS-format is the most widespread way of astronomical data storage now, but it does not allow make effective search of the necessary information, especially in conditions of constant growth of data volume. Existing industrial information technologies of management and performance of the data, such as database management systems, web-access, XML, Java, and their fast development give hope for realization of information resources interoperability in the close future. The virtual observatory concept forces in a new fashion the consideration of questions got in touch with science archives. SAO RAS archive consists from local archives that to receive observational data from diverse acquisition systems and the one is a good test example for development of similar systems. We have put a task to analyze the available data and to organize archival system for transparent access to diverse local archives, giving attention to data quality, automation of data re-processing and integration with the astronomical catalogues and surveys.

  11. Adding Hierarchical Objects to Relational Database General-Purpose XML-Based Information Managements

    NASA Technical Reports Server (NTRS)

    Lin, Shu-Chun; Knight, Chris; La, Tracy; Maluf, David; Bell, David; Tran, Khai Peter; Gawdiak, Yuri

    2006-01-01

    NETMARK is a flexible, high-throughput software system for managing, storing, and rapid searching of unstructured and semi-structured documents. NETMARK transforms such documents from their original highly complex, constantly changing, heterogeneous data formats into well-structured, common data formats in using Hypertext Markup Language (HTML) and/or Extensible Markup Language (XML). The software implements an object-relational database system that combines the best practices of the relational model utilizing Structured Query Language (SQL) with those of the object-oriented, semantic database model for creating complex data. In particular, NETMARK takes advantage of the Oracle 8i object-relational database model using physical-address data types for very efficient keyword searches of records across both context and content. NETMARK also supports multiple international standards such as WEBDAV for drag-and-drop file management and SOAP for integrated information management using Web services. The document-organization and -searching capabilities afforded by NETMARK are likely to make this software attractive for use in disciplines as diverse as science, auditing, and law enforcement.

  12. A pipeline VLSI design of fast singular value decomposition processor for real-time EEG system based on on-line recursive independent component analysis.

    PubMed

    Huang, Kuan-Ju; Shih, Wei-Yeh; Chang, Jui Chung; Feng, Chih Wei; Fang, Wai-Chi

    2013-01-01

    This paper presents a pipeline VLSI design of fast singular value decomposition (SVD) processor for real-time electroencephalography (EEG) system based on on-line recursive independent component analysis (ORICA). Since SVD is used frequently in computations of the real-time EEG system, a low-latency and high-accuracy SVD processor is essential. During the EEG system process, the proposed SVD processor aims to solve the diagonal, inverse and inverse square root matrices of the target matrices in real time. Generally, SVD requires a huge amount of computation in hardware implementation. Therefore, this work proposes a novel design concept for data flow updating to assist the pipeline VLSI implementation. The SVD processor can greatly improve the feasibility of real-time EEG system applications such as brain computer interfaces (BCIs). The proposed architecture is implemented using TSMC 90 nm CMOS technology. The sample rate of EEG raw data adopts 128 Hz. The core size of the SVD processor is 580×580 um(2), and the speed of operation frequency is 20MHz. It consumes 0.774mW of power during the 8-channel EEG system per execution time.

  13. Use of the recursion formula of the Gompertz survival function to evaluate life-table data.

    PubMed

    Bassukas, I D

    1996-08-29

    The recursion formula of the Gompertz function is an established method for the analysis of growth processes. In the present study the recursion formula of the Gompertz survival function 1n S(t + s) = a + b x ln S(t) is introduced for the analysis of survival data, where S(t) is the survival fraction at age 1, s is the constant age increment between two consecutive measurements of the survival fraction and a and b are parameters. With the help of this method--and provided stroboscopial measurements of rates of survival are available--the Gompertz survival function, instead of the corresponding mortality function, can be determined directly using linear regression analysis. The application of the present algorithm is demonstrated by analysing two sets of data taken from the literature (survival of Drosophila imagoes and of female centenarians) using linear regression analysis to fit survival or mortality rates to the corresponding models. In both cases the quality of fit was superior by using the algorithm presently introduced. Moreover, survival functions calculated from the fits to the mortality law only poorly predict the survival data. On the contrary, the results of the present method not only fit to the measurements, but, for both sets of data the mortality parameters calculated by the present method are essentially identical to those obtained by a corresponding application of a non-linear Marquardt-Levenberg algorithm to fit the same sets of data to the explicit form of the Gompertz survival function. Taking into consideration the advantages of using a linear fit (goodness-of-fit test and efficient statistical comparison of survival patterns) the method of the recursion formula of the Gompertz survival function is the most preferable method to fit survival data to the Gompertz function.

  14. SIM-Lite Update

    NASA Technical Reports Server (NTRS)

    Shao, Michael

    2008-01-01

    Discussion focus on: SIM-Lite Instrument Update - 6m baseline, 50cm, approximately 900M cost; Technology Update - Systematic errors and floor; SIM-Lite terrestrial planet discovery capability; Double blind multiple planet study summary; and the changing landscape of exoplanet science and the role of SIM-Lite. Slides include technology to flight component engineering; instrumental systematic errors; ultra deep search for Earth clones; double blind test, astrometric detection of Earths in multiplanet systems; the current era of exoplanet science and where SIM-Lite fits in; the next frontier and where SIM-Lite fits in, why SIM is unique in discovering Earths; imaging planet status is uncertain without masses and ages; SIM role in establishing how planetary systems form and evolve; and SIM probes of broad planet mass range around young stars.

  15. NXE pellicle: development update

    NASA Astrophysics Data System (ADS)

    Brouns, Derk; Bendiksen, Aage; Broman, Par; Casimiri, Eric; Colsters, Paul; de Graaf, Dennis; Harrold, Hilary; Hennus, Piet; Janssen, Paul; Kramer, Ronald; Kruizinga, Matthias; Kuntzel, Henk; Lafarre, Raymond; Mancuso, Andrea; Ockwell, David; Smith, Daniel; van de Weg, David; Wiley, Jim

    2016-09-01

    ASML introduced the NXE pellicle concept, a removable pellicle solution that is compatible with current and future patterned mask inspection methods. We will present results of how we have taken the idea from concept to a demonstrated solution enabling the use of EUV pellicle by the industry for high volume manufacturing. We will update on the development of the next generation of pellicle films with higher power capability. Further, we will provide an update on top level requirements for pellicles and external interface requirements needed to support NXE pellicle adoption at a mask shop. Finally, we will present ASML's pellicle handling equipment to enable pellicle use at mask shops and our NXE pellicle roadmap outlining future improvements.

  16. Sensors, Update 9

    NASA Astrophysics Data System (ADS)

    Baltes, Henry; Göpel, Wolfgang; Hesse, Joachim

    2001-10-01

    Sensors Update ensures that you stay at the cutting edge of the field. Built upon the series Sensors, it presents an overview of highlights in the field. Coverage includes current developments in materials, design, production, and applications of sensors, signal detection and processing, as well as new sensing principles. Each volume is divided into three sections. Sensor Technology, reviews highlights in applied and basic research, Sensor Applications, covers new or improved applications of sensors, Sensor Markets, provides a survey of suppliers and market trends for a particular area. With this unique combination of information in each volume, Sensors Update will be of value for scientists and engineers in industry and at universities, to sensors developers, distributors, and users.

  17. Sensors, Update 2

    NASA Astrophysics Data System (ADS)

    Baltes, Henry; Göpel, Wolfgang; Hesse, Joachim

    1996-10-01

    Sensors Update ensures that you stay at the cutting edge of the field. Built upon the series Sensors, it presents an overview of highlights in the field. Coverage includes current developments in materials, design, production, and applications of sensors, signal detection and processing, as well as new sensing principles. Furthermore, the sensor market as well as peripheral aspects such as standards are covered. Each volume is divided into four sections. Sensor Technology, reviews highlights in applied and basic research, Sensor Applications, covers new or improved applications of sensors, Sensor Markets, provides a survey of suppliers and market trends for a particular area. With this unique combination of information in each volume, Sensors Update will be of value for scientists and engineers in industry and at universities, to sensors developers, distributors, and users.

  18. Sensors, Update 10

    NASA Astrophysics Data System (ADS)

    Baltes, Henry; Fedder, Gary K.; Korvink, Jan G.

    2002-04-01

    Sensors Update ensures that you stay at the cutting edge of the field. Built upon the series Sensors, it presents an overview of highlights in the field. Coverage includes current developments in materials, design, production, and applications of sensors, signal detection and processing, as well as new sensing principles. Each volume is divided into three sections. Sensor Technology, reviews highlights in applied and basic research, Sensor Applications, covers new or improved applications of sensors, Sensor Markets, provides a survey of suppliers and market trends for a particular area. With this unique combination of information in each volume, Sensors Update will be of value for scientists and engineers in industry and at universities, to sensors developers, distributors, and users.

  19. Sensors, Update 1

    NASA Astrophysics Data System (ADS)

    Baltes, Henry; Göpel, Wolfgang; Hesse, Joachim

    1996-12-01

    Sensors Update ensures that you stay at the cutting edge of the field. Built upon the series Sensors, it presents an overview of highlights in the field. Treatments include current developments in materials, design, production, and applications of sensors, signal detection and processing, as well as new sensing principles. Furthermore, the sensor market as well as peripheral aspects such as standards are covered. Each volume is divided into four sections. Sensor Technology, reviews highlights in applied and basic research, Sensor Applications, covers new or improved applications of sensors, Sensor Markets, provides an overview of suppliers and market trends for a particular section, and Sensor Standards, reviews recent legislation and requirements for sensors. With this unique combination of information in each volume, Sensors Update will be of value for scientists and engineers in industry and at universities, to sensors developers, distributors, and users.

  20. Sensors, Update 12

    NASA Astrophysics Data System (ADS)

    Baltes, Henry; Fedder, Gary K.; Korvink, Jan G.

    2003-04-01

    Sensors Update ensures that you stay at the cutting edge of the field. Built upon the series Sensors, it presents an overview of highlights in the field. Coverage includes current developments in materials, design, production, and applications of sensors, signal detection and processing, as well as new sensing principles. Each volume is divided into three sections. Sensor Technology, reviews highlights in applied and basic research, Sensor Applications, covers new or improved applications of sensors, Sensor Markets, provides a survey of suppliers and market trends for a particular area. With this unique combination of information in each volume, Sensors Update will be of value for scientists and engineers in industry and at universities, to sensors developers, distributors, and users.

  1. Sensors, Update 8

    NASA Astrophysics Data System (ADS)

    Baltes, Henry; Göpel, Wolfgang; Hesse, Joachim

    2001-02-01

    Sensors Update ensures that you stay at the cutting edge of the field. Built upon the series Sensors, it presents an overview of highlights in the field. Coverage includes current developments in materials, design, production, and applications of sensors, signal detection and processing, as well as new sensing principles. Each volume is divided into three sections: Sensor Technology reviews highlights in applied and basic research, while Sensor Applications covers new or improved applications of sensors, and Sensor Markets provides a survey of suppliers and market trends for a particular area. With this unique combination of information in each volume, Sensors Update will be invaluable to scientists and engineers in industry and at universities, to sensors developers, distributors, and users.

  2. Sequence History Update Tool

    NASA Technical Reports Server (NTRS)

    Khanampompan, Teerapat; Gladden, Roy; Fisher, Forest; DelGuercio, Chris

    2008-01-01

    The Sequence History Update Tool performs Web-based sequence statistics archiving for Mars Reconnaissance Orbiter (MRO). Using a single UNIX command, the software takes advantage of sequencing conventions to automatically extract the needed statistics from multiple files. This information is then used to populate a PHP database, which is then seamlessly formatted into a dynamic Web page. This tool replaces a previous tedious and error-prone process of manually editing HTML code to construct a Web-based table. Because the tool manages all of the statistics gathering and file delivery to and from multiple data sources spread across multiple servers, there is also a considerable time and effort savings. With the use of The Sequence History Update Tool what previously took minutes is now done in less than 30 seconds, and now provides a more accurate archival record of the sequence commanding for MRO.

  3. Recursive construction of perfect DNA molecules and libraries from imperfect oligonucleotides.

    PubMed

    Linshiz, Gregory; Yehezkel, Tuval Ben; Shapiro, Ehud

    2012-01-01

    Making faultless complex objects from potentially faulty building blocks is a fundamental challenge in computer engineering, nanotechnology, and synthetic biology. We developed an error-correcting recursive construction procedure that attempts to address this challenge. Making DNA molecules from synthetic oligonucleotides using the procedure described here surpasses existing methods for de novo DNA synthesis in speed, precision, and amenability to automation. It provides for the first time a unified DNA construction platform for combining synthetic and natural DNA fragments, for constructing designer DNA libraries, and for making the faultless long synthetic DNA building blocks needed for de novo genome construction.

  4. System Simulation by Recursive Feedback: Coupling a Set of Stand-Alone Subsystem Simulations

    NASA Technical Reports Server (NTRS)

    Nixon, Douglas D.; Ryan, Stephen G. (Technical Monitor)

    2002-01-01

    Recursive feedback is defined and discussed as a framework for development of specific algorithms and procedures that propagate the time-domain solution for a dynamical system simulation consisting of multiple numerically coupled, self-contained, stand-alone subsystem simulations. A satellite motion example containing three subsystems (orbit dynamics, attitude dynamics, and aerodynamics) has been defined and constructed using this approach. Conventional solution methods are used in the subsystem simulations. Centralized and distributed versions of coupling structure have been addressed. Numerical results are evaluated by direct comparison with a standard total-system, simultaneous-solution approach.

  5. Chandrasekhar-type algorithms for fast recursive estimation in linear systems with constant parameters

    NASA Technical Reports Server (NTRS)

    Choudhury, A. K.; Djalali, M.

    1975-01-01

    In this recursive method proposed, the gain matrix for the Kalman filter and the convariance of the state vector are computed not via the Riccati equation, but from certain other equations. These differential equations are of Chandrasekhar-type. The 'invariant imbedding' idea resulted in the reduction of the basic boundary value problem of transport theory to an equivalent initial value system, a significant computational advance. Initial value experience showed that there is some computational savings in the method and the loss of positive definiteness of the covariance matrix is less vulnerable.

  6. System Simulation by Recursive Feedback: Coupling A Set of Stand-Alone Subsystem Simulations

    NASA Technical Reports Server (NTRS)

    Nixon, Douglas D.; Hanson, John M. (Technical Monitor)

    2002-01-01

    Recursive feedback is defined and discussed as a framework for development of specific algorithms and procedures that propagate the time-domain solution for a dynamical system simulation consisting of multiple numerically coupled self-contained stand-alone subsystem simulations. A satellite motion example containing three subsystems (other dynamics, attitude dynamics, and aerodynamics) has been defined and constructed using this approach. Conventional solution methods are used in the subsystem simulations. Centralized and distributed versions of coupling structure have been addressed. Numerical results are evaluated by direct comparison with a standard total-system simultaneous-solution approach.

  7. Recursive Recovery of Sparse Signal Sequences From Compressive Measurements: A Review

    NASA Astrophysics Data System (ADS)

    Vaswani, Namrata; Zhan, Jinchun

    2016-07-01

    In this article, we review the literature on design and analysis of recursive algorithms for reconstructing a time sequence of sparse signals from compressive measurements. The signals are assumed to be sparse in some transform domain or in some dictionary. Their sparsity patterns can change with time, although, in many practical applications, the changes are gradual. An important class of applications where this problem occurs is dynamic projection imaging, e.g., dynamic magnetic resonance imaging (MRI) for real-time medical applications such as interventional radiology, or dynamic computed tomography.

  8. Recursive identification and tracking of parameters for linear and nonlinear multivariable systems

    NASA Technical Reports Server (NTRS)

    Sidar, M.

    1975-01-01

    The problem of identifying constant and variable parameters in multi-input, multi-output, linear and nonlinear systems is considered, using the maximum likelihood approach. An iterative algorithm, leading to recursive identification and tracking of the unknown parameters and the noise covariance matrix, is developed. Agile tracking, and accurate and unbiased identified parameters are obtained. Necessary conditions for a globally, asymptotically stable identification process are provided; the conditions proved to be useful and efficient. Among different cases studied, the stability derivatives of an aircraft were identified and some of the results are shown as examples.

  9. Cyclic period-3 window in antiferromagnetic potts and Ising models on recursive lattices

    NASA Astrophysics Data System (ADS)

    Ananikian, N. S.; Ananikyan, L. N.; Chakhmakhchyan, L. A.

    2011-09-01

    The magnetic properties of the antiferromagnetic Potts model with two-site interaction and the antiferromagnetic Ising model with three-site interaction on recursive lattices have been studied. A cyclic period-3 window has been revealed by the recurrence relation method in the antiferromagnetic Q-state Potts model on the Bethe lattice (at Q < 2) and in the antiferromagnetic Ising model with three-site interaction on the Husimi cactus. The Lyapunov exponents have been calculated, modulated phases and a chaotic regime in the cyclic period-3 window have been found for one-dimensional rational mappings determined the properties of these systems.

  10. Recursive polarization of nuclear spins in diamond at arbitrary magnetic fields

    SciTech Connect

    Pagliero, Daniela; Laraoui, Abdelghani; Henshaw, Jacob D.; Meriles, Carlos A.

    2014-12-15

    We introduce an alternate route to dynamically polarize the nuclear spin host of nitrogen-vacancy (NV) centers in diamond. Our approach articulates optical, microwave, and radio-frequency pulses to recursively transfer spin polarization from the NV electronic spin. Using two complementary variants of the same underlying principle, we demonstrate nitrogen nuclear spin initialization approaching 80% at room temperature both in ensemble and single NV centers. Unlike existing schemes, our approach does not rely on level anti-crossings and is thus applicable at arbitrary magnetic fields. This versatility should prove useful in applications ranging from nanoscale metrology to sensitivity-enhanced NMR.

  11. Recursive solution of number of reachable states of a simple subclass of FMS

    NASA Astrophysics Data System (ADS)

    Chao, Daniel Yuh

    2014-03-01

    This paper aims to compute the number of reachable (forbidden, live and deadlock) states for flexible manufacturing systems (FMS) without the construction of reachability graph. The problem is nontrivial and takes, in general, an exponential amount of time to solve. Hence, this paper focusses on a simple version of Systems of Simple Sequential Processes with Resources (S3PR), called kth-order system, where each resource place holds one token to be shared between two processes. The exact number of reachable (forbidden, live and deadlock) states can be computed recursively.

  12. Endodontic diagnostic terminology update.

    PubMed

    McClannahan, Scott B; Baisden, Michael K; Bowles, Walter R

    2011-01-01

    Determination of the etiology of the patient's chief complaint and a correct diagnosis are paramount prior to a recommendation of endodontic therapy. Reproduction of the patient's chief complaint is critical. If the chief complaint cannot be reproduced, consider consultation with or referral to an endodontist or orofacial pain specialist. The diagnostic terminology presented in this update provides for a more accurate description and communication of the health or pathological conditions of both pulpal and apical tissues. This information is summarized in Table I.

  13. How Documentalists Update SIMBAD

    NASA Astrophysics Data System (ADS)

    Buga, M.; Bot, C.; Brouty, M.; Bruneau, C.; Brunet, C.; Cambresy, L.; Eisele, A.; Genova, F.; Lesteven, S.; Loup, C.; Neuville, M.; Oberto, A.; Ochsenbein, F.; Perret, E.; Siebert, A.; Son, E.; Vannier, P.; Vollmer, B.; Vonflie, P.; Wenger, M.; Woelfel, F.

    2015-04-01

    The Strasbourg astronomical Data Center (CDS) was created in 1972 and has had a major role in astronomy for more than forty years. CDS develops a service called SIMBAD that provides basic data, cross-identifications, bibliography, and measurements for astronomical objects outside the solar system. It brings to the scientific community an added value to content which is updated daily by a team of documentalists working together in close collaboration with astronomers and IT specialists. We explain how the CDS staff updates SIMBAD with object citations in the main astronomical journals, as well as with astronomical data and measurements. We also explain how the identification is made between the objects found in the literature and those already existing in SIMBAD. We show the steps followed by the documentalist team to update the database using different tools developed at CDS, like the sky visualizer Aladin, and the large catalogues and survey database VizieR. As a direct result of this teamwork, SIMBAD integrates almost 10.000 bibliographic references per year. The service receives more than 400.000 queries per day.

  14. SU-E-T-48: Automated Quality Assurance for XML Controlled Linacs

    SciTech Connect

    Valdes, G; Morin, O; Pouliot, J; Chuang, C

    2014-06-01

    Purpose: To automate routine imaging QA procedures so that complying with TG 142 and TG 179 can be efficient and reliable. Methods: Two QA tests for a True Beam Linac were automatized. A Winston Lutz test as described by Lutz et al{sup 1} using the Winston Lutz test kit from BrainLab, Germany and a CBCT Image Quality test as described in TG 179 using the EMMA phantom, Siemens Medical Physics, Germany were performed in our True Beam. For each QA procedure tested, a 3 step paradigm was used. First, the data was automatically acquired using True Beam Developer Mode and XML scripting. Second, the data acquired in the first step was automatically processed using in-home grown Matlab GUIs. Third, Machine Learning algorithms were used to automatically classify the processed data and reports generated. Results: The Winston Luzt test could be performed by an experienced medical physicist in 29.0 ± 8.0 min. The same test, if automated using our paradigm, could be performed in 3.0 ± 0.1 min. In the same lieu, time could be substantially saved for image quality tests. In this case, the amount of time saved will depend on the phantoms used and the initial localization method. Additionally, machine learning algorithms could automatically identify the roots of the problems if any and possibly help reduce machine down time. Conclusion: Modern linear accelerators are equipped with advanced 2D and 3D imaging that are used for patient alignment substantially improving IGRT protocols. However, this extra complexity exponentially increases the number of QA tests needed. Using the new paradigm described above, not only bare minimum but best practice QA programs could be implemented with the same manpower. This work is supported by Varian, Palo Alto, CA.

  15. A recursive field-normalized bibliometric performance indicator: an application to the field of library and information science.

    PubMed

    Waltman, Ludo; Yan, Erjia; van Eck, Nees Jan

    2011-10-01

    Two commonly used ideas in the development of citation-based research performance indicators are the idea of normalizing citation counts based on a field classification scheme and the idea of recursive citation weighing (like in PageRank-inspired indicators). We combine these two ideas in a single indicator, referred to as the recursive mean normalized citation score indicator, and we study the validity of this indicator. Our empirical analysis shows that the proposed indicator is highly sensitive to the field classification scheme that is used. The indicator also has a strong tendency to reinforce biases caused by the classification scheme. Based on these observations, we advise against the use of indicators in which the idea of normalization based on a field classification scheme and the idea of recursive citation weighing are combined.

  16. Recursive Averaging

    ERIC Educational Resources Information Center

    Smith, Scott G.

    2015-01-01

    In this article, Scott Smith presents an innocent problem (Problem 12 of the May 2001 Calendar from "Mathematics Teacher" ("MT" May 2001, vol. 94, no. 5, p. 384) that was transformed by several timely "what if?" questions into a rewarding investigation of some interesting mathematics. These investigations led to two…

  17. Poster — Thur Eve — 55: An automated XML technique for isocentre verification on the Varian TrueBeam

    SciTech Connect

    Asiev, Krum; Mullins, Joel; DeBlois, François; Liang, Liheng; Syme, Alasdair

    2014-08-15

    Isocentre verification tests, such as the Winston-Lutz (WL) test, have gained popularity in the recent years as techniques such as stereotactic radiosurgery/radiotherapy (SRS/SRT) treatments are more commonly performed on radiotherapy linacs. These highly conformal treatments require frequent monitoring of the geometrical accuracy of the isocentre to ensure proper radiation delivery. At our clinic, the WL test is performed by acquiring with the EPID a collection of 8 images of a WL phantom fixed on the couch for various couch/gantry angles. This set of images is later analyzed to determine the isocentre size. The current work addresses the acquisition process. A manual WL test acquisition performed by and experienced physicist takes in average 25 minutes and is prone to user manipulation errors. We have automated this acquisition on a Varian TrueBeam STx linac (Varian, Palo Alto, USA). The Varian developer mode allows the execution of custom-made XML script files to control all aspects of the linac operation. We have created an XML-WL script that cycles through each couch/gantry combinations taking an EPID image at each position. This automated acquisition is done in less than 4 minutes. The reproducibility of the method was verified by repeating the execution of the XML file 5 times. The analysis of the images showed variation of the isocenter size less than 0.1 mm along the X, Y and Z axes and compares favorably to a manual acquisition for which we typically observe variations up to 0.5 mm.

  18. A recursive regularization algorithm for estimating the particle size distribution from multiangle dynamic light scattering measurements

    NASA Astrophysics Data System (ADS)

    Li, Lei; Yang, Kecheng; Li, Wei; Wang, Wanyan; Guo, Wenping; Xia, Min

    2016-07-01

    Conventional regularization methods have been widely used for estimating particle size distribution (PSD) in single-angle dynamic light scattering, but they could not be used directly in multiangle dynamic light scattering (MDLS) measurements for lack of accurate angular weighting coefficients, which greatly affects the PSD determination and none of the regularization methods perform well for both unimodal and multimodal distributions. In this paper, we propose a recursive regularization method-Recursion Nonnegative Tikhonov-Phillips-Twomey (RNNT-PT) algorithm for estimating the weighting coefficients and PSD from MDLS data. This is a self-adaptive algorithm which distinguishes characteristics of PSDs and chooses the optimal inversion method from Nonnegative Tikhonov (NNT) and Nonnegative Phillips-Twomey (NNPT) regularization algorithm efficiently and automatically. In simulations, the proposed algorithm was able to estimate the PSDs more accurately than the classical regularization methods and performed stably against random noise and adaptable to both unimodal and multimodal distributions. Furthermore, we found that the six-angle analysis in the 30-130° range is an optimal angle set for both unimodal and multimodal PSDs.

  19. Kernel Recursive Least-Squares Temporal Difference Algorithms with Sparsification and Regularization

    PubMed Central

    Zhu, Qingxin; Niu, Xinzheng

    2016-01-01

    By combining with sparse kernel methods, least-squares temporal difference (LSTD) algorithms can construct the feature dictionary automatically and obtain a better generalization ability. However, the previous kernel-based LSTD algorithms do not consider regularization and their sparsification processes are batch or offline, which hinder their widespread applications in online learning problems. In this paper, we combine the following five techniques and propose two novel kernel recursive LSTD algorithms: (i) online sparsification, which can cope with unknown state regions and be used for online learning, (ii) L2 and L1 regularization, which can avoid overfitting and eliminate the influence of noise, (iii) recursive least squares, which can eliminate matrix-inversion operations and reduce computational complexity, (iv) a sliding-window approach, which can avoid caching all history samples and reduce the computational cost, and (v) the fixed-point subiteration and online pruning, which can make L1 regularization easy to implement. Finally, simulation results on two 50-state chain problems demonstrate the effectiveness of our algorithms. PMID:27436996

  20. A recursive genetic framework for evolutionary decision-making in problems with high dynamism

    NASA Astrophysics Data System (ADS)

    Pashaei, Kaveh; Taghiyareh, Fattaneh; Badie, Kambiz

    2015-11-01

    Communication and coordination are the main cores for reaching a constructive agreement among multi-agent systems (MASs). Dividing the overall performance of MAS to individual agents may lead to group learning as opposed to individual learning, which is one of the weak points of MASs. This paper proposes a recursive genetic framework for solving problems with high dynamism. In this framework, a combination of genetic algorithm and multi-agent capabilities is utilised to accelerate team learning and accurate credit assignment. The argumentation feature is used to accomplish agent learning and the negotiation features of MASs are used to achieve a credit assignment. The proposed framework is quite general and its recursive hierarchical structure could be extended. We have dedicated one special controlling module for increasing convergence time. Due to the complexity of blackjack, we have applied it as a possible test bed to evaluate the system's performance. The learning rate of agents is measured as well as their credit assignment. The analysis of the obtained results led us to believe that our robust framework with the proposed negotiation operator is a promising methodology to solve similar problems in other areas with high dynamism.